[ 511.979355] env[61957]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'linux_bridge' {{(pid=61957) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 511.979707] env[61957]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'noop' {{(pid=61957) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 511.979821] env[61957]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'ovs' {{(pid=61957) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 511.980127] env[61957]: INFO os_vif [-] Loaded VIF plugins: linux_bridge, noop, ovs [ 512.074944] env[61957]: DEBUG oslo_concurrency.processutils [-] Running cmd (subprocess): grep -F node.session.scan /sbin/iscsiadm {{(pid=61957) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:390}} [ 512.084717] env[61957]: DEBUG oslo_concurrency.processutils [-] CMD "grep -F node.session.scan /sbin/iscsiadm" returned: 0 in 0.010s {{(pid=61957) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:428}} [ 512.684109] env[61957]: INFO nova.virt.driver [None req-e7cc684f-7947-44de-b028-e3382f08c36d None None] Loading compute driver 'vmwareapi.VMwareVCDriver' [ 512.754141] env[61957]: DEBUG oslo_concurrency.lockutils [-] Acquiring lock "oslo_vmware_api_lock" by "oslo_vmware.api.VMwareAPISession._create_session" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 512.754350] env[61957]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" acquired by "oslo_vmware.api.VMwareAPISession._create_session" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 512.754472] env[61957]: DEBUG oslo_vmware.service [-] Creating suds client with soap_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk' and wsdl_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk/vimService.wsdl' {{(pid=61957) __init__ /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:242}} [ 515.897752] env[61957]: DEBUG oslo_vmware.service [-] Invoking ServiceInstance.RetrieveServiceContent with opID=oslo.vmware-7b5abf3e-8549-4833-9003-8ede713440da {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 515.913848] env[61957]: DEBUG oslo_vmware.api [-] Logging into host: vc1.osci.c.eu-de-1.cloud.sap. {{(pid=61957) _create_session /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:242}} [ 515.914091] env[61957]: DEBUG oslo_vmware.service [-] Invoking SessionManager.Login with opID=oslo.vmware-443d2eeb-ab48-4020-9a9b-a45f4430fec9 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 515.945161] env[61957]: INFO oslo_vmware.api [-] Successfully established new session; session ID is a0283. [ 515.945403] env[61957]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" "released" by "oslo_vmware.api.VMwareAPISession._create_session" :: held 3.191s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 515.945873] env[61957]: INFO nova.virt.vmwareapi.driver [None req-e7cc684f-7947-44de-b028-e3382f08c36d None None] VMware vCenter version: 7.0.3 [ 515.949360] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a9ffd38-fc49-4b13-a77e-af926465adc3 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 515.971088] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4d66525-f507-4c2b-b3ca-07d97bc0fb31 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 515.977044] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c51a9460-fc3b-44e3-94ce-bd420160795e {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 515.983646] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e01d3c08-058e-4dbd-9c43-a4274074f37e {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 515.996547] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a772516-df12-4631-95e1-58169692fcc2 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 516.002524] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eaa6190d-94cf-4385-be52-06a873723856 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 516.032777] env[61957]: DEBUG oslo_vmware.service [-] Invoking ExtensionManager.FindExtension with opID=oslo.vmware-bc8b087e-d4b5-4d05-8658-40518e1ead5d {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 516.037918] env[61957]: DEBUG nova.virt.vmwareapi.driver [None req-e7cc684f-7947-44de-b028-e3382f08c36d None None] Extension org.openstack.compute already exists. {{(pid=61957) _register_openstack_extension /opt/stack/nova/nova/virt/vmwareapi/driver.py:225}} [ 516.040597] env[61957]: INFO nova.compute.provider_config [None req-e7cc684f-7947-44de-b028-e3382f08c36d None None] No provider configs found in /etc/nova/provider_config/. If files are present, ensure the Nova process has access. [ 516.544346] env[61957]: DEBUG nova.context [None req-e7cc684f-7947-44de-b028-e3382f08c36d None None] Found 2 cells: 00000000-0000-0000-0000-000000000000(cell0),7ea34dbf-443a-43a7-a1c4-ecdb29a8d307(cell1) {{(pid=61957) load_cells /opt/stack/nova/nova/context.py:464}} [ 516.547430] env[61957]: DEBUG oslo_concurrency.lockutils [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 516.547675] env[61957]: DEBUG oslo_concurrency.lockutils [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 516.548363] env[61957]: DEBUG oslo_concurrency.lockutils [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 516.548782] env[61957]: DEBUG oslo_concurrency.lockutils [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] Acquiring lock "7ea34dbf-443a-43a7-a1c4-ecdb29a8d307" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 516.548976] env[61957]: DEBUG oslo_concurrency.lockutils [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] Lock "7ea34dbf-443a-43a7-a1c4-ecdb29a8d307" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 516.549968] env[61957]: DEBUG oslo_concurrency.lockutils [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] Lock "7ea34dbf-443a-43a7-a1c4-ecdb29a8d307" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 516.569662] env[61957]: INFO dbcounter [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] Registered counter for database nova_cell0 [ 516.577681] env[61957]: INFO dbcounter [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] Registered counter for database nova_cell1 [ 516.580856] env[61957]: DEBUG oslo_db.sqlalchemy.engines [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=61957) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:342}} [ 516.581222] env[61957]: DEBUG oslo_db.sqlalchemy.engines [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=61957) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:342}} [ 516.586015] env[61957]: ERROR nova.db.main.api [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 516.586015] env[61957]: result = function(*args, **kwargs) [ 516.586015] env[61957]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 516.586015] env[61957]: return func(*args, **kwargs) [ 516.586015] env[61957]: File "/opt/stack/nova/nova/context.py", line 422, in gather_result [ 516.586015] env[61957]: result = fn(*args, **kwargs) [ 516.586015] env[61957]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 516.586015] env[61957]: return f(*args, **kwargs) [ 516.586015] env[61957]: File "/opt/stack/nova/nova/objects/service.py", line 553, in _db_service_get_minimum_version [ 516.586015] env[61957]: return db.service_get_minimum_version(context, binaries) [ 516.586015] env[61957]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 516.586015] env[61957]: _check_db_access() [ 516.586015] env[61957]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 516.586015] env[61957]: stacktrace = ''.join(traceback.format_stack()) [ 516.586015] env[61957]: [ 516.586873] env[61957]: ERROR nova.db.main.api [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 516.586873] env[61957]: result = function(*args, **kwargs) [ 516.586873] env[61957]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 516.586873] env[61957]: return func(*args, **kwargs) [ 516.586873] env[61957]: File "/opt/stack/nova/nova/context.py", line 422, in gather_result [ 516.586873] env[61957]: result = fn(*args, **kwargs) [ 516.586873] env[61957]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 516.586873] env[61957]: return f(*args, **kwargs) [ 516.586873] env[61957]: File "/opt/stack/nova/nova/objects/service.py", line 553, in _db_service_get_minimum_version [ 516.586873] env[61957]: return db.service_get_minimum_version(context, binaries) [ 516.586873] env[61957]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 516.586873] env[61957]: _check_db_access() [ 516.586873] env[61957]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 516.586873] env[61957]: stacktrace = ''.join(traceback.format_stack()) [ 516.586873] env[61957]: [ 516.587338] env[61957]: WARNING nova.objects.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] Failed to get minimum service version for cell 00000000-0000-0000-0000-000000000000 [ 516.587388] env[61957]: WARNING nova.objects.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] Failed to get minimum service version for cell 7ea34dbf-443a-43a7-a1c4-ecdb29a8d307 [ 516.587804] env[61957]: DEBUG oslo_concurrency.lockutils [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] Acquiring lock "singleton_lock" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 516.588059] env[61957]: DEBUG oslo_concurrency.lockutils [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] Acquired lock "singleton_lock" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 516.588438] env[61957]: DEBUG oslo_concurrency.lockutils [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] Releasing lock "singleton_lock" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 516.588914] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] Full set of CONF: {{(pid=61957) _wait_for_exit_or_signal /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/service.py:363}} [ 516.589177] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] ******************************************************************************** {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2806}} [ 516.589406] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] Configuration options gathered from: {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2807}} [ 516.589638] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] command line args: ['--config-file', '/etc/nova/nova.conf', '--config-file', '/etc/nova/nova-cpu-common.conf', '--config-file', '/etc/nova/nova-cpu-1.conf'] {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2808}} [ 516.589929] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] config files: ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2809}} [ 516.590160] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] ================================================================================ {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2811}} [ 516.590429] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] allow_resize_to_same_host = True {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.590716] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] arq_binding_timeout = 300 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.590927] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] backdoor_port = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.591189] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] backdoor_socket = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.591479] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] block_device_allocate_retries = 60 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.591773] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] block_device_allocate_retries_interval = 3 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.592104] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] cert = self.pem {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.592420] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] compute_driver = vmwareapi.VMwareVCDriver {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.592642] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] compute_monitors = [] {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.592824] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] config_dir = [] {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.593010] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] config_drive_format = iso9660 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.593151] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] config_file = ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.593319] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] config_source = [] {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.593485] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] console_host = devstack {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.593647] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] control_exchange = nova {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.593808] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] cpu_allocation_ratio = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.593964] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] daemon = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.594143] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] debug = True {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.594303] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] default_access_ip_network_name = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.594468] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] default_availability_zone = nova {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.594646] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] default_ephemeral_format = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.594816] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] default_green_pool_size = 1000 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.595058] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] default_log_levels = ['amqp=WARN', 'amqplib=WARN', 'boto=WARN', 'qpid=WARN', 'sqlalchemy=WARN', 'suds=INFO', 'oslo.messaging=INFO', 'oslo_messaging=INFO', 'iso8601=WARN', 'requests.packages.urllib3.connectionpool=WARN', 'urllib3.connectionpool=WARN', 'websocket=WARN', 'requests.packages.urllib3.util.retry=WARN', 'urllib3.util.retry=WARN', 'keystonemiddleware=WARN', 'routes.middleware=WARN', 'stevedore=WARN', 'taskflow=WARN', 'keystoneauth=WARN', 'oslo.cache=INFO', 'oslo_policy=INFO', 'dogpile.core.dogpile=INFO', 'glanceclient=WARN', 'oslo.privsep.daemon=INFO'] {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.595227] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] default_schedule_zone = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.595386] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] disk_allocation_ratio = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.595544] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] enable_new_services = True {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.595719] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] enabled_apis = ['osapi_compute'] {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.595880] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] enabled_ssl_apis = [] {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.596049] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] flat_injected = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.596207] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] force_config_drive = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.596361] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] force_raw_images = True {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.596549] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] graceful_shutdown_timeout = 5 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.596718] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] heal_instance_info_cache_interval = 60 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.596930] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] host = cpu-1 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.597117] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] initial_cpu_allocation_ratio = 4.0 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.597280] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] initial_disk_allocation_ratio = 1.0 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.597439] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] initial_ram_allocation_ratio = 1.0 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.597650] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] injected_network_template = /opt/stack/nova/nova/virt/interfaces.template {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.597812] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] instance_build_timeout = 0 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.597969] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] instance_delete_interval = 300 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.598143] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] instance_format = [instance: %(uuid)s] {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.598308] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] instance_name_template = instance-%08x {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.598465] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] instance_usage_audit = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.598647] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] instance_usage_audit_period = month {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.598831] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] instance_uuid_format = [instance: %(uuid)s] {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.598996] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] instances_path = /opt/stack/data/nova/instances {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.599174] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] internal_service_availability_zone = internal {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.599327] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] key = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.599479] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] live_migration_retry_count = 30 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.599645] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] log_color = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.599806] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] log_config_append = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.599965] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] log_date_format = %Y-%m-%d %H:%M:%S {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.600136] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] log_dir = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.600291] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] log_file = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.600417] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] log_options = True {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.600574] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] log_rotate_interval = 1 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.600743] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] log_rotate_interval_type = days {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.600908] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] log_rotation_type = none {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.601040] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] logging_context_format_string = %(color)s%(levelname)s %(name)s [%(global_request_id)s %(request_id)s %(project_name)s %(user_name)s%(color)s] %(instance)s%(color)s%(message)s {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.601167] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] logging_debug_format_suffix = {{(pid=%(process)d) %(funcName)s %(pathname)s:%(lineno)d}} {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.601333] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] logging_default_format_string = %(color)s%(levelname)s %(name)s [-%(color)s] %(instance)s%(color)s%(message)s {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.601491] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] logging_exception_prefix = ERROR %(name)s %(instance)s {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.601615] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] logging_user_identity_format = %(user)s %(project)s %(domain)s %(system_scope)s %(user_domain)s %(project_domain)s {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.601780] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] long_rpc_timeout = 1800 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.601936] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] max_concurrent_builds = 10 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.602102] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] max_concurrent_live_migrations = 1 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.602261] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] max_concurrent_snapshots = 5 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.602418] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] max_local_block_devices = 3 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.602575] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] max_logfile_count = 30 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.602734] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] max_logfile_size_mb = 200 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.602892] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] maximum_instance_delete_attempts = 5 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.603066] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] metadata_listen = 0.0.0.0 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.603233] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] metadata_listen_port = 8775 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.603396] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] metadata_workers = 2 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.603553] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] migrate_max_retries = -1 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.603717] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] mkisofs_cmd = genisoimage {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.603919] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] my_block_storage_ip = 10.180.1.21 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.604061] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] my_ip = 10.180.1.21 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.604223] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] network_allocate_retries = 0 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.604433] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] non_inheritable_image_properties = ['cache_in_nova', 'bittorrent'] {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.604582] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] osapi_compute_listen = 0.0.0.0 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.604760] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] osapi_compute_listen_port = 8774 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.604931] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] osapi_compute_unique_server_name_scope = {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.605111] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] osapi_compute_workers = 2 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.605274] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] password_length = 12 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.605433] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] periodic_enable = True {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.605591] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] periodic_fuzzy_delay = 60 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.605757] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] pointer_model = usbtablet {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.605921] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] preallocate_images = none {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.606092] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] publish_errors = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.606220] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] pybasedir = /opt/stack/nova {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.606374] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] ram_allocation_ratio = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.606544] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] rate_limit_burst = 0 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.606729] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] rate_limit_except_level = CRITICAL {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.606896] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] rate_limit_interval = 0 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.607067] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] reboot_timeout = 0 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.607228] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] reclaim_instance_interval = 0 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.607382] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] record = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.607550] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] reimage_timeout_per_gb = 60 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.607714] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] report_interval = 120 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.607875] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] rescue_timeout = 0 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.608042] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] reserved_host_cpus = 0 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.608204] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] reserved_host_disk_mb = 0 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.608360] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] reserved_host_memory_mb = 512 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.608521] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] reserved_huge_pages = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.608678] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] resize_confirm_window = 0 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.608841] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] resize_fs_using_block_device = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.609028] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] resume_guests_state_on_host_boot = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.609201] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] rootwrap_config = /etc/nova/rootwrap.conf {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.609363] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] rpc_response_timeout = 60 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.609520] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] run_external_periodic_tasks = True {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.609687] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] running_deleted_instance_action = reap {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.609848] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] running_deleted_instance_poll_interval = 1800 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.610011] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] running_deleted_instance_timeout = 0 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.610174] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] scheduler_instance_sync_interval = 120 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.610342] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] service_down_time = 720 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.610506] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] servicegroup_driver = db {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.610675] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] shell_completion = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.610863] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] shelved_offload_time = 0 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.611036] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] shelved_poll_interval = 3600 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.611211] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] shutdown_timeout = 0 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.611374] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] source_is_ipv6 = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.611530] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] ssl_only = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.611775] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] state_path = /opt/stack/data/n-cpu-1 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.611939] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] sync_power_state_interval = 600 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.612112] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] sync_power_state_pool_size = 1000 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.612279] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] syslog_log_facility = LOG_USER {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.612436] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] tempdir = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.612591] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] timeout_nbd = 10 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.612758] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] transport_url = **** {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.612918] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] update_resources_interval = 0 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.613089] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] use_cow_images = True {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.613245] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] use_eventlog = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.613399] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] use_journal = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.613553] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] use_json = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.613709] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] use_rootwrap_daemon = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.613864] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] use_stderr = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.614028] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] use_syslog = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.614186] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] vcpu_pin_set = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.614347] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] vif_plugging_is_fatal = True {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.614514] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] vif_plugging_timeout = 300 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.614698] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] virt_mkfs = [] {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.614859] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] volume_usage_poll_interval = 0 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.615022] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] watch_log_file = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.615189] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] web = /usr/share/spice-html5 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 516.615369] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] oslo_concurrency.disable_process_locking = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.615654] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] oslo_concurrency.lock_path = /opt/stack/data/n-cpu-1 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.615834] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] oslo_messaging_metrics.metrics_buffer_size = 1000 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.615995] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] oslo_messaging_metrics.metrics_enabled = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.616179] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] oslo_messaging_metrics.metrics_process_name = {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.616348] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] oslo_messaging_metrics.metrics_socket_file = /var/tmp/metrics_collector.sock {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.616518] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] oslo_messaging_metrics.metrics_thread_stop_timeout = 10 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.616718] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] api.auth_strategy = keystone {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.616890] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] api.compute_link_prefix = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.617075] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] api.config_drive_skip_versions = 1.0 2007-01-19 2007-03-01 2007-08-29 2007-10-10 2007-12-15 2008-02-01 2008-09-01 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.617254] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] api.dhcp_domain = novalocal {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.617423] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] api.enable_instance_password = True {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.617589] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] api.glance_link_prefix = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.617754] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] api.instance_list_cells_batch_fixed_size = 100 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.617926] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] api.instance_list_cells_batch_strategy = distributed {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.618099] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] api.instance_list_per_project_cells = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.618268] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] api.list_records_by_skipping_down_cells = True {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.618433] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] api.local_metadata_per_cell = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.618600] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] api.max_limit = 1000 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.618770] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] api.metadata_cache_expiration = 15 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.618967] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] api.neutron_default_tenant_id = default {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.619160] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] api.response_validation = warn {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.619333] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] api.use_neutron_default_nets = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.619504] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] api.vendordata_dynamic_connect_timeout = 5 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.619669] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] api.vendordata_dynamic_failure_fatal = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.619841] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] api.vendordata_dynamic_read_timeout = 5 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.620025] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] api.vendordata_dynamic_ssl_certfile = {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.620195] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] api.vendordata_dynamic_targets = [] {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.620361] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] api.vendordata_jsonfile_path = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.620540] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] api.vendordata_providers = ['StaticJSON'] {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.620738] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] cache.backend = dogpile.cache.memcached {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.620901] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] cache.backend_argument = **** {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.621081] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] cache.config_prefix = cache.oslo {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.621254] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] cache.dead_timeout = 60.0 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.621414] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] cache.debug_cache_backend = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.621574] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] cache.enable_retry_client = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.621735] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] cache.enable_socket_keepalive = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.621907] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] cache.enabled = True {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.622081] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] cache.enforce_fips_mode = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.622248] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] cache.expiration_time = 600 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.622411] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] cache.hashclient_retry_attempts = 2 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.622576] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] cache.hashclient_retry_delay = 1.0 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.622741] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] cache.memcache_dead_retry = 300 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.622898] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] cache.memcache_password = **** {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.623071] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] cache.memcache_pool_connection_get_timeout = 10 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.623237] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] cache.memcache_pool_flush_on_reconnect = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.623405] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] cache.memcache_pool_maxsize = 10 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.623565] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] cache.memcache_pool_unused_timeout = 60 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.623730] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] cache.memcache_sasl_enabled = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.623905] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] cache.memcache_servers = ['localhost:11211'] {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.624085] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] cache.memcache_socket_timeout = 1.0 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.624246] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] cache.memcache_username = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.624434] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] cache.proxies = [] {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.624564] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] cache.redis_db = 0 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.624723] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] cache.redis_password = **** {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.624889] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] cache.redis_sentinel_service_name = mymaster {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.625073] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] cache.redis_sentinels = ['localhost:26379'] {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.625245] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] cache.redis_server = localhost:6379 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.625410] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] cache.redis_socket_timeout = 1.0 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.625567] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] cache.redis_username = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.625729] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] cache.retry_attempts = 2 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.625890] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] cache.retry_delay = 0.0 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.626064] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] cache.socket_keepalive_count = 1 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.626227] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] cache.socket_keepalive_idle = 1 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.626385] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] cache.socket_keepalive_interval = 1 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.626571] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] cache.tls_allowed_ciphers = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.626735] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] cache.tls_cafile = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.626894] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] cache.tls_certfile = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.627068] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] cache.tls_enabled = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.627229] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] cache.tls_keyfile = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.627399] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] cinder.auth_section = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.627591] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] cinder.auth_type = password {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.627773] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] cinder.cafile = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.627945] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] cinder.catalog_info = volumev3::publicURL {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.628121] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] cinder.certfile = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.628286] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] cinder.collect_timing = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.628444] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] cinder.cross_az_attach = True {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.628605] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] cinder.debug = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.628767] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] cinder.endpoint_template = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.628942] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] cinder.http_retries = 3 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.629119] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] cinder.insecure = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.629280] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] cinder.keyfile = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.629451] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] cinder.os_region_name = RegionOne {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.629616] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] cinder.split_loggers = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.629777] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] cinder.timeout = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.629947] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] compute.consecutive_build_service_disable_threshold = 10 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.630120] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] compute.cpu_dedicated_set = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.630281] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] compute.cpu_shared_set = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.630452] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] compute.image_type_exclude_list = [] {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.630615] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] compute.live_migration_wait_for_vif_plug = True {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.630779] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] compute.max_concurrent_disk_ops = 0 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.630941] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] compute.max_disk_devices_to_attach = -1 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.631114] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] compute.packing_host_numa_cells_allocation_strategy = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.631288] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] compute.provider_config_location = /etc/nova/provider_config/ {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.631446] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] compute.resource_provider_association_refresh = 300 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.631605] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] compute.sharing_providers_max_uuids_per_request = 200 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.631768] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] compute.shutdown_retry_interval = 10 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.631945] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] compute.vmdk_allowed_types = ['streamOptimized', 'monolithicSparse'] {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.632133] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] conductor.workers = 2 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.632312] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] console.allowed_origins = [] {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.632474] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] console.ssl_ciphers = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.632644] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] console.ssl_minimum_version = default {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.632817] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] consoleauth.enforce_session_timeout = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.632984] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] consoleauth.token_ttl = 600 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.633171] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] cyborg.cafile = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.633330] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] cyborg.certfile = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.633494] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] cyborg.collect_timing = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.633654] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] cyborg.connect_retries = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.633813] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] cyborg.connect_retry_delay = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.633971] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] cyborg.endpoint_override = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.634143] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] cyborg.insecure = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.634302] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] cyborg.keyfile = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.634463] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] cyborg.max_version = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.634616] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] cyborg.min_version = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.634775] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] cyborg.region_name = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.634934] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] cyborg.retriable_status_codes = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.635097] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] cyborg.service_name = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.635267] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] cyborg.service_type = accelerator {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.635427] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] cyborg.split_loggers = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.635583] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] cyborg.status_code_retries = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.635740] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] cyborg.status_code_retry_delay = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.635895] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] cyborg.timeout = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.636083] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] cyborg.valid_interfaces = ['internal', 'public'] {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.636247] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] cyborg.version = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.636424] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] database.backend = sqlalchemy {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.636621] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] database.connection = **** {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.636791] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] database.connection_debug = 0 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.636960] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] database.connection_parameters = {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.637137] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] database.connection_recycle_time = 3600 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.637301] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] database.connection_trace = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.637459] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] database.db_inc_retry_interval = True {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.637645] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] database.db_max_retries = 20 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.637813] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] database.db_max_retry_interval = 10 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.637973] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] database.db_retry_interval = 1 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.638148] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] database.max_overflow = 50 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.638313] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] database.max_pool_size = 5 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.638474] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] database.max_retries = 10 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.638643] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] database.mysql_sql_mode = TRADITIONAL {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.638804] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] database.mysql_wsrep_sync_wait = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.638959] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] database.pool_timeout = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.639132] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] database.retry_interval = 10 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.639291] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] database.slave_connection = **** {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.639451] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] database.sqlite_synchronous = True {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.639608] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] database.use_db_reconnect = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.639787] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] api_database.backend = sqlalchemy {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.639954] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] api_database.connection = **** {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.640129] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] api_database.connection_debug = 0 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.640300] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] api_database.connection_parameters = {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.640462] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] api_database.connection_recycle_time = 3600 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.640622] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] api_database.connection_trace = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.640785] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] api_database.db_inc_retry_interval = True {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.640945] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] api_database.db_max_retries = 20 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.641119] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] api_database.db_max_retry_interval = 10 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.641284] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] api_database.db_retry_interval = 1 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.641445] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] api_database.max_overflow = 50 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.641606] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] api_database.max_pool_size = 5 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.641768] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] api_database.max_retries = 10 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.641934] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] api_database.mysql_sql_mode = TRADITIONAL {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.642101] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] api_database.mysql_wsrep_sync_wait = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.642263] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] api_database.pool_timeout = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.642424] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] api_database.retry_interval = 10 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.642582] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] api_database.slave_connection = **** {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.642743] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] api_database.sqlite_synchronous = True {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.642916] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] devices.enabled_mdev_types = [] {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.643100] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] ephemeral_storage_encryption.cipher = aes-xts-plain64 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.643270] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] ephemeral_storage_encryption.default_format = luks {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.643431] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] ephemeral_storage_encryption.enabled = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.643591] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] ephemeral_storage_encryption.key_size = 512 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.643762] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] glance.api_servers = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.643922] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] glance.cafile = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.644095] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] glance.certfile = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.644260] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] glance.collect_timing = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.644421] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] glance.connect_retries = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.644575] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] glance.connect_retry_delay = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.644737] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] glance.debug = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.644900] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] glance.default_trusted_certificate_ids = [] {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.645072] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] glance.enable_certificate_validation = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.645237] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] glance.enable_rbd_download = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.645393] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] glance.endpoint_override = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.645555] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] glance.insecure = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.645715] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] glance.keyfile = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.645872] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] glance.max_version = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.646035] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] glance.min_version = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.646199] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] glance.num_retries = 3 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.646366] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] glance.rbd_ceph_conf = {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.646545] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] glance.rbd_connect_timeout = 5 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.646731] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] glance.rbd_pool = {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.646901] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] glance.rbd_user = {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.647071] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] glance.region_name = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.647233] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] glance.retriable_status_codes = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.647387] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] glance.service_name = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.647584] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] glance.service_type = image {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.647776] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] glance.split_loggers = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.647936] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] glance.status_code_retries = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.648104] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] glance.status_code_retry_delay = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.648263] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] glance.timeout = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.648451] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] glance.valid_interfaces = ['internal', 'public'] {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.648619] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] glance.verify_glance_signatures = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.648777] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] glance.version = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.648941] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] guestfs.debug = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.649118] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] mks.enabled = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.649484] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] mks.mksproxy_base_url = http://127.0.0.1:6090/ {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.649676] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] image_cache.manager_interval = 2400 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.649849] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] image_cache.precache_concurrency = 1 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.650028] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] image_cache.remove_unused_base_images = True {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.650205] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] image_cache.remove_unused_original_minimum_age_seconds = 86400 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.650373] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] image_cache.remove_unused_resized_minimum_age_seconds = 3600 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.650550] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] image_cache.subdirectory_name = _base {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.650728] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] ironic.api_max_retries = 60 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.650892] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] ironic.api_retry_interval = 2 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.651064] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] ironic.auth_section = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.651231] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] ironic.auth_type = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.651391] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] ironic.cafile = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.651550] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] ironic.certfile = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.651714] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] ironic.collect_timing = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.651874] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] ironic.conductor_group = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.652042] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] ironic.connect_retries = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.652206] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] ironic.connect_retry_delay = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.652361] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] ironic.endpoint_override = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.652522] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] ironic.insecure = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.652680] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] ironic.keyfile = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.652837] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] ironic.max_version = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.652992] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] ironic.min_version = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.653171] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] ironic.peer_list = [] {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.653331] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] ironic.region_name = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.653488] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] ironic.retriable_status_codes = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.653650] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] ironic.serial_console_state_timeout = 10 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.653811] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] ironic.service_name = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.653983] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] ironic.service_type = baremetal {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.654154] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] ironic.shard = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.654315] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] ironic.split_loggers = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.654473] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] ironic.status_code_retries = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.654628] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] ironic.status_code_retry_delay = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.654788] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] ironic.timeout = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.654967] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] ironic.valid_interfaces = ['internal', 'public'] {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.655140] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] ironic.version = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.655325] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] key_manager.backend = nova.keymgr.conf_key_mgr.ConfKeyManager {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.655497] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] key_manager.fixed_key = **** {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.655679] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] barbican.auth_endpoint = http://localhost/identity/v3 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.655845] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] barbican.barbican_api_version = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.656017] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] barbican.barbican_endpoint = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.656185] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] barbican.barbican_endpoint_type = public {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.656342] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] barbican.barbican_region_name = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.656502] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] barbican.cafile = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.656692] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] barbican.certfile = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.656864] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] barbican.collect_timing = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.657034] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] barbican.insecure = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.657197] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] barbican.keyfile = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.657361] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] barbican.number_of_retries = 60 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.657540] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] barbican.retry_delay = 1 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.657738] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] barbican.send_service_user_token = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.657907] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] barbican.split_loggers = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.658079] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] barbican.timeout = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.658244] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] barbican.verify_ssl = True {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.658403] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] barbican.verify_ssl_path = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.658570] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] barbican_service_user.auth_section = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.658734] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] barbican_service_user.auth_type = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.658892] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] barbican_service_user.cafile = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.659059] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] barbican_service_user.certfile = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.659224] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] barbican_service_user.collect_timing = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.659382] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] barbican_service_user.insecure = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.659538] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] barbican_service_user.keyfile = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.659701] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] barbican_service_user.split_loggers = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.659857] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] barbican_service_user.timeout = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.660031] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] vault.approle_role_id = **** {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.660193] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] vault.approle_secret_id = **** {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.660365] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] vault.kv_mountpoint = secret {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.660528] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] vault.kv_path = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.660694] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] vault.kv_version = 2 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.660851] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] vault.namespace = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.661016] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] vault.root_token_id = **** {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.661178] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] vault.ssl_ca_crt_file = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.661350] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] vault.timeout = 60.0 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.661513] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] vault.use_ssl = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.661683] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] vault.vault_url = http://127.0.0.1:8200 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.661861] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] keystone.auth_section = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.662034] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] keystone.auth_type = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.662198] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] keystone.cafile = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.662357] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] keystone.certfile = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.662520] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] keystone.collect_timing = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.662680] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] keystone.connect_retries = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.662838] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] keystone.connect_retry_delay = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.662996] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] keystone.endpoint_override = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.663170] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] keystone.insecure = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.663330] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] keystone.keyfile = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.663488] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] keystone.max_version = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.663645] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] keystone.min_version = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.663808] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] keystone.region_name = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.663966] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] keystone.retriable_status_codes = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.664140] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] keystone.service_name = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.664309] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] keystone.service_type = identity {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.664473] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] keystone.split_loggers = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.664634] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] keystone.status_code_retries = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.664796] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] keystone.status_code_retry_delay = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.664973] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] keystone.timeout = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.665167] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] keystone.valid_interfaces = ['internal', 'public'] {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.665332] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] keystone.version = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.665531] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] libvirt.connection_uri = {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.665695] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] libvirt.cpu_mode = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.665859] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] libvirt.cpu_model_extra_flags = [] {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.666034] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] libvirt.cpu_models = [] {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.666209] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] libvirt.cpu_power_governor_high = performance {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.666383] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] libvirt.cpu_power_governor_low = powersave {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.666565] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] libvirt.cpu_power_management = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.666750] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] libvirt.cpu_power_management_strategy = cpu_state {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.666915] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] libvirt.device_detach_attempts = 8 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.667091] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] libvirt.device_detach_timeout = 20 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.667258] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] libvirt.disk_cachemodes = [] {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.667415] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] libvirt.disk_prefix = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.667611] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] libvirt.enabled_perf_events = [] {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.667796] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] libvirt.file_backed_memory = 0 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.667969] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] libvirt.gid_maps = [] {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.668139] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] libvirt.hw_disk_discard = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.668299] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] libvirt.hw_machine_type = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.668466] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] libvirt.images_rbd_ceph_conf = {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.668628] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] libvirt.images_rbd_glance_copy_poll_interval = 15 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.668791] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] libvirt.images_rbd_glance_copy_timeout = 600 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.668960] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] libvirt.images_rbd_glance_store_name = {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.669149] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] libvirt.images_rbd_pool = rbd {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.669319] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] libvirt.images_type = default {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.669477] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] libvirt.images_volume_group = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.669660] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] libvirt.inject_key = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.669888] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] libvirt.inject_partition = -2 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.670076] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] libvirt.inject_password = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.670247] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] libvirt.iscsi_iface = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.670410] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] libvirt.iser_use_multipath = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.670574] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] libvirt.live_migration_bandwidth = 0 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.670738] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] libvirt.live_migration_completion_timeout = 800 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.670901] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] libvirt.live_migration_downtime = 500 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.671074] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] libvirt.live_migration_downtime_delay = 75 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.671241] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] libvirt.live_migration_downtime_steps = 10 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.671403] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] libvirt.live_migration_inbound_addr = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.671566] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] libvirt.live_migration_permit_auto_converge = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.671728] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] libvirt.live_migration_permit_post_copy = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.671885] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] libvirt.live_migration_scheme = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.672068] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] libvirt.live_migration_timeout_action = abort {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.672240] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] libvirt.live_migration_tunnelled = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.672404] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] libvirt.live_migration_uri = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.672570] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] libvirt.live_migration_with_native_tls = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.672797] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] libvirt.max_queues = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.672988] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] libvirt.mem_stats_period_seconds = 10 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.673254] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] libvirt.migration_inbound_addr = 10.180.1.21 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.673423] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] libvirt.nfs_mount_options = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.674662] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] libvirt.nfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.674853] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] libvirt.num_aoe_discover_tries = 3 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.675041] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] libvirt.num_iser_scan_tries = 5 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.675216] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] libvirt.num_memory_encrypted_guests = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.675387] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] libvirt.num_nvme_discover_tries = 5 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.675556] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] libvirt.num_pcie_ports = 0 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.675774] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] libvirt.num_volume_scan_tries = 5 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.675991] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] libvirt.pmem_namespaces = [] {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.676179] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] libvirt.quobyte_client_cfg = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.676481] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] libvirt.quobyte_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.676692] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] libvirt.rbd_connect_timeout = 5 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.676872] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] libvirt.rbd_destroy_volume_retries = 12 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.677054] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] libvirt.rbd_destroy_volume_retry_interval = 5 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.677223] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] libvirt.rbd_secret_uuid = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.677387] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] libvirt.rbd_user = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.677577] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] libvirt.realtime_scheduler_priority = 1 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.677778] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] libvirt.remote_filesystem_transport = ssh {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.677947] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] libvirt.rescue_image_id = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.678124] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] libvirt.rescue_kernel_id = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.678284] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] libvirt.rescue_ramdisk_id = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.678456] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] libvirt.rng_dev_path = /dev/urandom {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.678681] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] libvirt.rx_queue_size = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.678884] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] libvirt.smbfs_mount_options = {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.679183] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] libvirt.smbfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.679359] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] libvirt.snapshot_compression = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.679524] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] libvirt.snapshot_image_format = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.679748] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] libvirt.snapshots_directory = /opt/stack/data/nova/instances/snapshots {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.679915] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] libvirt.sparse_logical_volumes = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.680092] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] libvirt.swtpm_enabled = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.680266] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] libvirt.swtpm_group = tss {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.680434] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] libvirt.swtpm_user = tss {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.680604] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] libvirt.sysinfo_serial = unique {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.680766] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] libvirt.tb_cache_size = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.680926] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] libvirt.tx_queue_size = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.681105] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] libvirt.uid_maps = [] {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.681272] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] libvirt.use_virtio_for_bridges = True {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.681442] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] libvirt.virt_type = kvm {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.681613] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] libvirt.volume_clear = zero {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.681782] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] libvirt.volume_clear_size = 0 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.681949] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] libvirt.volume_use_multipath = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.682121] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] libvirt.vzstorage_cache_path = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.682294] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] libvirt.vzstorage_log_path = /var/log/vstorage/%(cluster_name)s/nova.log.gz {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.682463] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] libvirt.vzstorage_mount_group = qemu {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.682625] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] libvirt.vzstorage_mount_opts = [] {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.682795] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] libvirt.vzstorage_mount_perms = 0770 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.683083] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] libvirt.vzstorage_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.683264] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] libvirt.vzstorage_mount_user = stack {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.683431] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] libvirt.wait_soft_reboot_seconds = 120 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.683605] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] neutron.auth_section = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.683779] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] neutron.auth_type = password {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.683941] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] neutron.cafile = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.684114] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] neutron.certfile = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.684282] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] neutron.collect_timing = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.684459] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] neutron.connect_retries = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.684635] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] neutron.connect_retry_delay = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.684839] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] neutron.default_floating_pool = public {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.684969] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] neutron.endpoint_override = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.685147] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] neutron.extension_sync_interval = 600 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.685310] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] neutron.http_retries = 3 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.685470] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] neutron.insecure = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.685629] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] neutron.keyfile = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.685788] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] neutron.max_version = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.685957] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] neutron.metadata_proxy_shared_secret = **** {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.686130] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] neutron.min_version = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.686302] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] neutron.ovs_bridge = br-int {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.686468] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] neutron.physnets = [] {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.686677] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] neutron.region_name = RegionOne {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.686842] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] neutron.retriable_status_codes = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.687022] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] neutron.service_metadata_proxy = True {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.687189] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] neutron.service_name = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.687358] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] neutron.service_type = network {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.687537] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] neutron.split_loggers = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.687735] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] neutron.status_code_retries = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.687908] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] neutron.status_code_retry_delay = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.688082] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] neutron.timeout = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.688267] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] neutron.valid_interfaces = ['internal', 'public'] {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.688429] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] neutron.version = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.688632] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] notifications.bdms_in_notifications = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.688823] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] notifications.default_level = INFO {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.688997] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] notifications.notification_format = unversioned {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.689178] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] notifications.notify_on_state_change = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.689354] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] notifications.versioned_notifications_topics = ['versioned_notifications'] {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.689529] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] pci.alias = [] {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.689701] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] pci.device_spec = [] {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.689865] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] pci.report_in_placement = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.690048] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] placement.auth_section = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.690223] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] placement.auth_type = password {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.690393] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] placement.auth_url = http://10.180.1.21/identity {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.690553] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] placement.cafile = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.690712] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] placement.certfile = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.690874] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] placement.collect_timing = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.691041] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] placement.connect_retries = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.691203] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] placement.connect_retry_delay = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.691359] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] placement.default_domain_id = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.691515] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] placement.default_domain_name = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.691673] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] placement.domain_id = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.691830] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] placement.domain_name = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.691985] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] placement.endpoint_override = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.692158] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] placement.insecure = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.692315] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] placement.keyfile = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.692470] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] placement.max_version = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.692627] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] placement.min_version = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.692795] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] placement.password = **** {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.692980] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] placement.project_domain_id = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.693178] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] placement.project_domain_name = Default {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.693349] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] placement.project_id = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.693520] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] placement.project_name = service {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.693692] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] placement.region_name = RegionOne {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.693853] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] placement.retriable_status_codes = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.694018] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] placement.service_name = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.694191] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] placement.service_type = placement {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.694354] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] placement.split_loggers = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.694515] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] placement.status_code_retries = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.694680] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] placement.status_code_retry_delay = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.694839] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] placement.system_scope = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.694997] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] placement.timeout = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.695168] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] placement.trust_id = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.695326] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] placement.user_domain_id = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.695495] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] placement.user_domain_name = Default {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.695655] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] placement.user_id = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.695827] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] placement.username = nova {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.696021] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] placement.valid_interfaces = ['internal', 'public'] {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.696185] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] placement.version = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.696364] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] quota.cores = 20 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.696550] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] quota.count_usage_from_placement = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.696722] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] quota.driver = nova.quota.DbQuotaDriver {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.696897] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] quota.injected_file_content_bytes = 10240 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.697076] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] quota.injected_file_path_length = 255 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.697246] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] quota.injected_files = 5 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.697411] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] quota.instances = 10 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.697580] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] quota.key_pairs = 100 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.697743] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] quota.metadata_items = 128 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.697908] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] quota.ram = 51200 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.698081] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] quota.recheck_quota = True {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.698252] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] quota.server_group_members = 10 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.698417] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] quota.server_groups = 10 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.698587] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] scheduler.discover_hosts_in_cells_interval = -1 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.698753] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] scheduler.enable_isolated_aggregate_filtering = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.698918] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] scheduler.image_metadata_prefilter = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.699090] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] scheduler.limit_tenants_to_placement_aggregate = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.699256] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] scheduler.max_attempts = 3 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.699418] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] scheduler.max_placement_results = 1000 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.699579] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] scheduler.placement_aggregate_required_for_tenants = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.699752] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] scheduler.query_placement_for_image_type_support = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.699915] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] scheduler.query_placement_for_routed_network_aggregates = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.700099] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] scheduler.workers = 2 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.700283] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] filter_scheduler.aggregate_image_properties_isolation_namespace = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.700456] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] filter_scheduler.aggregate_image_properties_isolation_separator = . {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.700667] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] filter_scheduler.available_filters = ['nova.scheduler.filters.all_filters'] {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.700854] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] filter_scheduler.build_failure_weight_multiplier = 1000000.0 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.701034] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] filter_scheduler.cpu_weight_multiplier = 1.0 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.701205] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] filter_scheduler.cross_cell_move_weight_multiplier = 1000000.0 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.701370] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] filter_scheduler.disk_weight_multiplier = 1.0 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.701558] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] filter_scheduler.enabled_filters = ['ComputeFilter', 'ComputeCapabilitiesFilter', 'ImagePropertiesFilter', 'ServerGroupAntiAffinityFilter', 'ServerGroupAffinityFilter', 'SameHostFilter', 'DifferentHostFilter'] {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.701728] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] filter_scheduler.host_subset_size = 1 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.701893] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] filter_scheduler.hypervisor_version_weight_multiplier = 1.0 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.702063] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] filter_scheduler.image_properties_default_architecture = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.702227] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] filter_scheduler.io_ops_weight_multiplier = -1.0 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.702391] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] filter_scheduler.isolated_hosts = [] {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.702558] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] filter_scheduler.isolated_images = [] {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.702722] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] filter_scheduler.max_instances_per_host = 50 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.702881] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] filter_scheduler.max_io_ops_per_host = 8 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.703053] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] filter_scheduler.num_instances_weight_multiplier = 0.0 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.703217] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] filter_scheduler.pci_in_placement = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.703377] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] filter_scheduler.pci_weight_multiplier = 1.0 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.703536] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] filter_scheduler.ram_weight_multiplier = 1.0 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.703704] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] filter_scheduler.restrict_isolated_hosts_to_isolated_images = True {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.703867] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] filter_scheduler.shuffle_best_same_weighed_hosts = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.704038] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] filter_scheduler.soft_affinity_weight_multiplier = 1.0 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.704204] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] filter_scheduler.soft_anti_affinity_weight_multiplier = 1.0 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.704367] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] filter_scheduler.track_instance_changes = True {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.704565] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] filter_scheduler.weight_classes = ['nova.scheduler.weights.all_weighers'] {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.704742] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] metrics.required = True {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.704908] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] metrics.weight_multiplier = 1.0 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.705085] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] metrics.weight_of_unavailable = -10000.0 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.705250] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] metrics.weight_setting = [] {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.705570] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] serial_console.base_url = ws://127.0.0.1:6083/ {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.705749] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] serial_console.enabled = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.705925] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] serial_console.port_range = 10000:20000 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.706105] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] serial_console.proxyclient_address = 127.0.0.1 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.706277] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] serial_console.serialproxy_host = 0.0.0.0 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.706445] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] serial_console.serialproxy_port = 6083 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.706638] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] service_user.auth_section = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.706823] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] service_user.auth_type = password {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.706986] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] service_user.cafile = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.707159] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] service_user.certfile = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.707325] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] service_user.collect_timing = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.707505] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] service_user.insecure = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.707643] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] service_user.keyfile = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.707815] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] service_user.send_service_user_token = True {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.707977] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] service_user.split_loggers = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.708148] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] service_user.timeout = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.708316] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] spice.agent_enabled = True {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.708478] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] spice.enabled = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.708814] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] spice.html5proxy_base_url = http://127.0.0.1:6082/spice_auto.html {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.709044] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] spice.html5proxy_host = 0.0.0.0 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.709226] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] spice.html5proxy_port = 6082 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.709391] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] spice.image_compression = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.709553] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] spice.jpeg_compression = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.709715] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] spice.playback_compression = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.709882] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] spice.server_listen = 127.0.0.1 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.710061] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] spice.server_proxyclient_address = 127.0.0.1 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.710223] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] spice.streaming_mode = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.710383] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] spice.zlib_compression = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.710548] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] upgrade_levels.baseapi = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.710721] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] upgrade_levels.compute = auto {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.710880] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] upgrade_levels.conductor = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.711047] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] upgrade_levels.scheduler = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.711215] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] vendordata_dynamic_auth.auth_section = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.711378] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] vendordata_dynamic_auth.auth_type = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.711537] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] vendordata_dynamic_auth.cafile = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.711697] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] vendordata_dynamic_auth.certfile = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.711857] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] vendordata_dynamic_auth.collect_timing = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.712026] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] vendordata_dynamic_auth.insecure = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.712186] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] vendordata_dynamic_auth.keyfile = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.712348] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] vendordata_dynamic_auth.split_loggers = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.712503] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] vendordata_dynamic_auth.timeout = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.712708] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] vmware.api_retry_count = 10 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.712881] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] vmware.ca_file = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.713101] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] vmware.cache_prefix = devstack-image-cache {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.713279] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] vmware.cluster_name = testcl1 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.713447] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] vmware.connection_pool_size = 10 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.713605] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] vmware.console_delay_seconds = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.713773] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] vmware.datastore_regex = ^datastore.* {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.713986] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] vmware.host_ip = vc1.osci.c.eu-de-1.cloud.sap {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.714171] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] vmware.host_password = **** {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.714336] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] vmware.host_port = 443 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.714504] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] vmware.host_username = administrator@vsphere.local {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.714670] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] vmware.insecure = True {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.714830] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] vmware.integration_bridge = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.714990] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] vmware.maximum_objects = 100 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.715160] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] vmware.pbm_default_policy = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.715320] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] vmware.pbm_enabled = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.715473] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] vmware.pbm_wsdl_location = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.715637] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] vmware.serial_log_dir = /opt/vmware/vspc {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.715794] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] vmware.serial_port_proxy_uri = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.715947] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] vmware.serial_port_service_uri = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.716122] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] vmware.task_poll_interval = 0.5 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.716293] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] vmware.use_linked_clone = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.716456] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] vmware.vnc_keymap = en-us {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.716652] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] vmware.vnc_port = 5900 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.716821] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] vmware.vnc_port_total = 10000 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.717019] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] vnc.auth_schemes = ['none'] {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.717194] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] vnc.enabled = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.717504] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] vnc.novncproxy_base_url = http://127.0.0.1:6080/vnc_auto.html {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.717687] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] vnc.novncproxy_host = 0.0.0.0 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.717858] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] vnc.novncproxy_port = 6080 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.718044] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] vnc.server_listen = 127.0.0.1 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.718221] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] vnc.server_proxyclient_address = 127.0.0.1 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.718381] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] vnc.vencrypt_ca_certs = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.718538] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] vnc.vencrypt_client_cert = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.718696] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] vnc.vencrypt_client_key = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.718890] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] workarounds.disable_compute_service_check_for_ffu = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.719070] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] workarounds.disable_deep_image_inspection = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.719237] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] workarounds.disable_fallback_pcpu_query = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.719398] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] workarounds.disable_group_policy_check_upcall = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.719556] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] workarounds.disable_libvirt_livesnapshot = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.719716] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] workarounds.disable_rootwrap = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.719873] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] workarounds.enable_numa_live_migration = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.720040] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] workarounds.enable_qemu_monitor_announce_self = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.720206] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] workarounds.ensure_libvirt_rbd_instance_dir_cleanup = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.720361] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] workarounds.handle_virt_lifecycle_events = True {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.720519] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] workarounds.libvirt_disable_apic = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.720677] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] workarounds.never_download_image_if_on_rbd = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.720835] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] workarounds.qemu_monitor_announce_self_count = 3 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.720992] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] workarounds.qemu_monitor_announce_self_interval = 1 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.721163] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] workarounds.reserve_disk_resource_for_image_cache = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.721321] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] workarounds.skip_cpu_compare_at_startup = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.721477] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] workarounds.skip_cpu_compare_on_dest = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.721634] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] workarounds.skip_hypervisor_version_check_on_lm = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.721791] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] workarounds.skip_reserve_in_use_ironic_nodes = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.721945] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] workarounds.unified_limits_count_pcpu_as_vcpu = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.722120] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] workarounds.wait_for_vif_plugged_event_during_hard_reboot = [] {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.722306] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] wsgi.api_paste_config = /etc/nova/api-paste.ini {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.722475] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] wsgi.client_socket_timeout = 900 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.722641] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] wsgi.default_pool_size = 1000 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.722808] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] wsgi.keep_alive = True {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.722974] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] wsgi.max_header_line = 16384 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.723148] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] wsgi.secure_proxy_ssl_header = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.723309] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] wsgi.ssl_ca_file = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.723466] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] wsgi.ssl_cert_file = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.723623] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] wsgi.ssl_key_file = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.723788] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] wsgi.tcp_keepidle = 600 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.723964] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] wsgi.wsgi_log_format = %(client_ip)s "%(request_line)s" status: %(status_code)s len: %(body_length)s time: %(wall_seconds).7f {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.724143] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] zvm.ca_file = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.724304] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] zvm.cloud_connector_url = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.724634] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] zvm.image_tmp_path = /opt/stack/data/n-cpu-1/images {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.724816] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] zvm.reachable_timeout = 300 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.725029] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] oslo_policy.enforce_new_defaults = True {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.725436] env[61957]: WARNING oslo_config.cfg [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] Deprecated: Option "enforce_scope" from group "oslo_policy" is deprecated for removal (This configuration was added temporarily to facilitate a smooth transition to the new RBAC. OpenStack will always enforce scope checks. This configuration option is deprecated and will be removed in the 2025.2 cycle.). Its value may be silently ignored in the future. [ 516.725622] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] oslo_policy.enforce_scope = True {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.725804] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] oslo_policy.policy_default_rule = default {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.725988] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] oslo_policy.policy_dirs = ['policy.d'] {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.726182] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] oslo_policy.policy_file = policy.yaml {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.726356] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] oslo_policy.remote_content_type = application/x-www-form-urlencoded {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.726557] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] oslo_policy.remote_ssl_ca_crt_file = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.726708] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] oslo_policy.remote_ssl_client_crt_file = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.726873] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] oslo_policy.remote_ssl_client_key_file = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.727047] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] oslo_policy.remote_ssl_verify_server_crt = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.727220] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] oslo_versionedobjects.fatal_exception_format_errors = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.727393] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] oslo_middleware.http_basic_auth_user_file = /etc/htpasswd {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.727566] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] profiler.connection_string = messaging:// {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.727733] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] profiler.enabled = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.727900] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] profiler.es_doc_type = notification {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.728071] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] profiler.es_scroll_size = 10000 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.728240] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] profiler.es_scroll_time = 2m {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.728401] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] profiler.filter_error_trace = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.728566] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] profiler.hmac_keys = **** {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.728732] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] profiler.sentinel_service_name = mymaster {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.728920] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] profiler.socket_timeout = 0.1 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.729112] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] profiler.trace_requests = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.729278] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] profiler.trace_sqlalchemy = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.729454] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] profiler_jaeger.process_tags = {} {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.729613] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] profiler_jaeger.service_name_prefix = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.729778] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] profiler_otlp.service_name_prefix = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.729942] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] remote_debug.host = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.730110] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] remote_debug.port = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.730291] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] oslo_messaging_rabbit.amqp_auto_delete = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.730453] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] oslo_messaging_rabbit.amqp_durable_queues = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.730616] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] oslo_messaging_rabbit.conn_pool_min_size = 2 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.730778] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] oslo_messaging_rabbit.conn_pool_ttl = 1200 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.730938] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] oslo_messaging_rabbit.direct_mandatory_flag = True {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.731110] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] oslo_messaging_rabbit.enable_cancel_on_failover = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.731273] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] oslo_messaging_rabbit.heartbeat_in_pthread = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.731435] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] oslo_messaging_rabbit.heartbeat_rate = 3 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.731597] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] oslo_messaging_rabbit.heartbeat_timeout_threshold = 60 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.731766] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] oslo_messaging_rabbit.hostname = devstack {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.731924] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] oslo_messaging_rabbit.kombu_compression = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.732104] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] oslo_messaging_rabbit.kombu_failover_strategy = round-robin {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.732272] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] oslo_messaging_rabbit.kombu_missing_consumer_retry_timeout = 60 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.732440] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] oslo_messaging_rabbit.kombu_reconnect_delay = 1.0 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.732613] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] oslo_messaging_rabbit.processname = nova-compute {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.732778] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] oslo_messaging_rabbit.rabbit_ha_queues = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.732942] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] oslo_messaging_rabbit.rabbit_interval_max = 30 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.733129] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] oslo_messaging_rabbit.rabbit_login_method = AMQPLAIN {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.733293] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] oslo_messaging_rabbit.rabbit_qos_prefetch_count = 0 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.733455] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] oslo_messaging_rabbit.rabbit_quorum_delivery_limit = 0 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.733617] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_bytes = 0 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.733781] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_length = 0 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.733943] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] oslo_messaging_rabbit.rabbit_quorum_queue = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.734121] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] oslo_messaging_rabbit.rabbit_retry_backoff = 2 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.734284] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] oslo_messaging_rabbit.rabbit_retry_interval = 1 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.734467] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] oslo_messaging_rabbit.rabbit_stream_fanout = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.734650] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] oslo_messaging_rabbit.rabbit_transient_queues_ttl = 1800 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.734818] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] oslo_messaging_rabbit.rabbit_transient_quorum_queue = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.734987] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] oslo_messaging_rabbit.rpc_conn_pool_size = 30 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.735168] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] oslo_messaging_rabbit.ssl = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.735344] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] oslo_messaging_rabbit.ssl_ca_file = {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.735514] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] oslo_messaging_rabbit.ssl_cert_file = {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.735681] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] oslo_messaging_rabbit.ssl_enforce_fips_mode = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.735850] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] oslo_messaging_rabbit.ssl_key_file = {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.736030] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] oslo_messaging_rabbit.ssl_version = {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.736198] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] oslo_messaging_rabbit.use_queue_manager = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.736386] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] oslo_messaging_notifications.driver = ['messagingv2'] {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.736602] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] oslo_messaging_notifications.retry = -1 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.736806] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] oslo_messaging_notifications.topics = ['notifications'] {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.736984] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] oslo_messaging_notifications.transport_url = **** {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.737235] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] oslo_limit.auth_section = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.737414] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] oslo_limit.auth_type = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.737580] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] oslo_limit.cafile = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.737744] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] oslo_limit.certfile = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.737906] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] oslo_limit.collect_timing = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.738077] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] oslo_limit.connect_retries = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.738243] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] oslo_limit.connect_retry_delay = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.738400] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] oslo_limit.endpoint_id = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.738558] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] oslo_limit.endpoint_override = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.738719] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] oslo_limit.insecure = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.738876] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] oslo_limit.keyfile = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.739043] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] oslo_limit.max_version = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.739202] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] oslo_limit.min_version = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.739360] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] oslo_limit.region_name = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.739521] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] oslo_limit.retriable_status_codes = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.739678] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] oslo_limit.service_name = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.739834] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] oslo_limit.service_type = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.739992] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] oslo_limit.split_loggers = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.740161] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] oslo_limit.status_code_retries = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.740319] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] oslo_limit.status_code_retry_delay = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.740475] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] oslo_limit.timeout = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.740630] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] oslo_limit.valid_interfaces = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.740788] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] oslo_limit.version = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.740949] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] oslo_reports.file_event_handler = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.741127] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] oslo_reports.file_event_handler_interval = 1 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.741287] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] oslo_reports.log_dir = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.741457] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] vif_plug_linux_bridge_privileged.capabilities = [12] {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.741617] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] vif_plug_linux_bridge_privileged.group = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.741780] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] vif_plug_linux_bridge_privileged.helper_command = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.741947] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] vif_plug_linux_bridge_privileged.logger_name = oslo_privsep.daemon {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.742123] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] vif_plug_linux_bridge_privileged.thread_pool_size = 8 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.742281] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] vif_plug_linux_bridge_privileged.user = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.742450] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] vif_plug_ovs_privileged.capabilities = [12, 1] {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.742608] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] vif_plug_ovs_privileged.group = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.742769] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] vif_plug_ovs_privileged.helper_command = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.742933] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] vif_plug_ovs_privileged.logger_name = oslo_privsep.daemon {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.743109] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] vif_plug_ovs_privileged.thread_pool_size = 8 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.743271] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] vif_plug_ovs_privileged.user = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.743440] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] os_vif_linux_bridge.flat_interface = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.743618] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] os_vif_linux_bridge.forward_bridge_interface = ['all'] {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.743792] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] os_vif_linux_bridge.iptables_bottom_regex = {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.743962] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] os_vif_linux_bridge.iptables_drop_action = DROP {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.744147] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] os_vif_linux_bridge.iptables_top_regex = {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.744314] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] os_vif_linux_bridge.network_device_mtu = 1500 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.744503] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] os_vif_linux_bridge.use_ipv6 = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.744678] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] os_vif_linux_bridge.vlan_interface = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.744858] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] os_vif_ovs.default_qos_type = linux-noop {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.745038] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] os_vif_ovs.isolate_vif = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.745211] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] os_vif_ovs.network_device_mtu = 1500 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.745374] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] os_vif_ovs.ovs_vsctl_timeout = 120 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.745541] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] os_vif_ovs.ovsdb_connection = tcp:127.0.0.1:6640 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.745711] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] os_vif_ovs.ovsdb_interface = native {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.745871] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] os_vif_ovs.per_port_bridge = False {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.746046] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] os_brick.lock_path = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.746213] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] os_brick.wait_mpath_device_attempts = 4 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.746373] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] os_brick.wait_mpath_device_interval = 1 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.746573] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] privsep_osbrick.capabilities = [21] {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.746733] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] privsep_osbrick.group = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.746892] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] privsep_osbrick.helper_command = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.747067] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] privsep_osbrick.logger_name = os_brick.privileged {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.747233] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] privsep_osbrick.thread_pool_size = 8 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.747390] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] privsep_osbrick.user = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.747570] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] nova_sys_admin.capabilities = [0, 1, 2, 3, 12, 21] {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.747727] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] nova_sys_admin.group = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.747884] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] nova_sys_admin.helper_command = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.748058] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] nova_sys_admin.logger_name = oslo_privsep.daemon {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.748224] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] nova_sys_admin.thread_pool_size = 8 {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.748382] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] nova_sys_admin.user = None {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 516.748528] env[61957]: DEBUG oslo_service.service [None req-dbbbc43a-9753-4e75-9c02-b34494d62434 None None] ******************************************************************************** {{(pid=61957) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2830}} [ 516.749075] env[61957]: INFO nova.service [-] Starting compute node (version 0.1.0) [ 517.252789] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-d6d909f4-2ea9-412a-a880-cbc293fc0fb1 None None] Getting list of instances from cluster (obj){ [ 517.252789] env[61957]: value = "domain-c8" [ 517.252789] env[61957]: _type = "ClusterComputeResource" [ 517.252789] env[61957]: } {{(pid=61957) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2122}} [ 517.253981] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fa1309b-6ee5-4cdc-bcb4-059655a4ae2f {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 517.265559] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-d6d909f4-2ea9-412a-a880-cbc293fc0fb1 None None] Got total of 0 instances {{(pid=61957) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2131}} [ 517.266318] env[61957]: WARNING nova.virt.vmwareapi.driver [None req-d6d909f4-2ea9-412a-a880-cbc293fc0fb1 None None] The vmwareapi driver is not tested by the OpenStack project nor does it have clear maintainer(s) and thus its quality can not be ensured. It should be considered experimental and may be removed in a future release. If you are using the driver in production please let us know via the openstack-discuss mailing list. [ 517.266939] env[61957]: INFO nova.virt.node [None req-d6d909f4-2ea9-412a-a880-cbc293fc0fb1 None None] Generated node identity 0ceb6c9e-61c6-496d-8579-9d32627e96da [ 517.267255] env[61957]: INFO nova.virt.node [None req-d6d909f4-2ea9-412a-a880-cbc293fc0fb1 None None] Wrote node identity 0ceb6c9e-61c6-496d-8579-9d32627e96da to /opt/stack/data/n-cpu-1/compute_id [ 517.770609] env[61957]: WARNING nova.compute.manager [None req-d6d909f4-2ea9-412a-a880-cbc293fc0fb1 None None] Compute nodes ['0ceb6c9e-61c6-496d-8579-9d32627e96da'] for host cpu-1 were not found in the database. If this is the first time this service is starting on this host, then you can ignore this warning. [ 518.777901] env[61957]: INFO nova.compute.manager [None req-d6d909f4-2ea9-412a-a880-cbc293fc0fb1 None None] Looking for unclaimed instances stuck in BUILDING status for nodes managed by this host [ 519.784151] env[61957]: WARNING nova.compute.manager [None req-d6d909f4-2ea9-412a-a880-cbc293fc0fb1 None None] No compute node record found for host cpu-1. If this is the first time this service is starting on this host, then you can ignore this warning.: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host cpu-1 could not be found. [ 519.784514] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d6d909f4-2ea9-412a-a880-cbc293fc0fb1 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 519.784698] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d6d909f4-2ea9-412a-a880-cbc293fc0fb1 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 519.784858] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d6d909f4-2ea9-412a-a880-cbc293fc0fb1 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 519.785032] env[61957]: DEBUG nova.compute.resource_tracker [None req-d6d909f4-2ea9-412a-a880-cbc293fc0fb1 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61957) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 519.785935] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea5509ec-1169-4360-984a-998105dfbb26 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 519.794495] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c92da63-a1cc-4ed7-abc5-72eca16d124a {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 519.808078] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ad39c9a-269a-4e2c-baf7-9d17a9af7a68 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 519.814503] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6967f3e-7a5e-4b89-bae1-87d24144c250 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 519.843173] env[61957]: DEBUG nova.compute.resource_tracker [None req-d6d909f4-2ea9-412a-a880-cbc293fc0fb1 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181484MB free_disk=142GB free_vcpus=48 pci_devices=None {{(pid=61957) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 519.843395] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d6d909f4-2ea9-412a-a880-cbc293fc0fb1 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 519.843652] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d6d909f4-2ea9-412a-a880-cbc293fc0fb1 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 520.346461] env[61957]: WARNING nova.compute.resource_tracker [None req-d6d909f4-2ea9-412a-a880-cbc293fc0fb1 None None] No compute node record for cpu-1:0ceb6c9e-61c6-496d-8579-9d32627e96da: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host 0ceb6c9e-61c6-496d-8579-9d32627e96da could not be found. [ 520.850393] env[61957]: INFO nova.compute.resource_tracker [None req-d6d909f4-2ea9-412a-a880-cbc293fc0fb1 None None] Compute node record created for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 with uuid: 0ceb6c9e-61c6-496d-8579-9d32627e96da [ 522.358139] env[61957]: DEBUG nova.compute.resource_tracker [None req-d6d909f4-2ea9-412a-a880-cbc293fc0fb1 None None] Total usable vcpus: 48, total allocated vcpus: 0 {{(pid=61957) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 522.358477] env[61957]: DEBUG nova.compute.resource_tracker [None req-d6d909f4-2ea9-412a-a880-cbc293fc0fb1 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=512MB phys_disk=200GB used_disk=0GB total_vcpus=48 used_vcpus=0 pci_stats=[] {{(pid=61957) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 522.515206] env[61957]: INFO nova.scheduler.client.report [None req-d6d909f4-2ea9-412a-a880-cbc293fc0fb1 None None] [req-b73ace98-30c8-457a-aa16-672ea48d35a1] Created resource provider record via placement API for resource provider with UUID 0ceb6c9e-61c6-496d-8579-9d32627e96da and name domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28. [ 522.531861] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56dd7d6c-d62c-4711-91e8-0ec7f9ce061e {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.539640] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd6a1fd8-5edd-4ef2-b9ae-84fa5dc65218 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.569556] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d14506c-d6b4-4684-ae1c-4d4b904baf51 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.576287] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-332169c2-e829-4cc8-a1a0-f229abde3edb {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.588736] env[61957]: DEBUG nova.compute.provider_tree [None req-d6d909f4-2ea9-412a-a880-cbc293fc0fb1 None None] Updating inventory in ProviderTree for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 523.122111] env[61957]: DEBUG nova.scheduler.client.report [None req-d6d909f4-2ea9-412a-a880-cbc293fc0fb1 None None] Updated inventory for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da with generation 0 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 523.122357] env[61957]: DEBUG nova.compute.provider_tree [None req-d6d909f4-2ea9-412a-a880-cbc293fc0fb1 None None] Updating resource provider 0ceb6c9e-61c6-496d-8579-9d32627e96da generation from 0 to 1 during operation: update_inventory {{(pid=61957) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 523.122496] env[61957]: DEBUG nova.compute.provider_tree [None req-d6d909f4-2ea9-412a-a880-cbc293fc0fb1 None None] Updating inventory in ProviderTree for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 523.175383] env[61957]: DEBUG nova.compute.provider_tree [None req-d6d909f4-2ea9-412a-a880-cbc293fc0fb1 None None] Updating resource provider 0ceb6c9e-61c6-496d-8579-9d32627e96da generation from 1 to 2 during operation: update_traits {{(pid=61957) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 523.680037] env[61957]: DEBUG nova.compute.resource_tracker [None req-d6d909f4-2ea9-412a-a880-cbc293fc0fb1 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61957) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 523.680451] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d6d909f4-2ea9-412a-a880-cbc293fc0fb1 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.837s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 523.680451] env[61957]: DEBUG nova.service [None req-d6d909f4-2ea9-412a-a880-cbc293fc0fb1 None None] Creating RPC server for service compute {{(pid=61957) start /opt/stack/nova/nova/service.py:186}} [ 523.696131] env[61957]: DEBUG nova.service [None req-d6d909f4-2ea9-412a-a880-cbc293fc0fb1 None None] Join ServiceGroup membership for this service compute {{(pid=61957) start /opt/stack/nova/nova/service.py:203}} [ 523.696301] env[61957]: DEBUG nova.servicegroup.drivers.db [None req-d6d909f4-2ea9-412a-a880-cbc293fc0fb1 None None] DB_Driver: join new ServiceGroup member cpu-1 to the compute group, service = {{(pid=61957) join /opt/stack/nova/nova/servicegroup/drivers/db.py:44}} [ 528.698752] env[61957]: DEBUG oslo_service.periodic_task [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Running periodic task ComputeManager._sync_power_states {{(pid=61957) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 529.207224] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Getting list of instances from cluster (obj){ [ 529.207224] env[61957]: value = "domain-c8" [ 529.207224] env[61957]: _type = "ClusterComputeResource" [ 529.207224] env[61957]: } {{(pid=61957) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2122}} [ 529.208429] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f64f134e-d6c6-4c60-a566-987eccbff7c2 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.217107] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Got total of 0 instances {{(pid=61957) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2131}} [ 529.217337] env[61957]: DEBUG oslo_service.periodic_task [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Running periodic task ComputeManager._cleanup_running_deleted_instances {{(pid=61957) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 529.217631] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Getting list of instances from cluster (obj){ [ 529.217631] env[61957]: value = "domain-c8" [ 529.217631] env[61957]: _type = "ClusterComputeResource" [ 529.217631] env[61957]: } {{(pid=61957) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2122}} [ 529.218474] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3aeeb1bb-4816-40a5-9b7e-23c32fe26b4d {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.225267] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Got total of 0 instances {{(pid=61957) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2131}} [ 563.132014] env[61957]: DEBUG oslo_concurrency.lockutils [None req-abb0f6e8-eaec-4ec9-a6d4-89648c99b1d8 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] Acquiring lock "1a0e117e-aa56-44fa-8818-71f0fd0b6183" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 563.132014] env[61957]: DEBUG oslo_concurrency.lockutils [None req-abb0f6e8-eaec-4ec9-a6d4-89648c99b1d8 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] Lock "1a0e117e-aa56-44fa-8818-71f0fd0b6183" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 563.637762] env[61957]: DEBUG nova.compute.manager [None req-abb0f6e8-eaec-4ec9-a6d4-89648c99b1d8 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] Starting instance... {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 563.771209] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d3a5f917-ae8e-4fd7-b4eb-cbb6e49d67de tempest-ServerDiagnosticsNegativeTest-1621277291 tempest-ServerDiagnosticsNegativeTest-1621277291-project-member] Acquiring lock "f28611df-9528-4ec0-b575-ad54ebd4a958" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 563.772467] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d3a5f917-ae8e-4fd7-b4eb-cbb6e49d67de tempest-ServerDiagnosticsNegativeTest-1621277291 tempest-ServerDiagnosticsNegativeTest-1621277291-project-member] Lock "f28611df-9528-4ec0-b575-ad54ebd4a958" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 564.088210] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d102ca19-392e-4051-87d1-b58bed2da29d tempest-ServersAdminNegativeTestJSON-555579 tempest-ServersAdminNegativeTestJSON-555579-project-member] Acquiring lock "37837096-5ad0-4b6f-b32d-f2899d88b6dd" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 564.088615] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d102ca19-392e-4051-87d1-b58bed2da29d tempest-ServersAdminNegativeTestJSON-555579 tempest-ServersAdminNegativeTestJSON-555579-project-member] Lock "37837096-5ad0-4b6f-b32d-f2899d88b6dd" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 564.194313] env[61957]: DEBUG oslo_concurrency.lockutils [None req-abb0f6e8-eaec-4ec9-a6d4-89648c99b1d8 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 564.194313] env[61957]: DEBUG oslo_concurrency.lockutils [None req-abb0f6e8-eaec-4ec9-a6d4-89648c99b1d8 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.002s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 564.194313] env[61957]: INFO nova.compute.claims [None req-abb0f6e8-eaec-4ec9-a6d4-89648c99b1d8 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 564.276561] env[61957]: DEBUG nova.compute.manager [None req-d3a5f917-ae8e-4fd7-b4eb-cbb6e49d67de tempest-ServerDiagnosticsNegativeTest-1621277291 tempest-ServerDiagnosticsNegativeTest-1621277291-project-member] [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] Starting instance... {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 564.593222] env[61957]: DEBUG nova.compute.manager [None req-d102ca19-392e-4051-87d1-b58bed2da29d tempest-ServersAdminNegativeTestJSON-555579 tempest-ServersAdminNegativeTestJSON-555579-project-member] [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] Starting instance... {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 564.802381] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d3a5f917-ae8e-4fd7-b4eb-cbb6e49d67de tempest-ServerDiagnosticsNegativeTest-1621277291 tempest-ServerDiagnosticsNegativeTest-1621277291-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 565.124930] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d102ca19-392e-4051-87d1-b58bed2da29d tempest-ServersAdminNegativeTestJSON-555579 tempest-ServersAdminNegativeTestJSON-555579-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 565.272355] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d83719b8-6bd9-4ae7-99ea-7d0970401ba5 tempest-ServerExternalEventsTest-1886420864 tempest-ServerExternalEventsTest-1886420864-project-member] Acquiring lock "abdb0ba9-abef-4528-96fb-ae7d7d41f8fc" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 565.272355] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d83719b8-6bd9-4ae7-99ea-7d0970401ba5 tempest-ServerExternalEventsTest-1886420864 tempest-ServerExternalEventsTest-1886420864-project-member] Lock "abdb0ba9-abef-4528-96fb-ae7d7d41f8fc" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 565.319221] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71f069c2-0c53-4f6a-b803-f81bc9453604 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.328421] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69fe5658-a744-400a-93f9-906caab86a1a {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.363613] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1762d63f-c71b-4731-bed0-ada0b16a24bd {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.373409] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-300c4c0a-98c6-4979-9e0e-5736ddbb53d5 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.388380] env[61957]: DEBUG nova.compute.provider_tree [None req-abb0f6e8-eaec-4ec9-a6d4-89648c99b1d8 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 565.774915] env[61957]: DEBUG nova.compute.manager [None req-d83719b8-6bd9-4ae7-99ea-7d0970401ba5 tempest-ServerExternalEventsTest-1886420864 tempest-ServerExternalEventsTest-1886420864-project-member] [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] Starting instance... {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 565.891918] env[61957]: DEBUG nova.scheduler.client.report [None req-abb0f6e8-eaec-4ec9-a6d4-89648c99b1d8 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 566.309019] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d83719b8-6bd9-4ae7-99ea-7d0970401ba5 tempest-ServerExternalEventsTest-1886420864 tempest-ServerExternalEventsTest-1886420864-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 566.398173] env[61957]: DEBUG oslo_concurrency.lockutils [None req-abb0f6e8-eaec-4ec9-a6d4-89648c99b1d8 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.209s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 566.399061] env[61957]: DEBUG nova.compute.manager [None req-abb0f6e8-eaec-4ec9-a6d4-89648c99b1d8 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] Start building networks asynchronously for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 566.402508] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d3a5f917-ae8e-4fd7-b4eb-cbb6e49d67de tempest-ServerDiagnosticsNegativeTest-1621277291 tempest-ServerDiagnosticsNegativeTest-1621277291-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.600s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 566.404027] env[61957]: INFO nova.compute.claims [None req-d3a5f917-ae8e-4fd7-b4eb-cbb6e49d67de tempest-ServerDiagnosticsNegativeTest-1621277291 tempest-ServerDiagnosticsNegativeTest-1621277291-project-member] [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 566.911331] env[61957]: DEBUG nova.compute.utils [None req-abb0f6e8-eaec-4ec9-a6d4-89648c99b1d8 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] Using /dev/sd instead of None {{(pid=61957) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 566.916757] env[61957]: DEBUG nova.compute.manager [None req-abb0f6e8-eaec-4ec9-a6d4-89648c99b1d8 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] Allocating IP information in the background. {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 566.919112] env[61957]: DEBUG nova.network.neutron [None req-abb0f6e8-eaec-4ec9-a6d4-89648c99b1d8 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] allocate_for_instance() {{(pid=61957) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 567.387314] env[61957]: DEBUG nova.policy [None req-abb0f6e8-eaec-4ec9-a6d4-89648c99b1d8 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'dbec97161e3249759c8743b15ea72873', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e4b7e4d62d074c1f9d66b40b328e5373', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61957) authorize /opt/stack/nova/nova/policy.py:203}} [ 567.415613] env[61957]: DEBUG nova.compute.manager [None req-abb0f6e8-eaec-4ec9-a6d4-89648c99b1d8 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] Start building block device mappings for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 567.561457] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-303eef81-b814-49ed-bd06-760bd8bf01ac {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.572641] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2119b902-c3ae-4edf-aa19-96303e37228a {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.610016] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d33c8be6-5f5e-41ee-aac6-011fb73f96c0 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.619790] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cacd8667-2056-4c21-ac4d-8ab782fa770c {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.638930] env[61957]: DEBUG nova.compute.provider_tree [None req-d3a5f917-ae8e-4fd7-b4eb-cbb6e49d67de tempest-ServerDiagnosticsNegativeTest-1621277291 tempest-ServerDiagnosticsNegativeTest-1621277291-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 568.143859] env[61957]: DEBUG nova.scheduler.client.report [None req-d3a5f917-ae8e-4fd7-b4eb-cbb6e49d67de tempest-ServerDiagnosticsNegativeTest-1621277291 tempest-ServerDiagnosticsNegativeTest-1621277291-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 568.429262] env[61957]: DEBUG nova.compute.manager [None req-abb0f6e8-eaec-4ec9-a6d4-89648c99b1d8 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] Start spawning the instance on the hypervisor. {{(pid=61957) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 568.468172] env[61957]: DEBUG nova.virt.hardware [None req-abb0f6e8-eaec-4ec9-a6d4-89648c99b1d8 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T17:22:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T17:21:50Z,direct_url=,disk_format='vmdk',id=11c76a2c-f705-470a-ba9d-4657858bab38,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='81d3096d0f094373913fc3dc8f5c3c6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T17:21:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 568.468172] env[61957]: DEBUG nova.virt.hardware [None req-abb0f6e8-eaec-4ec9-a6d4-89648c99b1d8 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] Flavor limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 568.468172] env[61957]: DEBUG nova.virt.hardware [None req-abb0f6e8-eaec-4ec9-a6d4-89648c99b1d8 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] Image limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 568.468788] env[61957]: DEBUG nova.virt.hardware [None req-abb0f6e8-eaec-4ec9-a6d4-89648c99b1d8 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] Flavor pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 568.470122] env[61957]: DEBUG nova.virt.hardware [None req-abb0f6e8-eaec-4ec9-a6d4-89648c99b1d8 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] Image pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 568.470369] env[61957]: DEBUG nova.virt.hardware [None req-abb0f6e8-eaec-4ec9-a6d4-89648c99b1d8 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 568.470623] env[61957]: DEBUG nova.virt.hardware [None req-abb0f6e8-eaec-4ec9-a6d4-89648c99b1d8 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 568.473017] env[61957]: DEBUG nova.virt.hardware [None req-abb0f6e8-eaec-4ec9-a6d4-89648c99b1d8 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 568.473017] env[61957]: DEBUG nova.virt.hardware [None req-abb0f6e8-eaec-4ec9-a6d4-89648c99b1d8 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] Got 1 possible topologies {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 568.473017] env[61957]: DEBUG nova.virt.hardware [None req-abb0f6e8-eaec-4ec9-a6d4-89648c99b1d8 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 568.473017] env[61957]: DEBUG nova.virt.hardware [None req-abb0f6e8-eaec-4ec9-a6d4-89648c99b1d8 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 568.474547] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4514b4de-50ad-47f7-8553-7d54c14d9db6 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 568.478581] env[61957]: DEBUG nova.network.neutron [None req-abb0f6e8-eaec-4ec9-a6d4-89648c99b1d8 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] Successfully created port: 6bf15563-9c78-4c42-9a95-c660f6c0e3b5 {{(pid=61957) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 568.489804] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0ce40c8-8536-49dd-8b47-19e38e9d08d1 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 568.515315] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a19120a-3a71-484f-966d-fbe39cbaec60 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 568.651390] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d3a5f917-ae8e-4fd7-b4eb-cbb6e49d67de tempest-ServerDiagnosticsNegativeTest-1621277291 tempest-ServerDiagnosticsNegativeTest-1621277291-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.247s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 568.651390] env[61957]: DEBUG nova.compute.manager [None req-d3a5f917-ae8e-4fd7-b4eb-cbb6e49d67de tempest-ServerDiagnosticsNegativeTest-1621277291 tempest-ServerDiagnosticsNegativeTest-1621277291-project-member] [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] Start building networks asynchronously for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 568.654731] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d102ca19-392e-4051-87d1-b58bed2da29d tempest-ServersAdminNegativeTestJSON-555579 tempest-ServersAdminNegativeTestJSON-555579-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.529s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 568.656288] env[61957]: INFO nova.compute.claims [None req-d102ca19-392e-4051-87d1-b58bed2da29d tempest-ServersAdminNegativeTestJSON-555579 tempest-ServersAdminNegativeTestJSON-555579-project-member] [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 569.161546] env[61957]: DEBUG nova.compute.utils [None req-d3a5f917-ae8e-4fd7-b4eb-cbb6e49d67de tempest-ServerDiagnosticsNegativeTest-1621277291 tempest-ServerDiagnosticsNegativeTest-1621277291-project-member] Using /dev/sd instead of None {{(pid=61957) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 569.165622] env[61957]: DEBUG nova.compute.manager [None req-d3a5f917-ae8e-4fd7-b4eb-cbb6e49d67de tempest-ServerDiagnosticsNegativeTest-1621277291 tempest-ServerDiagnosticsNegativeTest-1621277291-project-member] [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] Allocating IP information in the background. {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 569.165622] env[61957]: DEBUG nova.network.neutron [None req-d3a5f917-ae8e-4fd7-b4eb-cbb6e49d67de tempest-ServerDiagnosticsNegativeTest-1621277291 tempest-ServerDiagnosticsNegativeTest-1621277291-project-member] [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] allocate_for_instance() {{(pid=61957) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 569.412423] env[61957]: DEBUG nova.policy [None req-d3a5f917-ae8e-4fd7-b4eb-cbb6e49d67de tempest-ServerDiagnosticsNegativeTest-1621277291 tempest-ServerDiagnosticsNegativeTest-1621277291-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '38fb6691b032470198af03b1750f7925', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0ca03d3fd8a54aeca1fa9a95bbc7c3f7', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61957) authorize /opt/stack/nova/nova/policy.py:203}} [ 569.666429] env[61957]: DEBUG nova.compute.manager [None req-d3a5f917-ae8e-4fd7-b4eb-cbb6e49d67de tempest-ServerDiagnosticsNegativeTest-1621277291 tempest-ServerDiagnosticsNegativeTest-1621277291-project-member] [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] Start building block device mappings for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 569.784109] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46ac9556-2dc1-447f-854f-427094f6fd69 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.793790] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2788334-5b33-4bc0-8e83-f71b031bbe5f {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.825746] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25973140-fed1-442a-9fe9-639490649640 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.835175] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7042e3a9-35a9-41b2-a5b9-30b88e8112bd {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.852444] env[61957]: DEBUG nova.compute.provider_tree [None req-d102ca19-392e-4051-87d1-b58bed2da29d tempest-ServersAdminNegativeTestJSON-555579 tempest-ServersAdminNegativeTestJSON-555579-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 570.357676] env[61957]: DEBUG nova.scheduler.client.report [None req-d102ca19-392e-4051-87d1-b58bed2da29d tempest-ServersAdminNegativeTestJSON-555579 tempest-ServersAdminNegativeTestJSON-555579-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 570.681470] env[61957]: DEBUG nova.compute.manager [None req-d3a5f917-ae8e-4fd7-b4eb-cbb6e49d67de tempest-ServerDiagnosticsNegativeTest-1621277291 tempest-ServerDiagnosticsNegativeTest-1621277291-project-member] [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] Start spawning the instance on the hypervisor. {{(pid=61957) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 570.704223] env[61957]: DEBUG nova.network.neutron [None req-d3a5f917-ae8e-4fd7-b4eb-cbb6e49d67de tempest-ServerDiagnosticsNegativeTest-1621277291 tempest-ServerDiagnosticsNegativeTest-1621277291-project-member] [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] Successfully created port: c759a216-a6b5-4bff-97b2-bdfda7cbc7d5 {{(pid=61957) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 570.731684] env[61957]: DEBUG nova.virt.hardware [None req-d3a5f917-ae8e-4fd7-b4eb-cbb6e49d67de tempest-ServerDiagnosticsNegativeTest-1621277291 tempest-ServerDiagnosticsNegativeTest-1621277291-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T17:22:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T17:21:50Z,direct_url=,disk_format='vmdk',id=11c76a2c-f705-470a-ba9d-4657858bab38,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='81d3096d0f094373913fc3dc8f5c3c6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T17:21:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 570.731684] env[61957]: DEBUG nova.virt.hardware [None req-d3a5f917-ae8e-4fd7-b4eb-cbb6e49d67de tempest-ServerDiagnosticsNegativeTest-1621277291 tempest-ServerDiagnosticsNegativeTest-1621277291-project-member] Flavor limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 570.731684] env[61957]: DEBUG nova.virt.hardware [None req-d3a5f917-ae8e-4fd7-b4eb-cbb6e49d67de tempest-ServerDiagnosticsNegativeTest-1621277291 tempest-ServerDiagnosticsNegativeTest-1621277291-project-member] Image limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 570.731827] env[61957]: DEBUG nova.virt.hardware [None req-d3a5f917-ae8e-4fd7-b4eb-cbb6e49d67de tempest-ServerDiagnosticsNegativeTest-1621277291 tempest-ServerDiagnosticsNegativeTest-1621277291-project-member] Flavor pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 570.731827] env[61957]: DEBUG nova.virt.hardware [None req-d3a5f917-ae8e-4fd7-b4eb-cbb6e49d67de tempest-ServerDiagnosticsNegativeTest-1621277291 tempest-ServerDiagnosticsNegativeTest-1621277291-project-member] Image pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 570.731827] env[61957]: DEBUG nova.virt.hardware [None req-d3a5f917-ae8e-4fd7-b4eb-cbb6e49d67de tempest-ServerDiagnosticsNegativeTest-1621277291 tempest-ServerDiagnosticsNegativeTest-1621277291-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 570.731827] env[61957]: DEBUG nova.virt.hardware [None req-d3a5f917-ae8e-4fd7-b4eb-cbb6e49d67de tempest-ServerDiagnosticsNegativeTest-1621277291 tempest-ServerDiagnosticsNegativeTest-1621277291-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 570.731827] env[61957]: DEBUG nova.virt.hardware [None req-d3a5f917-ae8e-4fd7-b4eb-cbb6e49d67de tempest-ServerDiagnosticsNegativeTest-1621277291 tempest-ServerDiagnosticsNegativeTest-1621277291-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 570.731964] env[61957]: DEBUG nova.virt.hardware [None req-d3a5f917-ae8e-4fd7-b4eb-cbb6e49d67de tempest-ServerDiagnosticsNegativeTest-1621277291 tempest-ServerDiagnosticsNegativeTest-1621277291-project-member] Got 1 possible topologies {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 570.731964] env[61957]: DEBUG nova.virt.hardware [None req-d3a5f917-ae8e-4fd7-b4eb-cbb6e49d67de tempest-ServerDiagnosticsNegativeTest-1621277291 tempest-ServerDiagnosticsNegativeTest-1621277291-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 570.731964] env[61957]: DEBUG nova.virt.hardware [None req-d3a5f917-ae8e-4fd7-b4eb-cbb6e49d67de tempest-ServerDiagnosticsNegativeTest-1621277291 tempest-ServerDiagnosticsNegativeTest-1621277291-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 570.732888] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a94ec9d2-a2e0-4774-9c5b-40fe01691980 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.750812] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f428bb5e-c4af-4455-8fd6-532b368981d2 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.864361] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d102ca19-392e-4051-87d1-b58bed2da29d tempest-ServersAdminNegativeTestJSON-555579 tempest-ServersAdminNegativeTestJSON-555579-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.210s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 570.866358] env[61957]: DEBUG nova.compute.manager [None req-d102ca19-392e-4051-87d1-b58bed2da29d tempest-ServersAdminNegativeTestJSON-555579 tempest-ServersAdminNegativeTestJSON-555579-project-member] [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] Start building networks asynchronously for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 570.869683] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d83719b8-6bd9-4ae7-99ea-7d0970401ba5 tempest-ServerExternalEventsTest-1886420864 tempest-ServerExternalEventsTest-1886420864-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.562s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 570.872107] env[61957]: INFO nova.compute.claims [None req-d83719b8-6bd9-4ae7-99ea-7d0970401ba5 tempest-ServerExternalEventsTest-1886420864 tempest-ServerExternalEventsTest-1886420864-project-member] [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 571.380041] env[61957]: DEBUG nova.compute.utils [None req-d102ca19-392e-4051-87d1-b58bed2da29d tempest-ServersAdminNegativeTestJSON-555579 tempest-ServersAdminNegativeTestJSON-555579-project-member] Using /dev/sd instead of None {{(pid=61957) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 571.380041] env[61957]: DEBUG nova.compute.manager [None req-d102ca19-392e-4051-87d1-b58bed2da29d tempest-ServersAdminNegativeTestJSON-555579 tempest-ServersAdminNegativeTestJSON-555579-project-member] [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] Allocating IP information in the background. {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 571.380041] env[61957]: DEBUG nova.network.neutron [None req-d102ca19-392e-4051-87d1-b58bed2da29d tempest-ServersAdminNegativeTestJSON-555579 tempest-ServersAdminNegativeTestJSON-555579-project-member] [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] allocate_for_instance() {{(pid=61957) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 571.542844] env[61957]: DEBUG nova.policy [None req-d102ca19-392e-4051-87d1-b58bed2da29d tempest-ServersAdminNegativeTestJSON-555579 tempest-ServersAdminNegativeTestJSON-555579-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fddfc6f9d8a246e8b4bf4a8410d16f2a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0f6995674d39444a93b11f94b3ff0eaf', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61957) authorize /opt/stack/nova/nova/policy.py:203}} [ 571.888026] env[61957]: DEBUG nova.compute.manager [None req-d102ca19-392e-4051-87d1-b58bed2da29d tempest-ServersAdminNegativeTestJSON-555579 tempest-ServersAdminNegativeTestJSON-555579-project-member] [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] Start building block device mappings for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 571.974307] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbc9d057-0e0d-4aba-84bd-84afed1c4e1d {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.983213] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da4bac41-a367-42db-af77-dafd4b3cfa83 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.017067] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe0b459b-4cbf-4a51-a114-839309a43df4 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.025774] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19c554b7-2250-4bc0-a118-1660c05100c9 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.043783] env[61957]: DEBUG nova.compute.provider_tree [None req-d83719b8-6bd9-4ae7-99ea-7d0970401ba5 tempest-ServerExternalEventsTest-1886420864 tempest-ServerExternalEventsTest-1886420864-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 572.134688] env[61957]: DEBUG oslo_service.periodic_task [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61957) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 572.135198] env[61957]: DEBUG oslo_service.periodic_task [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61957) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 572.135361] env[61957]: DEBUG nova.compute.manager [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Starting heal instance info cache {{(pid=61957) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 572.135491] env[61957]: DEBUG nova.compute.manager [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Rebuilding the list of instances to heal {{(pid=61957) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9930}} [ 572.550425] env[61957]: DEBUG nova.scheduler.client.report [None req-d83719b8-6bd9-4ae7-99ea-7d0970401ba5 tempest-ServerExternalEventsTest-1886420864 tempest-ServerExternalEventsTest-1886420864-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 572.641269] env[61957]: DEBUG nova.compute.manager [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] Skipping network cache update for instance because it is Building. {{(pid=61957) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 572.641472] env[61957]: DEBUG nova.compute.manager [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] Skipping network cache update for instance because it is Building. {{(pid=61957) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 572.641607] env[61957]: DEBUG nova.compute.manager [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] Skipping network cache update for instance because it is Building. {{(pid=61957) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 572.643728] env[61957]: DEBUG nova.compute.manager [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] Skipping network cache update for instance because it is Building. {{(pid=61957) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 572.643728] env[61957]: DEBUG nova.compute.manager [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Didn't find any instances for network info cache update. {{(pid=61957) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10012}} [ 572.643728] env[61957]: DEBUG oslo_service.periodic_task [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61957) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 572.643728] env[61957]: DEBUG oslo_service.periodic_task [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61957) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 572.643728] env[61957]: DEBUG oslo_service.periodic_task [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61957) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 572.643728] env[61957]: DEBUG oslo_service.periodic_task [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61957) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 572.644212] env[61957]: DEBUG oslo_service.periodic_task [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61957) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 572.644212] env[61957]: DEBUG oslo_service.periodic_task [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61957) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 572.644364] env[61957]: DEBUG nova.compute.manager [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61957) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 572.645123] env[61957]: DEBUG oslo_service.periodic_task [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61957) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 572.767091] env[61957]: ERROR nova.compute.manager [None req-abb0f6e8-eaec-4ec9-a6d4-89648c99b1d8 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 6bf15563-9c78-4c42-9a95-c660f6c0e3b5, please check neutron logs for more information. [ 572.767091] env[61957]: ERROR nova.compute.manager Traceback (most recent call last): [ 572.767091] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 572.767091] env[61957]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 572.767091] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 572.767091] env[61957]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 572.767091] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 572.767091] env[61957]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 572.767091] env[61957]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 572.767091] env[61957]: ERROR nova.compute.manager self.force_reraise() [ 572.767091] env[61957]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 572.767091] env[61957]: ERROR nova.compute.manager raise self.value [ 572.767091] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 572.767091] env[61957]: ERROR nova.compute.manager updated_port = self._update_port( [ 572.767091] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 572.767091] env[61957]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 572.767936] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 572.767936] env[61957]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 572.767936] env[61957]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 6bf15563-9c78-4c42-9a95-c660f6c0e3b5, please check neutron logs for more information. [ 572.767936] env[61957]: ERROR nova.compute.manager [ 572.767936] env[61957]: Traceback (most recent call last): [ 572.767936] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 572.767936] env[61957]: listener.cb(fileno) [ 572.767936] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 572.767936] env[61957]: result = function(*args, **kwargs) [ 572.767936] env[61957]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 572.767936] env[61957]: return func(*args, **kwargs) [ 572.767936] env[61957]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 572.767936] env[61957]: raise e [ 572.767936] env[61957]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 572.767936] env[61957]: nwinfo = self.network_api.allocate_for_instance( [ 572.767936] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 572.767936] env[61957]: created_port_ids = self._update_ports_for_instance( [ 572.767936] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 572.767936] env[61957]: with excutils.save_and_reraise_exception(): [ 572.767936] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 572.767936] env[61957]: self.force_reraise() [ 572.767936] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 572.767936] env[61957]: raise self.value [ 572.767936] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 572.767936] env[61957]: updated_port = self._update_port( [ 572.767936] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 572.767936] env[61957]: _ensure_no_port_binding_failure(port) [ 572.767936] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 572.767936] env[61957]: raise exception.PortBindingFailed(port_id=port['id']) [ 572.769817] env[61957]: nova.exception.PortBindingFailed: Binding failed for port 6bf15563-9c78-4c42-9a95-c660f6c0e3b5, please check neutron logs for more information. [ 572.769817] env[61957]: Removing descriptor: 15 [ 572.769817] env[61957]: ERROR nova.compute.manager [None req-abb0f6e8-eaec-4ec9-a6d4-89648c99b1d8 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 6bf15563-9c78-4c42-9a95-c660f6c0e3b5, please check neutron logs for more information. [ 572.769817] env[61957]: ERROR nova.compute.manager [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] Traceback (most recent call last): [ 572.769817] env[61957]: ERROR nova.compute.manager [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 572.769817] env[61957]: ERROR nova.compute.manager [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] yield resources [ 572.769817] env[61957]: ERROR nova.compute.manager [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 572.769817] env[61957]: ERROR nova.compute.manager [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] self.driver.spawn(context, instance, image_meta, [ 572.769817] env[61957]: ERROR nova.compute.manager [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 572.769817] env[61957]: ERROR nova.compute.manager [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] self._vmops.spawn(context, instance, image_meta, injected_files, [ 572.769817] env[61957]: ERROR nova.compute.manager [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 572.769817] env[61957]: ERROR nova.compute.manager [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] vm_ref = self.build_virtual_machine(instance, [ 572.770165] env[61957]: ERROR nova.compute.manager [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 572.770165] env[61957]: ERROR nova.compute.manager [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] vif_infos = vmwarevif.get_vif_info(self._session, [ 572.770165] env[61957]: ERROR nova.compute.manager [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 572.770165] env[61957]: ERROR nova.compute.manager [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] for vif in network_info: [ 572.770165] env[61957]: ERROR nova.compute.manager [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 572.770165] env[61957]: ERROR nova.compute.manager [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] return self._sync_wrapper(fn, *args, **kwargs) [ 572.770165] env[61957]: ERROR nova.compute.manager [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 572.770165] env[61957]: ERROR nova.compute.manager [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] self.wait() [ 572.770165] env[61957]: ERROR nova.compute.manager [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 572.770165] env[61957]: ERROR nova.compute.manager [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] self[:] = self._gt.wait() [ 572.770165] env[61957]: ERROR nova.compute.manager [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 572.770165] env[61957]: ERROR nova.compute.manager [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] return self._exit_event.wait() [ 572.770165] env[61957]: ERROR nova.compute.manager [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 572.770497] env[61957]: ERROR nova.compute.manager [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] result = hub.switch() [ 572.770497] env[61957]: ERROR nova.compute.manager [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 572.770497] env[61957]: ERROR nova.compute.manager [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] return self.greenlet.switch() [ 572.770497] env[61957]: ERROR nova.compute.manager [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 572.770497] env[61957]: ERROR nova.compute.manager [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] result = function(*args, **kwargs) [ 572.770497] env[61957]: ERROR nova.compute.manager [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 572.770497] env[61957]: ERROR nova.compute.manager [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] return func(*args, **kwargs) [ 572.770497] env[61957]: ERROR nova.compute.manager [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 572.770497] env[61957]: ERROR nova.compute.manager [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] raise e [ 572.770497] env[61957]: ERROR nova.compute.manager [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 572.770497] env[61957]: ERROR nova.compute.manager [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] nwinfo = self.network_api.allocate_for_instance( [ 572.770497] env[61957]: ERROR nova.compute.manager [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 572.770497] env[61957]: ERROR nova.compute.manager [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] created_port_ids = self._update_ports_for_instance( [ 572.770936] env[61957]: ERROR nova.compute.manager [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 572.770936] env[61957]: ERROR nova.compute.manager [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] with excutils.save_and_reraise_exception(): [ 572.770936] env[61957]: ERROR nova.compute.manager [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 572.770936] env[61957]: ERROR nova.compute.manager [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] self.force_reraise() [ 572.770936] env[61957]: ERROR nova.compute.manager [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 572.770936] env[61957]: ERROR nova.compute.manager [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] raise self.value [ 572.770936] env[61957]: ERROR nova.compute.manager [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 572.770936] env[61957]: ERROR nova.compute.manager [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] updated_port = self._update_port( [ 572.770936] env[61957]: ERROR nova.compute.manager [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 572.770936] env[61957]: ERROR nova.compute.manager [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] _ensure_no_port_binding_failure(port) [ 572.770936] env[61957]: ERROR nova.compute.manager [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 572.770936] env[61957]: ERROR nova.compute.manager [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] raise exception.PortBindingFailed(port_id=port['id']) [ 572.771255] env[61957]: ERROR nova.compute.manager [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] nova.exception.PortBindingFailed: Binding failed for port 6bf15563-9c78-4c42-9a95-c660f6c0e3b5, please check neutron logs for more information. [ 572.771255] env[61957]: ERROR nova.compute.manager [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] [ 572.771255] env[61957]: INFO nova.compute.manager [None req-abb0f6e8-eaec-4ec9-a6d4-89648c99b1d8 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] Terminating instance [ 572.774111] env[61957]: DEBUG oslo_concurrency.lockutils [None req-abb0f6e8-eaec-4ec9-a6d4-89648c99b1d8 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] Acquiring lock "refresh_cache-1a0e117e-aa56-44fa-8818-71f0fd0b6183" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 572.774281] env[61957]: DEBUG oslo_concurrency.lockutils [None req-abb0f6e8-eaec-4ec9-a6d4-89648c99b1d8 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] Acquired lock "refresh_cache-1a0e117e-aa56-44fa-8818-71f0fd0b6183" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 572.774447] env[61957]: DEBUG nova.network.neutron [None req-abb0f6e8-eaec-4ec9-a6d4-89648c99b1d8 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 572.815330] env[61957]: DEBUG nova.network.neutron [None req-d102ca19-392e-4051-87d1-b58bed2da29d tempest-ServersAdminNegativeTestJSON-555579 tempest-ServersAdminNegativeTestJSON-555579-project-member] [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] Successfully created port: 178d7697-f823-4565-9fc6-706777f5e312 {{(pid=61957) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 572.899912] env[61957]: DEBUG nova.compute.manager [None req-d102ca19-392e-4051-87d1-b58bed2da29d tempest-ServersAdminNegativeTestJSON-555579 tempest-ServersAdminNegativeTestJSON-555579-project-member] [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] Start spawning the instance on the hypervisor. {{(pid=61957) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 572.938142] env[61957]: DEBUG nova.virt.hardware [None req-d102ca19-392e-4051-87d1-b58bed2da29d tempest-ServersAdminNegativeTestJSON-555579 tempest-ServersAdminNegativeTestJSON-555579-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T17:22:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T17:21:50Z,direct_url=,disk_format='vmdk',id=11c76a2c-f705-470a-ba9d-4657858bab38,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='81d3096d0f094373913fc3dc8f5c3c6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T17:21:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 572.938360] env[61957]: DEBUG nova.virt.hardware [None req-d102ca19-392e-4051-87d1-b58bed2da29d tempest-ServersAdminNegativeTestJSON-555579 tempest-ServersAdminNegativeTestJSON-555579-project-member] Flavor limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 572.938512] env[61957]: DEBUG nova.virt.hardware [None req-d102ca19-392e-4051-87d1-b58bed2da29d tempest-ServersAdminNegativeTestJSON-555579 tempest-ServersAdminNegativeTestJSON-555579-project-member] Image limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 572.939642] env[61957]: DEBUG nova.virt.hardware [None req-d102ca19-392e-4051-87d1-b58bed2da29d tempest-ServersAdminNegativeTestJSON-555579 tempest-ServersAdminNegativeTestJSON-555579-project-member] Flavor pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 572.939642] env[61957]: DEBUG nova.virt.hardware [None req-d102ca19-392e-4051-87d1-b58bed2da29d tempest-ServersAdminNegativeTestJSON-555579 tempest-ServersAdminNegativeTestJSON-555579-project-member] Image pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 572.939642] env[61957]: DEBUG nova.virt.hardware [None req-d102ca19-392e-4051-87d1-b58bed2da29d tempest-ServersAdminNegativeTestJSON-555579 tempest-ServersAdminNegativeTestJSON-555579-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 572.939642] env[61957]: DEBUG nova.virt.hardware [None req-d102ca19-392e-4051-87d1-b58bed2da29d tempest-ServersAdminNegativeTestJSON-555579 tempest-ServersAdminNegativeTestJSON-555579-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 572.939882] env[61957]: DEBUG nova.virt.hardware [None req-d102ca19-392e-4051-87d1-b58bed2da29d tempest-ServersAdminNegativeTestJSON-555579 tempest-ServersAdminNegativeTestJSON-555579-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 572.939979] env[61957]: DEBUG nova.virt.hardware [None req-d102ca19-392e-4051-87d1-b58bed2da29d tempest-ServersAdminNegativeTestJSON-555579 tempest-ServersAdminNegativeTestJSON-555579-project-member] Got 1 possible topologies {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 572.940162] env[61957]: DEBUG nova.virt.hardware [None req-d102ca19-392e-4051-87d1-b58bed2da29d tempest-ServersAdminNegativeTestJSON-555579 tempest-ServersAdminNegativeTestJSON-555579-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 572.940333] env[61957]: DEBUG nova.virt.hardware [None req-d102ca19-392e-4051-87d1-b58bed2da29d tempest-ServersAdminNegativeTestJSON-555579 tempest-ServersAdminNegativeTestJSON-555579-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 572.941246] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e7411e3-5a94-409a-9107-02069e697447 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.952998] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2f57fee-f927-4178-af65-72228e4e4986 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 573.057053] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d83719b8-6bd9-4ae7-99ea-7d0970401ba5 tempest-ServerExternalEventsTest-1886420864 tempest-ServerExternalEventsTest-1886420864-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.187s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 573.057638] env[61957]: DEBUG nova.compute.manager [None req-d83719b8-6bd9-4ae7-99ea-7d0970401ba5 tempest-ServerExternalEventsTest-1886420864 tempest-ServerExternalEventsTest-1886420864-project-member] [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] Start building networks asynchronously for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 573.154250] env[61957]: DEBUG oslo_concurrency.lockutils [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 573.154346] env[61957]: DEBUG oslo_concurrency.lockutils [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.001s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 573.154853] env[61957]: DEBUG oslo_concurrency.lockutils [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 573.155040] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61957) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 573.156109] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef9bbd69-f426-48c9-8b05-47fa01b136a2 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 573.167674] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1341a82-4661-4b2d-89f1-691912f12938 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 573.190031] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3577372-2de2-4aee-b117-faab77f9afd4 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 573.198928] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-226242d7-11a4-4d5e-8ba6-4847d1d812a3 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 573.249444] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181483MB free_disk=142GB free_vcpus=48 pci_devices=None {{(pid=61957) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 573.249744] env[61957]: DEBUG oslo_concurrency.lockutils [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 573.249811] env[61957]: DEBUG oslo_concurrency.lockutils [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 573.333028] env[61957]: DEBUG nova.network.neutron [None req-abb0f6e8-eaec-4ec9-a6d4-89648c99b1d8 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 573.506815] env[61957]: DEBUG nova.network.neutron [None req-abb0f6e8-eaec-4ec9-a6d4-89648c99b1d8 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 573.564408] env[61957]: DEBUG nova.compute.utils [None req-d83719b8-6bd9-4ae7-99ea-7d0970401ba5 tempest-ServerExternalEventsTest-1886420864 tempest-ServerExternalEventsTest-1886420864-project-member] Using /dev/sd instead of None {{(pid=61957) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 573.566759] env[61957]: DEBUG nova.compute.manager [None req-d83719b8-6bd9-4ae7-99ea-7d0970401ba5 tempest-ServerExternalEventsTest-1886420864 tempest-ServerExternalEventsTest-1886420864-project-member] [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] Allocating IP information in the background. {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 573.567155] env[61957]: DEBUG nova.network.neutron [None req-d83719b8-6bd9-4ae7-99ea-7d0970401ba5 tempest-ServerExternalEventsTest-1886420864 tempest-ServerExternalEventsTest-1886420864-project-member] [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] allocate_for_instance() {{(pid=61957) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 573.833158] env[61957]: DEBUG nova.policy [None req-d83719b8-6bd9-4ae7-99ea-7d0970401ba5 tempest-ServerExternalEventsTest-1886420864 tempest-ServerExternalEventsTest-1886420864-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5759c30016ff4f1382b9bda3bd3ec351', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8655d663549047b99084e82820679ab4', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61957) authorize /opt/stack/nova/nova/policy.py:203}} [ 573.851435] env[61957]: DEBUG nova.compute.manager [req-ce1396c4-e625-41ff-9e99-797e724f8383 req-78f5f04a-45ee-4914-99be-e2b2b7c1f44c service nova] [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] Received event network-changed-6bf15563-9c78-4c42-9a95-c660f6c0e3b5 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 573.851435] env[61957]: DEBUG nova.compute.manager [req-ce1396c4-e625-41ff-9e99-797e724f8383 req-78f5f04a-45ee-4914-99be-e2b2b7c1f44c service nova] [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] Refreshing instance network info cache due to event network-changed-6bf15563-9c78-4c42-9a95-c660f6c0e3b5. {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 573.851435] env[61957]: DEBUG oslo_concurrency.lockutils [req-ce1396c4-e625-41ff-9e99-797e724f8383 req-78f5f04a-45ee-4914-99be-e2b2b7c1f44c service nova] Acquiring lock "refresh_cache-1a0e117e-aa56-44fa-8818-71f0fd0b6183" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 573.928858] env[61957]: DEBUG oslo_concurrency.lockutils [None req-837238f2-cf29-45f6-8541-231ac85b718c tempest-ServerDiagnosticsTest-1677027478 tempest-ServerDiagnosticsTest-1677027478-project-member] Acquiring lock "43942e78-7703-444d-91e3-bc7913299654" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 573.929137] env[61957]: DEBUG oslo_concurrency.lockutils [None req-837238f2-cf29-45f6-8541-231ac85b718c tempest-ServerDiagnosticsTest-1677027478 tempest-ServerDiagnosticsTest-1677027478-project-member] Lock "43942e78-7703-444d-91e3-bc7913299654" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 573.948295] env[61957]: DEBUG oslo_concurrency.lockutils [None req-423a0985-b17f-45fa-b4ad-54e09005ad46 tempest-TenantUsagesTestJSON-2011428287 tempest-TenantUsagesTestJSON-2011428287-project-member] Acquiring lock "ed456ff1-249c-45d0-a007-3fef96ae8a2d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 573.948554] env[61957]: DEBUG oslo_concurrency.lockutils [None req-423a0985-b17f-45fa-b4ad-54e09005ad46 tempest-TenantUsagesTestJSON-2011428287 tempest-TenantUsagesTestJSON-2011428287-project-member] Lock "ed456ff1-249c-45d0-a007-3fef96ae8a2d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 574.009189] env[61957]: DEBUG oslo_concurrency.lockutils [None req-abb0f6e8-eaec-4ec9-a6d4-89648c99b1d8 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] Releasing lock "refresh_cache-1a0e117e-aa56-44fa-8818-71f0fd0b6183" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 574.009649] env[61957]: DEBUG nova.compute.manager [None req-abb0f6e8-eaec-4ec9-a6d4-89648c99b1d8 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] Start destroying the instance on the hypervisor. {{(pid=61957) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 574.009826] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-abb0f6e8-eaec-4ec9-a6d4-89648c99b1d8 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] Destroying instance {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 574.010196] env[61957]: DEBUG oslo_concurrency.lockutils [req-ce1396c4-e625-41ff-9e99-797e724f8383 req-78f5f04a-45ee-4914-99be-e2b2b7c1f44c service nova] Acquired lock "refresh_cache-1a0e117e-aa56-44fa-8818-71f0fd0b6183" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 574.010322] env[61957]: DEBUG nova.network.neutron [req-ce1396c4-e625-41ff-9e99-797e724f8383 req-78f5f04a-45ee-4914-99be-e2b2b7c1f44c service nova] [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] Refreshing network info cache for port 6bf15563-9c78-4c42-9a95-c660f6c0e3b5 {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 574.011444] env[61957]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-57e7f48c-f248-43d9-a87c-807211a06f56 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.026433] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88d03a6b-44f3-4e8d-b7a9-72df48899600 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.055047] env[61957]: WARNING nova.virt.vmwareapi.vmops [None req-abb0f6e8-eaec-4ec9-a6d4-89648c99b1d8 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 1a0e117e-aa56-44fa-8818-71f0fd0b6183 could not be found. [ 574.055593] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-abb0f6e8-eaec-4ec9-a6d4-89648c99b1d8 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] Instance destroyed {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 574.055688] env[61957]: INFO nova.compute.manager [None req-abb0f6e8-eaec-4ec9-a6d4-89648c99b1d8 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] Took 0.05 seconds to destroy the instance on the hypervisor. [ 574.056053] env[61957]: DEBUG oslo.service.loopingcall [None req-abb0f6e8-eaec-4ec9-a6d4-89648c99b1d8 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 574.057260] env[61957]: DEBUG nova.compute.manager [-] [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 574.057260] env[61957]: DEBUG nova.network.neutron [-] [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 574.070756] env[61957]: DEBUG nova.compute.manager [None req-d83719b8-6bd9-4ae7-99ea-7d0970401ba5 tempest-ServerExternalEventsTest-1886420864 tempest-ServerExternalEventsTest-1886420864-project-member] [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] Start building block device mappings for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 574.133118] env[61957]: DEBUG nova.network.neutron [-] [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 574.231245] env[61957]: ERROR nova.compute.manager [None req-d3a5f917-ae8e-4fd7-b4eb-cbb6e49d67de tempest-ServerDiagnosticsNegativeTest-1621277291 tempest-ServerDiagnosticsNegativeTest-1621277291-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port c759a216-a6b5-4bff-97b2-bdfda7cbc7d5, please check neutron logs for more information. [ 574.231245] env[61957]: ERROR nova.compute.manager Traceback (most recent call last): [ 574.231245] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 574.231245] env[61957]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 574.231245] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 574.231245] env[61957]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 574.231245] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 574.231245] env[61957]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 574.231245] env[61957]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 574.231245] env[61957]: ERROR nova.compute.manager self.force_reraise() [ 574.231245] env[61957]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 574.231245] env[61957]: ERROR nova.compute.manager raise self.value [ 574.231245] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 574.231245] env[61957]: ERROR nova.compute.manager updated_port = self._update_port( [ 574.231245] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 574.231245] env[61957]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 574.231728] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 574.231728] env[61957]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 574.231728] env[61957]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port c759a216-a6b5-4bff-97b2-bdfda7cbc7d5, please check neutron logs for more information. [ 574.231728] env[61957]: ERROR nova.compute.manager [ 574.231728] env[61957]: Traceback (most recent call last): [ 574.231728] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 574.231728] env[61957]: listener.cb(fileno) [ 574.231728] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 574.231728] env[61957]: result = function(*args, **kwargs) [ 574.231728] env[61957]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 574.231728] env[61957]: return func(*args, **kwargs) [ 574.231728] env[61957]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 574.231728] env[61957]: raise e [ 574.231728] env[61957]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 574.231728] env[61957]: nwinfo = self.network_api.allocate_for_instance( [ 574.231728] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 574.231728] env[61957]: created_port_ids = self._update_ports_for_instance( [ 574.231728] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 574.231728] env[61957]: with excutils.save_and_reraise_exception(): [ 574.231728] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 574.231728] env[61957]: self.force_reraise() [ 574.231728] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 574.231728] env[61957]: raise self.value [ 574.231728] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 574.231728] env[61957]: updated_port = self._update_port( [ 574.231728] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 574.231728] env[61957]: _ensure_no_port_binding_failure(port) [ 574.231728] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 574.231728] env[61957]: raise exception.PortBindingFailed(port_id=port['id']) [ 574.232664] env[61957]: nova.exception.PortBindingFailed: Binding failed for port c759a216-a6b5-4bff-97b2-bdfda7cbc7d5, please check neutron logs for more information. [ 574.232664] env[61957]: Removing descriptor: 16 [ 574.232664] env[61957]: ERROR nova.compute.manager [None req-d3a5f917-ae8e-4fd7-b4eb-cbb6e49d67de tempest-ServerDiagnosticsNegativeTest-1621277291 tempest-ServerDiagnosticsNegativeTest-1621277291-project-member] [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port c759a216-a6b5-4bff-97b2-bdfda7cbc7d5, please check neutron logs for more information. [ 574.232664] env[61957]: ERROR nova.compute.manager [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] Traceback (most recent call last): [ 574.232664] env[61957]: ERROR nova.compute.manager [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 574.232664] env[61957]: ERROR nova.compute.manager [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] yield resources [ 574.232664] env[61957]: ERROR nova.compute.manager [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 574.232664] env[61957]: ERROR nova.compute.manager [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] self.driver.spawn(context, instance, image_meta, [ 574.232664] env[61957]: ERROR nova.compute.manager [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 574.232664] env[61957]: ERROR nova.compute.manager [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] self._vmops.spawn(context, instance, image_meta, injected_files, [ 574.232664] env[61957]: ERROR nova.compute.manager [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 574.232664] env[61957]: ERROR nova.compute.manager [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] vm_ref = self.build_virtual_machine(instance, [ 574.233081] env[61957]: ERROR nova.compute.manager [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 574.233081] env[61957]: ERROR nova.compute.manager [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] vif_infos = vmwarevif.get_vif_info(self._session, [ 574.233081] env[61957]: ERROR nova.compute.manager [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 574.233081] env[61957]: ERROR nova.compute.manager [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] for vif in network_info: [ 574.233081] env[61957]: ERROR nova.compute.manager [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 574.233081] env[61957]: ERROR nova.compute.manager [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] return self._sync_wrapper(fn, *args, **kwargs) [ 574.233081] env[61957]: ERROR nova.compute.manager [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 574.233081] env[61957]: ERROR nova.compute.manager [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] self.wait() [ 574.233081] env[61957]: ERROR nova.compute.manager [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 574.233081] env[61957]: ERROR nova.compute.manager [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] self[:] = self._gt.wait() [ 574.233081] env[61957]: ERROR nova.compute.manager [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 574.233081] env[61957]: ERROR nova.compute.manager [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] return self._exit_event.wait() [ 574.233081] env[61957]: ERROR nova.compute.manager [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 574.233410] env[61957]: ERROR nova.compute.manager [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] result = hub.switch() [ 574.233410] env[61957]: ERROR nova.compute.manager [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 574.233410] env[61957]: ERROR nova.compute.manager [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] return self.greenlet.switch() [ 574.233410] env[61957]: ERROR nova.compute.manager [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 574.233410] env[61957]: ERROR nova.compute.manager [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] result = function(*args, **kwargs) [ 574.233410] env[61957]: ERROR nova.compute.manager [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 574.233410] env[61957]: ERROR nova.compute.manager [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] return func(*args, **kwargs) [ 574.233410] env[61957]: ERROR nova.compute.manager [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 574.233410] env[61957]: ERROR nova.compute.manager [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] raise e [ 574.233410] env[61957]: ERROR nova.compute.manager [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 574.233410] env[61957]: ERROR nova.compute.manager [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] nwinfo = self.network_api.allocate_for_instance( [ 574.233410] env[61957]: ERROR nova.compute.manager [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 574.233410] env[61957]: ERROR nova.compute.manager [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] created_port_ids = self._update_ports_for_instance( [ 574.233975] env[61957]: ERROR nova.compute.manager [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 574.233975] env[61957]: ERROR nova.compute.manager [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] with excutils.save_and_reraise_exception(): [ 574.233975] env[61957]: ERROR nova.compute.manager [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 574.233975] env[61957]: ERROR nova.compute.manager [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] self.force_reraise() [ 574.233975] env[61957]: ERROR nova.compute.manager [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 574.233975] env[61957]: ERROR nova.compute.manager [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] raise self.value [ 574.233975] env[61957]: ERROR nova.compute.manager [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 574.233975] env[61957]: ERROR nova.compute.manager [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] updated_port = self._update_port( [ 574.233975] env[61957]: ERROR nova.compute.manager [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 574.233975] env[61957]: ERROR nova.compute.manager [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] _ensure_no_port_binding_failure(port) [ 574.233975] env[61957]: ERROR nova.compute.manager [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 574.233975] env[61957]: ERROR nova.compute.manager [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] raise exception.PortBindingFailed(port_id=port['id']) [ 574.234301] env[61957]: ERROR nova.compute.manager [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] nova.exception.PortBindingFailed: Binding failed for port c759a216-a6b5-4bff-97b2-bdfda7cbc7d5, please check neutron logs for more information. [ 574.234301] env[61957]: ERROR nova.compute.manager [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] [ 574.234301] env[61957]: INFO nova.compute.manager [None req-d3a5f917-ae8e-4fd7-b4eb-cbb6e49d67de tempest-ServerDiagnosticsNegativeTest-1621277291 tempest-ServerDiagnosticsNegativeTest-1621277291-project-member] [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] Terminating instance [ 574.238870] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d3a5f917-ae8e-4fd7-b4eb-cbb6e49d67de tempest-ServerDiagnosticsNegativeTest-1621277291 tempest-ServerDiagnosticsNegativeTest-1621277291-project-member] Acquiring lock "refresh_cache-f28611df-9528-4ec0-b575-ad54ebd4a958" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 574.238870] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d3a5f917-ae8e-4fd7-b4eb-cbb6e49d67de tempest-ServerDiagnosticsNegativeTest-1621277291 tempest-ServerDiagnosticsNegativeTest-1621277291-project-member] Acquired lock "refresh_cache-f28611df-9528-4ec0-b575-ad54ebd4a958" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 574.238870] env[61957]: DEBUG nova.network.neutron [None req-d3a5f917-ae8e-4fd7-b4eb-cbb6e49d67de tempest-ServerDiagnosticsNegativeTest-1621277291 tempest-ServerDiagnosticsNegativeTest-1621277291-project-member] [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 574.284751] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Instance 1a0e117e-aa56-44fa-8818-71f0fd0b6183 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61957) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 574.284980] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Instance f28611df-9528-4ec0-b575-ad54ebd4a958 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61957) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 574.285138] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Instance 37837096-5ad0-4b6f-b32d-f2899d88b6dd actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61957) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 574.285259] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Instance abdb0ba9-abef-4528-96fb-ae7d7d41f8fc actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61957) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 574.435437] env[61957]: DEBUG nova.compute.manager [None req-837238f2-cf29-45f6-8541-231ac85b718c tempest-ServerDiagnosticsTest-1677027478 tempest-ServerDiagnosticsTest-1677027478-project-member] [instance: 43942e78-7703-444d-91e3-bc7913299654] Starting instance... {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 574.452303] env[61957]: DEBUG nova.compute.manager [None req-423a0985-b17f-45fa-b4ad-54e09005ad46 tempest-TenantUsagesTestJSON-2011428287 tempest-TenantUsagesTestJSON-2011428287-project-member] [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] Starting instance... {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 574.561466] env[61957]: DEBUG nova.network.neutron [req-ce1396c4-e625-41ff-9e99-797e724f8383 req-78f5f04a-45ee-4914-99be-e2b2b7c1f44c service nova] [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 574.638414] env[61957]: DEBUG nova.network.neutron [-] [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 574.707768] env[61957]: DEBUG nova.network.neutron [req-ce1396c4-e625-41ff-9e99-797e724f8383 req-78f5f04a-45ee-4914-99be-e2b2b7c1f44c service nova] [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 574.786820] env[61957]: DEBUG nova.network.neutron [None req-d3a5f917-ae8e-4fd7-b4eb-cbb6e49d67de tempest-ServerDiagnosticsNegativeTest-1621277291 tempest-ServerDiagnosticsNegativeTest-1621277291-project-member] [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 574.789493] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Instance 43942e78-7703-444d-91e3-bc7913299654 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61957) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 574.956377] env[61957]: DEBUG oslo_concurrency.lockutils [None req-837238f2-cf29-45f6-8541-231ac85b718c tempest-ServerDiagnosticsTest-1677027478 tempest-ServerDiagnosticsTest-1677027478-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 574.960998] env[61957]: DEBUG nova.network.neutron [None req-d83719b8-6bd9-4ae7-99ea-7d0970401ba5 tempest-ServerExternalEventsTest-1886420864 tempest-ServerExternalEventsTest-1886420864-project-member] [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] Successfully created port: 3d42e5dd-f713-445a-8855-3e0e765e2c4a {{(pid=61957) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 574.968908] env[61957]: DEBUG nova.network.neutron [None req-d3a5f917-ae8e-4fd7-b4eb-cbb6e49d67de tempest-ServerDiagnosticsNegativeTest-1621277291 tempest-ServerDiagnosticsNegativeTest-1621277291-project-member] [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 574.978052] env[61957]: DEBUG oslo_concurrency.lockutils [None req-423a0985-b17f-45fa-b4ad-54e09005ad46 tempest-TenantUsagesTestJSON-2011428287 tempest-TenantUsagesTestJSON-2011428287-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 575.085154] env[61957]: DEBUG nova.compute.manager [None req-d83719b8-6bd9-4ae7-99ea-7d0970401ba5 tempest-ServerExternalEventsTest-1886420864 tempest-ServerExternalEventsTest-1886420864-project-member] [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] Start spawning the instance on the hypervisor. {{(pid=61957) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 575.116088] env[61957]: DEBUG nova.virt.hardware [None req-d83719b8-6bd9-4ae7-99ea-7d0970401ba5 tempest-ServerExternalEventsTest-1886420864 tempest-ServerExternalEventsTest-1886420864-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T17:22:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T17:21:50Z,direct_url=,disk_format='vmdk',id=11c76a2c-f705-470a-ba9d-4657858bab38,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='81d3096d0f094373913fc3dc8f5c3c6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T17:21:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 575.116295] env[61957]: DEBUG nova.virt.hardware [None req-d83719b8-6bd9-4ae7-99ea-7d0970401ba5 tempest-ServerExternalEventsTest-1886420864 tempest-ServerExternalEventsTest-1886420864-project-member] Flavor limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 575.116295] env[61957]: DEBUG nova.virt.hardware [None req-d83719b8-6bd9-4ae7-99ea-7d0970401ba5 tempest-ServerExternalEventsTest-1886420864 tempest-ServerExternalEventsTest-1886420864-project-member] Image limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 575.117045] env[61957]: DEBUG nova.virt.hardware [None req-d83719b8-6bd9-4ae7-99ea-7d0970401ba5 tempest-ServerExternalEventsTest-1886420864 tempest-ServerExternalEventsTest-1886420864-project-member] Flavor pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 575.117045] env[61957]: DEBUG nova.virt.hardware [None req-d83719b8-6bd9-4ae7-99ea-7d0970401ba5 tempest-ServerExternalEventsTest-1886420864 tempest-ServerExternalEventsTest-1886420864-project-member] Image pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 575.117045] env[61957]: DEBUG nova.virt.hardware [None req-d83719b8-6bd9-4ae7-99ea-7d0970401ba5 tempest-ServerExternalEventsTest-1886420864 tempest-ServerExternalEventsTest-1886420864-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 575.117045] env[61957]: DEBUG nova.virt.hardware [None req-d83719b8-6bd9-4ae7-99ea-7d0970401ba5 tempest-ServerExternalEventsTest-1886420864 tempest-ServerExternalEventsTest-1886420864-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 575.117336] env[61957]: DEBUG nova.virt.hardware [None req-d83719b8-6bd9-4ae7-99ea-7d0970401ba5 tempest-ServerExternalEventsTest-1886420864 tempest-ServerExternalEventsTest-1886420864-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 575.117336] env[61957]: DEBUG nova.virt.hardware [None req-d83719b8-6bd9-4ae7-99ea-7d0970401ba5 tempest-ServerExternalEventsTest-1886420864 tempest-ServerExternalEventsTest-1886420864-project-member] Got 1 possible topologies {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 575.117435] env[61957]: DEBUG nova.virt.hardware [None req-d83719b8-6bd9-4ae7-99ea-7d0970401ba5 tempest-ServerExternalEventsTest-1886420864 tempest-ServerExternalEventsTest-1886420864-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 575.117600] env[61957]: DEBUG nova.virt.hardware [None req-d83719b8-6bd9-4ae7-99ea-7d0970401ba5 tempest-ServerExternalEventsTest-1886420864 tempest-ServerExternalEventsTest-1886420864-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 575.118502] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43a62469-bfac-4688-b426-15a4f00c7b43 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.128697] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f64cecd1-98bb-4c94-b10a-010d44ca3bff {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.145562] env[61957]: INFO nova.compute.manager [-] [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] Took 1.09 seconds to deallocate network for instance. [ 575.150212] env[61957]: DEBUG nova.compute.claims [None req-abb0f6e8-eaec-4ec9-a6d4-89648c99b1d8 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] Aborting claim: {{(pid=61957) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 575.150390] env[61957]: DEBUG oslo_concurrency.lockutils [None req-abb0f6e8-eaec-4ec9-a6d4-89648c99b1d8 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 575.214744] env[61957]: DEBUG oslo_concurrency.lockutils [req-ce1396c4-e625-41ff-9e99-797e724f8383 req-78f5f04a-45ee-4914-99be-e2b2b7c1f44c service nova] Releasing lock "refresh_cache-1a0e117e-aa56-44fa-8818-71f0fd0b6183" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 575.292381] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Instance ed456ff1-249c-45d0-a007-3fef96ae8a2d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61957) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 575.295020] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Total usable vcpus: 48, total allocated vcpus: 4 {{(pid=61957) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 575.295020] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1280MB phys_disk=200GB used_disk=4GB total_vcpus=48 used_vcpus=4 pci_stats=[] {{(pid=61957) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 575.370192] env[61957]: DEBUG oslo_concurrency.lockutils [None req-4b8bfc5c-9cfa-4b72-9c9f-71ae820b9856 tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] Acquiring lock "c5af147e-3526-4014-98bc-7ad163ae89d9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 575.370414] env[61957]: DEBUG oslo_concurrency.lockutils [None req-4b8bfc5c-9cfa-4b72-9c9f-71ae820b9856 tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] Lock "c5af147e-3526-4014-98bc-7ad163ae89d9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 575.430212] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-396e0784-d7ff-42a6-82af-ea3d6da5c9de {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.443010] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51fb27ce-f1ca-4844-b106-889f1422062b {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.479832] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d3a5f917-ae8e-4fd7-b4eb-cbb6e49d67de tempest-ServerDiagnosticsNegativeTest-1621277291 tempest-ServerDiagnosticsNegativeTest-1621277291-project-member] Releasing lock "refresh_cache-f28611df-9528-4ec0-b575-ad54ebd4a958" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 575.480398] env[61957]: DEBUG nova.compute.manager [None req-d3a5f917-ae8e-4fd7-b4eb-cbb6e49d67de tempest-ServerDiagnosticsNegativeTest-1621277291 tempest-ServerDiagnosticsNegativeTest-1621277291-project-member] [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] Start destroying the instance on the hypervisor. {{(pid=61957) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 575.480518] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-d3a5f917-ae8e-4fd7-b4eb-cbb6e49d67de tempest-ServerDiagnosticsNegativeTest-1621277291 tempest-ServerDiagnosticsNegativeTest-1621277291-project-member] [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] Destroying instance {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 575.483118] env[61957]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2106d540-4700-4407-ba45-f9fa6cb01003 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.486705] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4ba034f-cffe-479a-b6ad-7b9e8ecadd35 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.500502] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b627283-41e5-4c7b-b57d-cefaebe8ed9f {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.518879] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c30867d-25ed-4b27-b6a9-eda09d4be72f {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.529395] env[61957]: WARNING nova.virt.vmwareapi.vmops [None req-d3a5f917-ae8e-4fd7-b4eb-cbb6e49d67de tempest-ServerDiagnosticsNegativeTest-1621277291 tempest-ServerDiagnosticsNegativeTest-1621277291-project-member] [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance f28611df-9528-4ec0-b575-ad54ebd4a958 could not be found. [ 575.529619] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-d3a5f917-ae8e-4fd7-b4eb-cbb6e49d67de tempest-ServerDiagnosticsNegativeTest-1621277291 tempest-ServerDiagnosticsNegativeTest-1621277291-project-member] [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] Instance destroyed {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 575.529830] env[61957]: INFO nova.compute.manager [None req-d3a5f917-ae8e-4fd7-b4eb-cbb6e49d67de tempest-ServerDiagnosticsNegativeTest-1621277291 tempest-ServerDiagnosticsNegativeTest-1621277291-project-member] [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] Took 0.05 seconds to destroy the instance on the hypervisor. [ 575.530047] env[61957]: DEBUG oslo.service.loopingcall [None req-d3a5f917-ae8e-4fd7-b4eb-cbb6e49d67de tempest-ServerDiagnosticsNegativeTest-1621277291 tempest-ServerDiagnosticsNegativeTest-1621277291-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 575.530992] env[61957]: DEBUG nova.compute.manager [-] [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 575.530992] env[61957]: DEBUG nova.network.neutron [-] [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 575.540815] env[61957]: DEBUG nova.compute.provider_tree [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 575.572020] env[61957]: DEBUG nova.network.neutron [-] [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 575.840296] env[61957]: ERROR nova.compute.manager [None req-d102ca19-392e-4051-87d1-b58bed2da29d tempest-ServersAdminNegativeTestJSON-555579 tempest-ServersAdminNegativeTestJSON-555579-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 178d7697-f823-4565-9fc6-706777f5e312, please check neutron logs for more information. [ 575.840296] env[61957]: ERROR nova.compute.manager Traceback (most recent call last): [ 575.840296] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 575.840296] env[61957]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 575.840296] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 575.840296] env[61957]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 575.840296] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 575.840296] env[61957]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 575.840296] env[61957]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 575.840296] env[61957]: ERROR nova.compute.manager self.force_reraise() [ 575.840296] env[61957]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 575.840296] env[61957]: ERROR nova.compute.manager raise self.value [ 575.840296] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 575.840296] env[61957]: ERROR nova.compute.manager updated_port = self._update_port( [ 575.840296] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 575.840296] env[61957]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 575.840996] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 575.840996] env[61957]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 575.840996] env[61957]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 178d7697-f823-4565-9fc6-706777f5e312, please check neutron logs for more information. [ 575.840996] env[61957]: ERROR nova.compute.manager [ 575.840996] env[61957]: Traceback (most recent call last): [ 575.840996] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 575.840996] env[61957]: listener.cb(fileno) [ 575.840996] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 575.840996] env[61957]: result = function(*args, **kwargs) [ 575.840996] env[61957]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 575.840996] env[61957]: return func(*args, **kwargs) [ 575.840996] env[61957]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 575.840996] env[61957]: raise e [ 575.840996] env[61957]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 575.840996] env[61957]: nwinfo = self.network_api.allocate_for_instance( [ 575.840996] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 575.840996] env[61957]: created_port_ids = self._update_ports_for_instance( [ 575.840996] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 575.840996] env[61957]: with excutils.save_and_reraise_exception(): [ 575.840996] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 575.840996] env[61957]: self.force_reraise() [ 575.840996] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 575.840996] env[61957]: raise self.value [ 575.840996] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 575.840996] env[61957]: updated_port = self._update_port( [ 575.840996] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 575.840996] env[61957]: _ensure_no_port_binding_failure(port) [ 575.840996] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 575.840996] env[61957]: raise exception.PortBindingFailed(port_id=port['id']) [ 575.842131] env[61957]: nova.exception.PortBindingFailed: Binding failed for port 178d7697-f823-4565-9fc6-706777f5e312, please check neutron logs for more information. [ 575.842131] env[61957]: Removing descriptor: 17 [ 575.842131] env[61957]: ERROR nova.compute.manager [None req-d102ca19-392e-4051-87d1-b58bed2da29d tempest-ServersAdminNegativeTestJSON-555579 tempest-ServersAdminNegativeTestJSON-555579-project-member] [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 178d7697-f823-4565-9fc6-706777f5e312, please check neutron logs for more information. [ 575.842131] env[61957]: ERROR nova.compute.manager [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] Traceback (most recent call last): [ 575.842131] env[61957]: ERROR nova.compute.manager [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 575.842131] env[61957]: ERROR nova.compute.manager [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] yield resources [ 575.842131] env[61957]: ERROR nova.compute.manager [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 575.842131] env[61957]: ERROR nova.compute.manager [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] self.driver.spawn(context, instance, image_meta, [ 575.842131] env[61957]: ERROR nova.compute.manager [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 575.842131] env[61957]: ERROR nova.compute.manager [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] self._vmops.spawn(context, instance, image_meta, injected_files, [ 575.842131] env[61957]: ERROR nova.compute.manager [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 575.842131] env[61957]: ERROR nova.compute.manager [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] vm_ref = self.build_virtual_machine(instance, [ 575.843184] env[61957]: ERROR nova.compute.manager [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 575.843184] env[61957]: ERROR nova.compute.manager [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] vif_infos = vmwarevif.get_vif_info(self._session, [ 575.843184] env[61957]: ERROR nova.compute.manager [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 575.843184] env[61957]: ERROR nova.compute.manager [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] for vif in network_info: [ 575.843184] env[61957]: ERROR nova.compute.manager [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 575.843184] env[61957]: ERROR nova.compute.manager [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] return self._sync_wrapper(fn, *args, **kwargs) [ 575.843184] env[61957]: ERROR nova.compute.manager [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 575.843184] env[61957]: ERROR nova.compute.manager [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] self.wait() [ 575.843184] env[61957]: ERROR nova.compute.manager [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 575.843184] env[61957]: ERROR nova.compute.manager [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] self[:] = self._gt.wait() [ 575.843184] env[61957]: ERROR nova.compute.manager [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 575.843184] env[61957]: ERROR nova.compute.manager [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] return self._exit_event.wait() [ 575.843184] env[61957]: ERROR nova.compute.manager [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 575.843578] env[61957]: ERROR nova.compute.manager [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] result = hub.switch() [ 575.843578] env[61957]: ERROR nova.compute.manager [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 575.843578] env[61957]: ERROR nova.compute.manager [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] return self.greenlet.switch() [ 575.843578] env[61957]: ERROR nova.compute.manager [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 575.843578] env[61957]: ERROR nova.compute.manager [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] result = function(*args, **kwargs) [ 575.843578] env[61957]: ERROR nova.compute.manager [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 575.843578] env[61957]: ERROR nova.compute.manager [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] return func(*args, **kwargs) [ 575.843578] env[61957]: ERROR nova.compute.manager [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 575.843578] env[61957]: ERROR nova.compute.manager [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] raise e [ 575.843578] env[61957]: ERROR nova.compute.manager [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 575.843578] env[61957]: ERROR nova.compute.manager [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] nwinfo = self.network_api.allocate_for_instance( [ 575.843578] env[61957]: ERROR nova.compute.manager [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 575.843578] env[61957]: ERROR nova.compute.manager [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] created_port_ids = self._update_ports_for_instance( [ 575.844904] env[61957]: ERROR nova.compute.manager [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 575.844904] env[61957]: ERROR nova.compute.manager [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] with excutils.save_and_reraise_exception(): [ 575.844904] env[61957]: ERROR nova.compute.manager [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 575.844904] env[61957]: ERROR nova.compute.manager [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] self.force_reraise() [ 575.844904] env[61957]: ERROR nova.compute.manager [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 575.844904] env[61957]: ERROR nova.compute.manager [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] raise self.value [ 575.844904] env[61957]: ERROR nova.compute.manager [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 575.844904] env[61957]: ERROR nova.compute.manager [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] updated_port = self._update_port( [ 575.844904] env[61957]: ERROR nova.compute.manager [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 575.844904] env[61957]: ERROR nova.compute.manager [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] _ensure_no_port_binding_failure(port) [ 575.844904] env[61957]: ERROR nova.compute.manager [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 575.844904] env[61957]: ERROR nova.compute.manager [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] raise exception.PortBindingFailed(port_id=port['id']) [ 575.846298] env[61957]: ERROR nova.compute.manager [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] nova.exception.PortBindingFailed: Binding failed for port 178d7697-f823-4565-9fc6-706777f5e312, please check neutron logs for more information. [ 575.846298] env[61957]: ERROR nova.compute.manager [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] [ 575.846298] env[61957]: INFO nova.compute.manager [None req-d102ca19-392e-4051-87d1-b58bed2da29d tempest-ServersAdminNegativeTestJSON-555579 tempest-ServersAdminNegativeTestJSON-555579-project-member] [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] Terminating instance [ 575.846298] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d102ca19-392e-4051-87d1-b58bed2da29d tempest-ServersAdminNegativeTestJSON-555579 tempest-ServersAdminNegativeTestJSON-555579-project-member] Acquiring lock "refresh_cache-37837096-5ad0-4b6f-b32d-f2899d88b6dd" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 575.846298] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d102ca19-392e-4051-87d1-b58bed2da29d tempest-ServersAdminNegativeTestJSON-555579 tempest-ServersAdminNegativeTestJSON-555579-project-member] Acquired lock "refresh_cache-37837096-5ad0-4b6f-b32d-f2899d88b6dd" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 575.848090] env[61957]: DEBUG nova.network.neutron [None req-d102ca19-392e-4051-87d1-b58bed2da29d tempest-ServersAdminNegativeTestJSON-555579 tempest-ServersAdminNegativeTestJSON-555579-project-member] [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 575.872825] env[61957]: DEBUG nova.compute.manager [None req-4b8bfc5c-9cfa-4b72-9c9f-71ae820b9856 tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] [instance: c5af147e-3526-4014-98bc-7ad163ae89d9] Starting instance... {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 576.046015] env[61957]: DEBUG nova.scheduler.client.report [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 576.075524] env[61957]: DEBUG nova.network.neutron [-] [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 576.371362] env[61957]: DEBUG nova.network.neutron [None req-d102ca19-392e-4051-87d1-b58bed2da29d tempest-ServersAdminNegativeTestJSON-555579 tempest-ServersAdminNegativeTestJSON-555579-project-member] [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 576.382970] env[61957]: DEBUG oslo_concurrency.lockutils [None req-840b2d8a-f610-4c89-a367-da388d81d5aa tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] Acquiring lock "f9e04dbd-d5ae-4f03-9406-14b2caeaf010" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 576.383881] env[61957]: DEBUG oslo_concurrency.lockutils [None req-840b2d8a-f610-4c89-a367-da388d81d5aa tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] Lock "f9e04dbd-d5ae-4f03-9406-14b2caeaf010" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 576.409697] env[61957]: DEBUG oslo_concurrency.lockutils [None req-4b8bfc5c-9cfa-4b72-9c9f-71ae820b9856 tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 576.498153] env[61957]: DEBUG nova.network.neutron [None req-d102ca19-392e-4051-87d1-b58bed2da29d tempest-ServersAdminNegativeTestJSON-555579 tempest-ServersAdminNegativeTestJSON-555579-project-member] [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 576.554566] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61957) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 576.554566] env[61957]: DEBUG oslo_concurrency.lockutils [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.305s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 576.555345] env[61957]: DEBUG oslo_concurrency.lockutils [None req-837238f2-cf29-45f6-8541-231ac85b718c tempest-ServerDiagnosticsTest-1677027478 tempest-ServerDiagnosticsTest-1677027478-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.599s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 576.557092] env[61957]: INFO nova.compute.claims [None req-837238f2-cf29-45f6-8541-231ac85b718c tempest-ServerDiagnosticsTest-1677027478 tempest-ServerDiagnosticsTest-1677027478-project-member] [instance: 43942e78-7703-444d-91e3-bc7913299654] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 576.578296] env[61957]: INFO nova.compute.manager [-] [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] Took 1.05 seconds to deallocate network for instance. [ 576.583120] env[61957]: DEBUG nova.compute.claims [None req-d3a5f917-ae8e-4fd7-b4eb-cbb6e49d67de tempest-ServerDiagnosticsNegativeTest-1621277291 tempest-ServerDiagnosticsNegativeTest-1621277291-project-member] [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] Aborting claim: {{(pid=61957) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 576.583120] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d3a5f917-ae8e-4fd7-b4eb-cbb6e49d67de tempest-ServerDiagnosticsNegativeTest-1621277291 tempest-ServerDiagnosticsNegativeTest-1621277291-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 576.888072] env[61957]: DEBUG nova.compute.manager [None req-840b2d8a-f610-4c89-a367-da388d81d5aa tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] Starting instance... {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 577.001143] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d102ca19-392e-4051-87d1-b58bed2da29d tempest-ServersAdminNegativeTestJSON-555579 tempest-ServersAdminNegativeTestJSON-555579-project-member] Releasing lock "refresh_cache-37837096-5ad0-4b6f-b32d-f2899d88b6dd" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 577.001572] env[61957]: DEBUG nova.compute.manager [None req-d102ca19-392e-4051-87d1-b58bed2da29d tempest-ServersAdminNegativeTestJSON-555579 tempest-ServersAdminNegativeTestJSON-555579-project-member] [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] Start destroying the instance on the hypervisor. {{(pid=61957) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 577.001760] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-d102ca19-392e-4051-87d1-b58bed2da29d tempest-ServersAdminNegativeTestJSON-555579 tempest-ServersAdminNegativeTestJSON-555579-project-member] [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] Destroying instance {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 577.002057] env[61957]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-06e9177f-511c-4a59-bcf3-ece6b0665b46 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.014132] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5460bda0-9cb8-4410-b184-a5741687462c {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.040973] env[61957]: WARNING nova.virt.vmwareapi.vmops [None req-d102ca19-392e-4051-87d1-b58bed2da29d tempest-ServersAdminNegativeTestJSON-555579 tempest-ServersAdminNegativeTestJSON-555579-project-member] [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 37837096-5ad0-4b6f-b32d-f2899d88b6dd could not be found. [ 577.041234] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-d102ca19-392e-4051-87d1-b58bed2da29d tempest-ServersAdminNegativeTestJSON-555579 tempest-ServersAdminNegativeTestJSON-555579-project-member] [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] Instance destroyed {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 577.041408] env[61957]: INFO nova.compute.manager [None req-d102ca19-392e-4051-87d1-b58bed2da29d tempest-ServersAdminNegativeTestJSON-555579 tempest-ServersAdminNegativeTestJSON-555579-project-member] [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] Took 0.04 seconds to destroy the instance on the hypervisor. [ 577.041649] env[61957]: DEBUG oslo.service.loopingcall [None req-d102ca19-392e-4051-87d1-b58bed2da29d tempest-ServersAdminNegativeTestJSON-555579 tempest-ServersAdminNegativeTestJSON-555579-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 577.041855] env[61957]: DEBUG nova.compute.manager [-] [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 577.041969] env[61957]: DEBUG nova.network.neutron [-] [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 577.087683] env[61957]: ERROR nova.compute.manager [None req-d83719b8-6bd9-4ae7-99ea-7d0970401ba5 tempest-ServerExternalEventsTest-1886420864 tempest-ServerExternalEventsTest-1886420864-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 3d42e5dd-f713-445a-8855-3e0e765e2c4a, please check neutron logs for more information. [ 577.087683] env[61957]: ERROR nova.compute.manager Traceback (most recent call last): [ 577.087683] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 577.087683] env[61957]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 577.087683] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 577.087683] env[61957]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 577.087683] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 577.087683] env[61957]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 577.087683] env[61957]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 577.087683] env[61957]: ERROR nova.compute.manager self.force_reraise() [ 577.087683] env[61957]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 577.087683] env[61957]: ERROR nova.compute.manager raise self.value [ 577.087683] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 577.087683] env[61957]: ERROR nova.compute.manager updated_port = self._update_port( [ 577.087683] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 577.087683] env[61957]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 577.088540] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 577.088540] env[61957]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 577.088540] env[61957]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 3d42e5dd-f713-445a-8855-3e0e765e2c4a, please check neutron logs for more information. [ 577.088540] env[61957]: ERROR nova.compute.manager [ 577.088540] env[61957]: Traceback (most recent call last): [ 577.088540] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 577.088540] env[61957]: listener.cb(fileno) [ 577.088540] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 577.088540] env[61957]: result = function(*args, **kwargs) [ 577.088540] env[61957]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 577.088540] env[61957]: return func(*args, **kwargs) [ 577.088540] env[61957]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 577.088540] env[61957]: raise e [ 577.088540] env[61957]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 577.088540] env[61957]: nwinfo = self.network_api.allocate_for_instance( [ 577.088540] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 577.088540] env[61957]: created_port_ids = self._update_ports_for_instance( [ 577.088540] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 577.088540] env[61957]: with excutils.save_and_reraise_exception(): [ 577.088540] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 577.088540] env[61957]: self.force_reraise() [ 577.088540] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 577.088540] env[61957]: raise self.value [ 577.088540] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 577.088540] env[61957]: updated_port = self._update_port( [ 577.088540] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 577.088540] env[61957]: _ensure_no_port_binding_failure(port) [ 577.088540] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 577.088540] env[61957]: raise exception.PortBindingFailed(port_id=port['id']) [ 577.089383] env[61957]: nova.exception.PortBindingFailed: Binding failed for port 3d42e5dd-f713-445a-8855-3e0e765e2c4a, please check neutron logs for more information. [ 577.089383] env[61957]: Removing descriptor: 15 [ 577.089383] env[61957]: ERROR nova.compute.manager [None req-d83719b8-6bd9-4ae7-99ea-7d0970401ba5 tempest-ServerExternalEventsTest-1886420864 tempest-ServerExternalEventsTest-1886420864-project-member] [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 3d42e5dd-f713-445a-8855-3e0e765e2c4a, please check neutron logs for more information. [ 577.089383] env[61957]: ERROR nova.compute.manager [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] Traceback (most recent call last): [ 577.089383] env[61957]: ERROR nova.compute.manager [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 577.089383] env[61957]: ERROR nova.compute.manager [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] yield resources [ 577.089383] env[61957]: ERROR nova.compute.manager [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 577.089383] env[61957]: ERROR nova.compute.manager [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] self.driver.spawn(context, instance, image_meta, [ 577.089383] env[61957]: ERROR nova.compute.manager [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 577.089383] env[61957]: ERROR nova.compute.manager [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] self._vmops.spawn(context, instance, image_meta, injected_files, [ 577.089383] env[61957]: ERROR nova.compute.manager [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 577.089383] env[61957]: ERROR nova.compute.manager [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] vm_ref = self.build_virtual_machine(instance, [ 577.091989] env[61957]: ERROR nova.compute.manager [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 577.091989] env[61957]: ERROR nova.compute.manager [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] vif_infos = vmwarevif.get_vif_info(self._session, [ 577.091989] env[61957]: ERROR nova.compute.manager [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 577.091989] env[61957]: ERROR nova.compute.manager [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] for vif in network_info: [ 577.091989] env[61957]: ERROR nova.compute.manager [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 577.091989] env[61957]: ERROR nova.compute.manager [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] return self._sync_wrapper(fn, *args, **kwargs) [ 577.091989] env[61957]: ERROR nova.compute.manager [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 577.091989] env[61957]: ERROR nova.compute.manager [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] self.wait() [ 577.091989] env[61957]: ERROR nova.compute.manager [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 577.091989] env[61957]: ERROR nova.compute.manager [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] self[:] = self._gt.wait() [ 577.091989] env[61957]: ERROR nova.compute.manager [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 577.091989] env[61957]: ERROR nova.compute.manager [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] return self._exit_event.wait() [ 577.091989] env[61957]: ERROR nova.compute.manager [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 577.093029] env[61957]: ERROR nova.compute.manager [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] result = hub.switch() [ 577.093029] env[61957]: ERROR nova.compute.manager [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 577.093029] env[61957]: ERROR nova.compute.manager [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] return self.greenlet.switch() [ 577.093029] env[61957]: ERROR nova.compute.manager [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 577.093029] env[61957]: ERROR nova.compute.manager [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] result = function(*args, **kwargs) [ 577.093029] env[61957]: ERROR nova.compute.manager [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 577.093029] env[61957]: ERROR nova.compute.manager [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] return func(*args, **kwargs) [ 577.093029] env[61957]: ERROR nova.compute.manager [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 577.093029] env[61957]: ERROR nova.compute.manager [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] raise e [ 577.093029] env[61957]: ERROR nova.compute.manager [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 577.093029] env[61957]: ERROR nova.compute.manager [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] nwinfo = self.network_api.allocate_for_instance( [ 577.093029] env[61957]: ERROR nova.compute.manager [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 577.093029] env[61957]: ERROR nova.compute.manager [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] created_port_ids = self._update_ports_for_instance( [ 577.094042] env[61957]: ERROR nova.compute.manager [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 577.094042] env[61957]: ERROR nova.compute.manager [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] with excutils.save_and_reraise_exception(): [ 577.094042] env[61957]: ERROR nova.compute.manager [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 577.094042] env[61957]: ERROR nova.compute.manager [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] self.force_reraise() [ 577.094042] env[61957]: ERROR nova.compute.manager [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 577.094042] env[61957]: ERROR nova.compute.manager [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] raise self.value [ 577.094042] env[61957]: ERROR nova.compute.manager [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 577.094042] env[61957]: ERROR nova.compute.manager [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] updated_port = self._update_port( [ 577.094042] env[61957]: ERROR nova.compute.manager [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 577.094042] env[61957]: ERROR nova.compute.manager [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] _ensure_no_port_binding_failure(port) [ 577.094042] env[61957]: ERROR nova.compute.manager [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 577.094042] env[61957]: ERROR nova.compute.manager [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] raise exception.PortBindingFailed(port_id=port['id']) [ 577.095039] env[61957]: ERROR nova.compute.manager [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] nova.exception.PortBindingFailed: Binding failed for port 3d42e5dd-f713-445a-8855-3e0e765e2c4a, please check neutron logs for more information. [ 577.095039] env[61957]: ERROR nova.compute.manager [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] [ 577.095039] env[61957]: INFO nova.compute.manager [None req-d83719b8-6bd9-4ae7-99ea-7d0970401ba5 tempest-ServerExternalEventsTest-1886420864 tempest-ServerExternalEventsTest-1886420864-project-member] [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] Terminating instance [ 577.095039] env[61957]: DEBUG nova.network.neutron [-] [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 577.095039] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d83719b8-6bd9-4ae7-99ea-7d0970401ba5 tempest-ServerExternalEventsTest-1886420864 tempest-ServerExternalEventsTest-1886420864-project-member] Acquiring lock "refresh_cache-abdb0ba9-abef-4528-96fb-ae7d7d41f8fc" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 577.095039] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d83719b8-6bd9-4ae7-99ea-7d0970401ba5 tempest-ServerExternalEventsTest-1886420864 tempest-ServerExternalEventsTest-1886420864-project-member] Acquired lock "refresh_cache-abdb0ba9-abef-4528-96fb-ae7d7d41f8fc" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 577.095039] env[61957]: DEBUG nova.network.neutron [None req-d83719b8-6bd9-4ae7-99ea-7d0970401ba5 tempest-ServerExternalEventsTest-1886420864 tempest-ServerExternalEventsTest-1886420864-project-member] [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 577.307138] env[61957]: DEBUG nova.compute.manager [req-3160967d-05da-46f7-8f17-bb2f4f958f6a req-22a23d0f-676f-4ddb-9914-ae42527b9c10 service nova] [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] Received event network-vif-deleted-6bf15563-9c78-4c42-9a95-c660f6c0e3b5 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 577.308816] env[61957]: DEBUG nova.compute.manager [req-3160967d-05da-46f7-8f17-bb2f4f958f6a req-22a23d0f-676f-4ddb-9914-ae42527b9c10 service nova] [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] Received event network-changed-c759a216-a6b5-4bff-97b2-bdfda7cbc7d5 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 577.308816] env[61957]: DEBUG nova.compute.manager [req-3160967d-05da-46f7-8f17-bb2f4f958f6a req-22a23d0f-676f-4ddb-9914-ae42527b9c10 service nova] [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] Refreshing instance network info cache due to event network-changed-c759a216-a6b5-4bff-97b2-bdfda7cbc7d5. {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 577.308816] env[61957]: DEBUG oslo_concurrency.lockutils [req-3160967d-05da-46f7-8f17-bb2f4f958f6a req-22a23d0f-676f-4ddb-9914-ae42527b9c10 service nova] Acquiring lock "refresh_cache-f28611df-9528-4ec0-b575-ad54ebd4a958" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 577.308816] env[61957]: DEBUG oslo_concurrency.lockutils [req-3160967d-05da-46f7-8f17-bb2f4f958f6a req-22a23d0f-676f-4ddb-9914-ae42527b9c10 service nova] Acquired lock "refresh_cache-f28611df-9528-4ec0-b575-ad54ebd4a958" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 577.308816] env[61957]: DEBUG nova.network.neutron [req-3160967d-05da-46f7-8f17-bb2f4f958f6a req-22a23d0f-676f-4ddb-9914-ae42527b9c10 service nova] [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] Refreshing network info cache for port c759a216-a6b5-4bff-97b2-bdfda7cbc7d5 {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 577.422034] env[61957]: DEBUG oslo_concurrency.lockutils [None req-840b2d8a-f610-4c89-a367-da388d81d5aa tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 577.595017] env[61957]: DEBUG nova.network.neutron [-] [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 577.628540] env[61957]: DEBUG nova.network.neutron [None req-d83719b8-6bd9-4ae7-99ea-7d0970401ba5 tempest-ServerExternalEventsTest-1886420864 tempest-ServerExternalEventsTest-1886420864-project-member] [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 577.708960] env[61957]: DEBUG nova.network.neutron [None req-d83719b8-6bd9-4ae7-99ea-7d0970401ba5 tempest-ServerExternalEventsTest-1886420864 tempest-ServerExternalEventsTest-1886420864-project-member] [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 577.748743] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46b6949c-0024-43ca-8947-db97525329a4 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.757288] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a96b0938-16c7-43ea-819c-913f511129b5 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.793192] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25c61a0e-2ba9-42a6-ae1b-f6efd9b527b5 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.801946] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c5030e7-494d-4057-b9a4-5cc28a0f9b90 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.820966] env[61957]: DEBUG nova.compute.provider_tree [None req-837238f2-cf29-45f6-8541-231ac85b718c tempest-ServerDiagnosticsTest-1677027478 tempest-ServerDiagnosticsTest-1677027478-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 577.838999] env[61957]: DEBUG nova.network.neutron [req-3160967d-05da-46f7-8f17-bb2f4f958f6a req-22a23d0f-676f-4ddb-9914-ae42527b9c10 service nova] [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 577.934032] env[61957]: DEBUG nova.network.neutron [req-3160967d-05da-46f7-8f17-bb2f4f958f6a req-22a23d0f-676f-4ddb-9914-ae42527b9c10 service nova] [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 578.100513] env[61957]: INFO nova.compute.manager [-] [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] Took 1.06 seconds to deallocate network for instance. [ 578.108071] env[61957]: DEBUG nova.compute.claims [None req-d102ca19-392e-4051-87d1-b58bed2da29d tempest-ServersAdminNegativeTestJSON-555579 tempest-ServersAdminNegativeTestJSON-555579-project-member] [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] Aborting claim: {{(pid=61957) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 578.108071] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d102ca19-392e-4051-87d1-b58bed2da29d tempest-ServersAdminNegativeTestJSON-555579 tempest-ServersAdminNegativeTestJSON-555579-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 578.213728] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d83719b8-6bd9-4ae7-99ea-7d0970401ba5 tempest-ServerExternalEventsTest-1886420864 tempest-ServerExternalEventsTest-1886420864-project-member] Releasing lock "refresh_cache-abdb0ba9-abef-4528-96fb-ae7d7d41f8fc" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 578.214722] env[61957]: DEBUG nova.compute.manager [None req-d83719b8-6bd9-4ae7-99ea-7d0970401ba5 tempest-ServerExternalEventsTest-1886420864 tempest-ServerExternalEventsTest-1886420864-project-member] [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] Start destroying the instance on the hypervisor. {{(pid=61957) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 578.214861] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-d83719b8-6bd9-4ae7-99ea-7d0970401ba5 tempest-ServerExternalEventsTest-1886420864 tempest-ServerExternalEventsTest-1886420864-project-member] [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] Destroying instance {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 578.215176] env[61957]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-eeb965df-9702-4b96-a6c1-70d7c59eedfe {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.226393] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe62e36d-bd65-4f23-9749-89bc150902b2 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.254222] env[61957]: WARNING nova.virt.vmwareapi.vmops [None req-d83719b8-6bd9-4ae7-99ea-7d0970401ba5 tempest-ServerExternalEventsTest-1886420864 tempest-ServerExternalEventsTest-1886420864-project-member] [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance abdb0ba9-abef-4528-96fb-ae7d7d41f8fc could not be found. [ 578.255299] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-d83719b8-6bd9-4ae7-99ea-7d0970401ba5 tempest-ServerExternalEventsTest-1886420864 tempest-ServerExternalEventsTest-1886420864-project-member] [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] Instance destroyed {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 578.255299] env[61957]: INFO nova.compute.manager [None req-d83719b8-6bd9-4ae7-99ea-7d0970401ba5 tempest-ServerExternalEventsTest-1886420864 tempest-ServerExternalEventsTest-1886420864-project-member] [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] Took 0.04 seconds to destroy the instance on the hypervisor. [ 578.255299] env[61957]: DEBUG oslo.service.loopingcall [None req-d83719b8-6bd9-4ae7-99ea-7d0970401ba5 tempest-ServerExternalEventsTest-1886420864 tempest-ServerExternalEventsTest-1886420864-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 578.255299] env[61957]: DEBUG nova.compute.manager [-] [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 578.255299] env[61957]: DEBUG nova.network.neutron [-] [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 578.289393] env[61957]: DEBUG nova.network.neutron [-] [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 578.324781] env[61957]: DEBUG nova.scheduler.client.report [None req-837238f2-cf29-45f6-8541-231ac85b718c tempest-ServerDiagnosticsTest-1677027478 tempest-ServerDiagnosticsTest-1677027478-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 578.439918] env[61957]: DEBUG oslo_concurrency.lockutils [req-3160967d-05da-46f7-8f17-bb2f4f958f6a req-22a23d0f-676f-4ddb-9914-ae42527b9c10 service nova] Releasing lock "refresh_cache-f28611df-9528-4ec0-b575-ad54ebd4a958" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 578.440241] env[61957]: DEBUG nova.compute.manager [req-3160967d-05da-46f7-8f17-bb2f4f958f6a req-22a23d0f-676f-4ddb-9914-ae42527b9c10 service nova] [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] Received event network-vif-deleted-c759a216-a6b5-4bff-97b2-bdfda7cbc7d5 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 578.440309] env[61957]: DEBUG nova.compute.manager [req-3160967d-05da-46f7-8f17-bb2f4f958f6a req-22a23d0f-676f-4ddb-9914-ae42527b9c10 service nova] [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] Received event network-changed-178d7697-f823-4565-9fc6-706777f5e312 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 578.440622] env[61957]: DEBUG nova.compute.manager [req-3160967d-05da-46f7-8f17-bb2f4f958f6a req-22a23d0f-676f-4ddb-9914-ae42527b9c10 service nova] [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] Refreshing instance network info cache due to event network-changed-178d7697-f823-4565-9fc6-706777f5e312. {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 578.440705] env[61957]: DEBUG oslo_concurrency.lockutils [req-3160967d-05da-46f7-8f17-bb2f4f958f6a req-22a23d0f-676f-4ddb-9914-ae42527b9c10 service nova] Acquiring lock "refresh_cache-37837096-5ad0-4b6f-b32d-f2899d88b6dd" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 578.440787] env[61957]: DEBUG oslo_concurrency.lockutils [req-3160967d-05da-46f7-8f17-bb2f4f958f6a req-22a23d0f-676f-4ddb-9914-ae42527b9c10 service nova] Acquired lock "refresh_cache-37837096-5ad0-4b6f-b32d-f2899d88b6dd" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 578.441017] env[61957]: DEBUG nova.network.neutron [req-3160967d-05da-46f7-8f17-bb2f4f958f6a req-22a23d0f-676f-4ddb-9914-ae42527b9c10 service nova] [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] Refreshing network info cache for port 178d7697-f823-4565-9fc6-706777f5e312 {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 578.746151] env[61957]: DEBUG oslo_concurrency.lockutils [None req-e4765209-cb68-4173-9918-09c43244c91d tempest-ServerMetadataTestJSON-1199197451 tempest-ServerMetadataTestJSON-1199197451-project-member] Acquiring lock "d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 578.747225] env[61957]: DEBUG oslo_concurrency.lockutils [None req-e4765209-cb68-4173-9918-09c43244c91d tempest-ServerMetadataTestJSON-1199197451 tempest-ServerMetadataTestJSON-1199197451-project-member] Lock "d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.002s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 578.797099] env[61957]: DEBUG nova.network.neutron [-] [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 578.832737] env[61957]: DEBUG oslo_concurrency.lockutils [None req-837238f2-cf29-45f6-8541-231ac85b718c tempest-ServerDiagnosticsTest-1677027478 tempest-ServerDiagnosticsTest-1677027478-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.276s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 578.832737] env[61957]: DEBUG nova.compute.manager [None req-837238f2-cf29-45f6-8541-231ac85b718c tempest-ServerDiagnosticsTest-1677027478 tempest-ServerDiagnosticsTest-1677027478-project-member] [instance: 43942e78-7703-444d-91e3-bc7913299654] Start building networks asynchronously for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 578.834585] env[61957]: DEBUG oslo_concurrency.lockutils [None req-423a0985-b17f-45fa-b4ad-54e09005ad46 tempest-TenantUsagesTestJSON-2011428287 tempest-TenantUsagesTestJSON-2011428287-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.856s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 578.840018] env[61957]: INFO nova.compute.claims [None req-423a0985-b17f-45fa-b4ad-54e09005ad46 tempest-TenantUsagesTestJSON-2011428287 tempest-TenantUsagesTestJSON-2011428287-project-member] [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 579.080435] env[61957]: DEBUG nova.network.neutron [req-3160967d-05da-46f7-8f17-bb2f4f958f6a req-22a23d0f-676f-4ddb-9914-ae42527b9c10 service nova] [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 579.232690] env[61957]: DEBUG nova.network.neutron [req-3160967d-05da-46f7-8f17-bb2f4f958f6a req-22a23d0f-676f-4ddb-9914-ae42527b9c10 service nova] [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 579.250654] env[61957]: DEBUG nova.compute.manager [None req-e4765209-cb68-4173-9918-09c43244c91d tempest-ServerMetadataTestJSON-1199197451 tempest-ServerMetadataTestJSON-1199197451-project-member] [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] Starting instance... {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 579.302357] env[61957]: INFO nova.compute.manager [-] [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] Took 1.05 seconds to deallocate network for instance. [ 579.305605] env[61957]: DEBUG nova.compute.claims [None req-d83719b8-6bd9-4ae7-99ea-7d0970401ba5 tempest-ServerExternalEventsTest-1886420864 tempest-ServerExternalEventsTest-1886420864-project-member] [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] Aborting claim: {{(pid=61957) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 579.305831] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d83719b8-6bd9-4ae7-99ea-7d0970401ba5 tempest-ServerExternalEventsTest-1886420864 tempest-ServerExternalEventsTest-1886420864-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 579.344376] env[61957]: DEBUG nova.compute.utils [None req-837238f2-cf29-45f6-8541-231ac85b718c tempest-ServerDiagnosticsTest-1677027478 tempest-ServerDiagnosticsTest-1677027478-project-member] Using /dev/sd instead of None {{(pid=61957) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 579.349046] env[61957]: DEBUG nova.compute.manager [None req-837238f2-cf29-45f6-8541-231ac85b718c tempest-ServerDiagnosticsTest-1677027478 tempest-ServerDiagnosticsTest-1677027478-project-member] [instance: 43942e78-7703-444d-91e3-bc7913299654] Allocating IP information in the background. {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 579.349046] env[61957]: DEBUG nova.network.neutron [None req-837238f2-cf29-45f6-8541-231ac85b718c tempest-ServerDiagnosticsTest-1677027478 tempest-ServerDiagnosticsTest-1677027478-project-member] [instance: 43942e78-7703-444d-91e3-bc7913299654] allocate_for_instance() {{(pid=61957) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 579.490063] env[61957]: DEBUG nova.policy [None req-837238f2-cf29-45f6-8541-231ac85b718c tempest-ServerDiagnosticsTest-1677027478 tempest-ServerDiagnosticsTest-1677027478-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1e863727661d4313a3b27b964dd39d9a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7252ccfdfcc643a095f21557397891d3', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61957) authorize /opt/stack/nova/nova/policy.py:203}} [ 579.736826] env[61957]: DEBUG oslo_concurrency.lockutils [req-3160967d-05da-46f7-8f17-bb2f4f958f6a req-22a23d0f-676f-4ddb-9914-ae42527b9c10 service nova] Releasing lock "refresh_cache-37837096-5ad0-4b6f-b32d-f2899d88b6dd" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 579.738374] env[61957]: DEBUG nova.compute.manager [req-3160967d-05da-46f7-8f17-bb2f4f958f6a req-22a23d0f-676f-4ddb-9914-ae42527b9c10 service nova] [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] Received event network-vif-deleted-178d7697-f823-4565-9fc6-706777f5e312 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 579.771974] env[61957]: DEBUG oslo_concurrency.lockutils [None req-e4765209-cb68-4173-9918-09c43244c91d tempest-ServerMetadataTestJSON-1199197451 tempest-ServerMetadataTestJSON-1199197451-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 579.851027] env[61957]: DEBUG nova.compute.manager [None req-837238f2-cf29-45f6-8541-231ac85b718c tempest-ServerDiagnosticsTest-1677027478 tempest-ServerDiagnosticsTest-1677027478-project-member] [instance: 43942e78-7703-444d-91e3-bc7913299654] Start building block device mappings for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 580.048369] env[61957]: DEBUG oslo_concurrency.lockutils [None req-1b94927c-57d2-4b99-982c-53a5e69214ea tempest-VolumesAssistedSnapshotsTest-190365324 tempest-VolumesAssistedSnapshotsTest-190365324-project-member] Acquiring lock "ee479d3a-239f-426c-8bdb-484fdf8dcd76" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 580.048606] env[61957]: DEBUG oslo_concurrency.lockutils [None req-1b94927c-57d2-4b99-982c-53a5e69214ea tempest-VolumesAssistedSnapshotsTest-190365324 tempest-VolumesAssistedSnapshotsTest-190365324-project-member] Lock "ee479d3a-239f-426c-8bdb-484fdf8dcd76" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 580.125223] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2671380-03cb-4681-ba04-9684252ca337 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.140117] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e17addc-062a-4de8-a446-23d3c7fca452 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.173490] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-caeae448-6906-4b1c-a1f2-cbe14ff871c7 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.186121] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-163d84bd-1c33-469a-b8ce-9a921d67df31 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.200374] env[61957]: DEBUG nova.compute.provider_tree [None req-423a0985-b17f-45fa-b4ad-54e09005ad46 tempest-TenantUsagesTestJSON-2011428287 tempest-TenantUsagesTestJSON-2011428287-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 580.365059] env[61957]: DEBUG nova.compute.manager [req-8941d92a-35d3-4b43-a1d8-113dd6b28b2a req-2cb12383-97a1-4a05-9332-201ab1c610ab service nova] [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] Received event network-changed-3d42e5dd-f713-445a-8855-3e0e765e2c4a {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 580.365392] env[61957]: DEBUG nova.compute.manager [req-8941d92a-35d3-4b43-a1d8-113dd6b28b2a req-2cb12383-97a1-4a05-9332-201ab1c610ab service nova] [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] Refreshing instance network info cache due to event network-changed-3d42e5dd-f713-445a-8855-3e0e765e2c4a. {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 580.365464] env[61957]: DEBUG oslo_concurrency.lockutils [req-8941d92a-35d3-4b43-a1d8-113dd6b28b2a req-2cb12383-97a1-4a05-9332-201ab1c610ab service nova] Acquiring lock "refresh_cache-abdb0ba9-abef-4528-96fb-ae7d7d41f8fc" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 580.365604] env[61957]: DEBUG oslo_concurrency.lockutils [req-8941d92a-35d3-4b43-a1d8-113dd6b28b2a req-2cb12383-97a1-4a05-9332-201ab1c610ab service nova] Acquired lock "refresh_cache-abdb0ba9-abef-4528-96fb-ae7d7d41f8fc" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 580.365769] env[61957]: DEBUG nova.network.neutron [req-8941d92a-35d3-4b43-a1d8-113dd6b28b2a req-2cb12383-97a1-4a05-9332-201ab1c610ab service nova] [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] Refreshing network info cache for port 3d42e5dd-f713-445a-8855-3e0e765e2c4a {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 580.420620] env[61957]: DEBUG nova.network.neutron [None req-837238f2-cf29-45f6-8541-231ac85b718c tempest-ServerDiagnosticsTest-1677027478 tempest-ServerDiagnosticsTest-1677027478-project-member] [instance: 43942e78-7703-444d-91e3-bc7913299654] Successfully created port: 021f5ba1-78f4-4aa6-b030-243c371db8ac {{(pid=61957) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 580.553662] env[61957]: DEBUG nova.compute.manager [None req-1b94927c-57d2-4b99-982c-53a5e69214ea tempest-VolumesAssistedSnapshotsTest-190365324 tempest-VolumesAssistedSnapshotsTest-190365324-project-member] [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] Starting instance... {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 580.709378] env[61957]: DEBUG nova.scheduler.client.report [None req-423a0985-b17f-45fa-b4ad-54e09005ad46 tempest-TenantUsagesTestJSON-2011428287 tempest-TenantUsagesTestJSON-2011428287-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 580.798526] env[61957]: DEBUG oslo_concurrency.lockutils [None req-61bcf42d-e42b-40a8-98f8-af8cb211ebed tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] Acquiring lock "96c6b747-7293-4252-a1d0-b0d7684c5529" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 580.799424] env[61957]: DEBUG oslo_concurrency.lockutils [None req-61bcf42d-e42b-40a8-98f8-af8cb211ebed tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] Lock "96c6b747-7293-4252-a1d0-b0d7684c5529" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 580.860077] env[61957]: DEBUG nova.compute.manager [None req-837238f2-cf29-45f6-8541-231ac85b718c tempest-ServerDiagnosticsTest-1677027478 tempest-ServerDiagnosticsTest-1677027478-project-member] [instance: 43942e78-7703-444d-91e3-bc7913299654] Start spawning the instance on the hypervisor. {{(pid=61957) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 580.889076] env[61957]: DEBUG nova.virt.hardware [None req-837238f2-cf29-45f6-8541-231ac85b718c tempest-ServerDiagnosticsTest-1677027478 tempest-ServerDiagnosticsTest-1677027478-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T17:22:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T17:21:50Z,direct_url=,disk_format='vmdk',id=11c76a2c-f705-470a-ba9d-4657858bab38,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='81d3096d0f094373913fc3dc8f5c3c6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T17:21:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 580.889391] env[61957]: DEBUG nova.virt.hardware [None req-837238f2-cf29-45f6-8541-231ac85b718c tempest-ServerDiagnosticsTest-1677027478 tempest-ServerDiagnosticsTest-1677027478-project-member] Flavor limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 580.889544] env[61957]: DEBUG nova.virt.hardware [None req-837238f2-cf29-45f6-8541-231ac85b718c tempest-ServerDiagnosticsTest-1677027478 tempest-ServerDiagnosticsTest-1677027478-project-member] Image limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 580.889718] env[61957]: DEBUG nova.virt.hardware [None req-837238f2-cf29-45f6-8541-231ac85b718c tempest-ServerDiagnosticsTest-1677027478 tempest-ServerDiagnosticsTest-1677027478-project-member] Flavor pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 580.889856] env[61957]: DEBUG nova.virt.hardware [None req-837238f2-cf29-45f6-8541-231ac85b718c tempest-ServerDiagnosticsTest-1677027478 tempest-ServerDiagnosticsTest-1677027478-project-member] Image pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 580.890013] env[61957]: DEBUG nova.virt.hardware [None req-837238f2-cf29-45f6-8541-231ac85b718c tempest-ServerDiagnosticsTest-1677027478 tempest-ServerDiagnosticsTest-1677027478-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 580.890275] env[61957]: DEBUG nova.virt.hardware [None req-837238f2-cf29-45f6-8541-231ac85b718c tempest-ServerDiagnosticsTest-1677027478 tempest-ServerDiagnosticsTest-1677027478-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 580.890450] env[61957]: DEBUG nova.virt.hardware [None req-837238f2-cf29-45f6-8541-231ac85b718c tempest-ServerDiagnosticsTest-1677027478 tempest-ServerDiagnosticsTest-1677027478-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 580.890612] env[61957]: DEBUG nova.virt.hardware [None req-837238f2-cf29-45f6-8541-231ac85b718c tempest-ServerDiagnosticsTest-1677027478 tempest-ServerDiagnosticsTest-1677027478-project-member] Got 1 possible topologies {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 580.890766] env[61957]: DEBUG nova.virt.hardware [None req-837238f2-cf29-45f6-8541-231ac85b718c tempest-ServerDiagnosticsTest-1677027478 tempest-ServerDiagnosticsTest-1677027478-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 580.891468] env[61957]: DEBUG nova.virt.hardware [None req-837238f2-cf29-45f6-8541-231ac85b718c tempest-ServerDiagnosticsTest-1677027478 tempest-ServerDiagnosticsTest-1677027478-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 580.891834] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ceb79a1f-7e91-41c7-b830-952756c56137 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.900711] env[61957]: DEBUG nova.network.neutron [req-8941d92a-35d3-4b43-a1d8-113dd6b28b2a req-2cb12383-97a1-4a05-9332-201ab1c610ab service nova] [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 580.905481] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88be38f1-1905-405c-af3e-b1bec223a677 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.028814] env[61957]: DEBUG nova.network.neutron [req-8941d92a-35d3-4b43-a1d8-113dd6b28b2a req-2cb12383-97a1-4a05-9332-201ab1c610ab service nova] [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 581.074972] env[61957]: DEBUG oslo_concurrency.lockutils [None req-1b94927c-57d2-4b99-982c-53a5e69214ea tempest-VolumesAssistedSnapshotsTest-190365324 tempest-VolumesAssistedSnapshotsTest-190365324-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 581.215027] env[61957]: DEBUG oslo_concurrency.lockutils [None req-423a0985-b17f-45fa-b4ad-54e09005ad46 tempest-TenantUsagesTestJSON-2011428287 tempest-TenantUsagesTestJSON-2011428287-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.381s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 581.215607] env[61957]: DEBUG nova.compute.manager [None req-423a0985-b17f-45fa-b4ad-54e09005ad46 tempest-TenantUsagesTestJSON-2011428287 tempest-TenantUsagesTestJSON-2011428287-project-member] [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] Start building networks asynchronously for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 581.219690] env[61957]: DEBUG oslo_concurrency.lockutils [None req-abb0f6e8-eaec-4ec9-a6d4-89648c99b1d8 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 6.068s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 581.458146] env[61957]: DEBUG oslo_concurrency.lockutils [None req-7527ee8e-e60c-46dd-bb63-3e8733859a46 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] Acquiring lock "4ad746d0-4076-4c97-8ea9-20e49fda59ca" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 581.458146] env[61957]: DEBUG oslo_concurrency.lockutils [None req-7527ee8e-e60c-46dd-bb63-3e8733859a46 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] Lock "4ad746d0-4076-4c97-8ea9-20e49fda59ca" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 581.532188] env[61957]: DEBUG oslo_concurrency.lockutils [req-8941d92a-35d3-4b43-a1d8-113dd6b28b2a req-2cb12383-97a1-4a05-9332-201ab1c610ab service nova] Releasing lock "refresh_cache-abdb0ba9-abef-4528-96fb-ae7d7d41f8fc" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 581.532450] env[61957]: DEBUG nova.compute.manager [req-8941d92a-35d3-4b43-a1d8-113dd6b28b2a req-2cb12383-97a1-4a05-9332-201ab1c610ab service nova] [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] Received event network-vif-deleted-3d42e5dd-f713-445a-8855-3e0e765e2c4a {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 581.728379] env[61957]: DEBUG nova.compute.utils [None req-423a0985-b17f-45fa-b4ad-54e09005ad46 tempest-TenantUsagesTestJSON-2011428287 tempest-TenantUsagesTestJSON-2011428287-project-member] Using /dev/sd instead of None {{(pid=61957) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 581.733982] env[61957]: DEBUG nova.compute.manager [None req-423a0985-b17f-45fa-b4ad-54e09005ad46 tempest-TenantUsagesTestJSON-2011428287 tempest-TenantUsagesTestJSON-2011428287-project-member] [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] Allocating IP information in the background. {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 581.733982] env[61957]: DEBUG nova.network.neutron [None req-423a0985-b17f-45fa-b4ad-54e09005ad46 tempest-TenantUsagesTestJSON-2011428287 tempest-TenantUsagesTestJSON-2011428287-project-member] [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] allocate_for_instance() {{(pid=61957) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 581.862812] env[61957]: DEBUG nova.policy [None req-423a0985-b17f-45fa-b4ad-54e09005ad46 tempest-TenantUsagesTestJSON-2011428287 tempest-TenantUsagesTestJSON-2011428287-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1c16304ad0194544942de9cf2f21c58d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3e6bd2c455f84a669efd35ddbdf0530d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61957) authorize /opt/stack/nova/nova/policy.py:203}} [ 581.961426] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08b1c98c-09db-418a-881a-99d848f724f6 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.975180] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2be162f9-6f03-4c18-8318-1d1f2991ea19 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.029068] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-183d54fd-463d-4758-9b96-37c6b889f875 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.038758] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed923d2f-f7d1-43da-838c-a050463c94ab {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.054519] env[61957]: DEBUG nova.compute.provider_tree [None req-abb0f6e8-eaec-4ec9-a6d4-89648c99b1d8 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 582.169973] env[61957]: DEBUG oslo_concurrency.lockutils [None req-1b6e3bde-e95e-4682-9e04-c3f69f3be74f tempest-ServersWithSpecificFlavorTestJSON-568377525 tempest-ServersWithSpecificFlavorTestJSON-568377525-project-member] Acquiring lock "f660c725-fc06-42e4-8a34-5d80568b535e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 582.169973] env[61957]: DEBUG oslo_concurrency.lockutils [None req-1b6e3bde-e95e-4682-9e04-c3f69f3be74f tempest-ServersWithSpecificFlavorTestJSON-568377525 tempest-ServersWithSpecificFlavorTestJSON-568377525-project-member] Lock "f660c725-fc06-42e4-8a34-5d80568b535e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 582.234140] env[61957]: DEBUG nova.compute.manager [None req-423a0985-b17f-45fa-b4ad-54e09005ad46 tempest-TenantUsagesTestJSON-2011428287 tempest-TenantUsagesTestJSON-2011428287-project-member] [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] Start building block device mappings for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 582.382663] env[61957]: DEBUG oslo_concurrency.lockutils [None req-42e67e16-6377-41f9-9501-426c4335c678 tempest-ServersTestBootFromVolume-482289819 tempest-ServersTestBootFromVolume-482289819-project-member] Acquiring lock "32e56693-5100-4ed8-b26a-7d377f9ab7b6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 582.382946] env[61957]: DEBUG oslo_concurrency.lockutils [None req-42e67e16-6377-41f9-9501-426c4335c678 tempest-ServersTestBootFromVolume-482289819 tempest-ServersTestBootFromVolume-482289819-project-member] Lock "32e56693-5100-4ed8-b26a-7d377f9ab7b6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 582.558065] env[61957]: DEBUG nova.scheduler.client.report [None req-abb0f6e8-eaec-4ec9-a6d4-89648c99b1d8 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 582.937127] env[61957]: DEBUG nova.network.neutron [None req-423a0985-b17f-45fa-b4ad-54e09005ad46 tempest-TenantUsagesTestJSON-2011428287 tempest-TenantUsagesTestJSON-2011428287-project-member] [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] Successfully created port: edd470af-41f9-45fb-a01b-74445e4d2f66 {{(pid=61957) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 582.979379] env[61957]: ERROR nova.compute.manager [None req-837238f2-cf29-45f6-8541-231ac85b718c tempest-ServerDiagnosticsTest-1677027478 tempest-ServerDiagnosticsTest-1677027478-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 021f5ba1-78f4-4aa6-b030-243c371db8ac, please check neutron logs for more information. [ 582.979379] env[61957]: ERROR nova.compute.manager Traceback (most recent call last): [ 582.979379] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 582.979379] env[61957]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 582.979379] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 582.979379] env[61957]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 582.979379] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 582.979379] env[61957]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 582.979379] env[61957]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 582.979379] env[61957]: ERROR nova.compute.manager self.force_reraise() [ 582.979379] env[61957]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 582.979379] env[61957]: ERROR nova.compute.manager raise self.value [ 582.979379] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 582.979379] env[61957]: ERROR nova.compute.manager updated_port = self._update_port( [ 582.979379] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 582.979379] env[61957]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 582.979960] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 582.979960] env[61957]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 582.979960] env[61957]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 021f5ba1-78f4-4aa6-b030-243c371db8ac, please check neutron logs for more information. [ 582.979960] env[61957]: ERROR nova.compute.manager [ 582.979960] env[61957]: Traceback (most recent call last): [ 582.979960] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 582.979960] env[61957]: listener.cb(fileno) [ 582.979960] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 582.979960] env[61957]: result = function(*args, **kwargs) [ 582.979960] env[61957]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 582.979960] env[61957]: return func(*args, **kwargs) [ 582.979960] env[61957]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 582.979960] env[61957]: raise e [ 582.979960] env[61957]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 582.979960] env[61957]: nwinfo = self.network_api.allocate_for_instance( [ 582.979960] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 582.979960] env[61957]: created_port_ids = self._update_ports_for_instance( [ 582.979960] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 582.979960] env[61957]: with excutils.save_and_reraise_exception(): [ 582.979960] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 582.979960] env[61957]: self.force_reraise() [ 582.979960] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 582.979960] env[61957]: raise self.value [ 582.979960] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 582.979960] env[61957]: updated_port = self._update_port( [ 582.979960] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 582.979960] env[61957]: _ensure_no_port_binding_failure(port) [ 582.979960] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 582.979960] env[61957]: raise exception.PortBindingFailed(port_id=port['id']) [ 582.981896] env[61957]: nova.exception.PortBindingFailed: Binding failed for port 021f5ba1-78f4-4aa6-b030-243c371db8ac, please check neutron logs for more information. [ 582.981896] env[61957]: Removing descriptor: 17 [ 582.981896] env[61957]: ERROR nova.compute.manager [None req-837238f2-cf29-45f6-8541-231ac85b718c tempest-ServerDiagnosticsTest-1677027478 tempest-ServerDiagnosticsTest-1677027478-project-member] [instance: 43942e78-7703-444d-91e3-bc7913299654] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 021f5ba1-78f4-4aa6-b030-243c371db8ac, please check neutron logs for more information. [ 582.981896] env[61957]: ERROR nova.compute.manager [instance: 43942e78-7703-444d-91e3-bc7913299654] Traceback (most recent call last): [ 582.981896] env[61957]: ERROR nova.compute.manager [instance: 43942e78-7703-444d-91e3-bc7913299654] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 582.981896] env[61957]: ERROR nova.compute.manager [instance: 43942e78-7703-444d-91e3-bc7913299654] yield resources [ 582.981896] env[61957]: ERROR nova.compute.manager [instance: 43942e78-7703-444d-91e3-bc7913299654] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 582.981896] env[61957]: ERROR nova.compute.manager [instance: 43942e78-7703-444d-91e3-bc7913299654] self.driver.spawn(context, instance, image_meta, [ 582.981896] env[61957]: ERROR nova.compute.manager [instance: 43942e78-7703-444d-91e3-bc7913299654] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 582.981896] env[61957]: ERROR nova.compute.manager [instance: 43942e78-7703-444d-91e3-bc7913299654] self._vmops.spawn(context, instance, image_meta, injected_files, [ 582.981896] env[61957]: ERROR nova.compute.manager [instance: 43942e78-7703-444d-91e3-bc7913299654] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 582.981896] env[61957]: ERROR nova.compute.manager [instance: 43942e78-7703-444d-91e3-bc7913299654] vm_ref = self.build_virtual_machine(instance, [ 582.982322] env[61957]: ERROR nova.compute.manager [instance: 43942e78-7703-444d-91e3-bc7913299654] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 582.982322] env[61957]: ERROR nova.compute.manager [instance: 43942e78-7703-444d-91e3-bc7913299654] vif_infos = vmwarevif.get_vif_info(self._session, [ 582.982322] env[61957]: ERROR nova.compute.manager [instance: 43942e78-7703-444d-91e3-bc7913299654] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 582.982322] env[61957]: ERROR nova.compute.manager [instance: 43942e78-7703-444d-91e3-bc7913299654] for vif in network_info: [ 582.982322] env[61957]: ERROR nova.compute.manager [instance: 43942e78-7703-444d-91e3-bc7913299654] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 582.982322] env[61957]: ERROR nova.compute.manager [instance: 43942e78-7703-444d-91e3-bc7913299654] return self._sync_wrapper(fn, *args, **kwargs) [ 582.982322] env[61957]: ERROR nova.compute.manager [instance: 43942e78-7703-444d-91e3-bc7913299654] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 582.982322] env[61957]: ERROR nova.compute.manager [instance: 43942e78-7703-444d-91e3-bc7913299654] self.wait() [ 582.982322] env[61957]: ERROR nova.compute.manager [instance: 43942e78-7703-444d-91e3-bc7913299654] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 582.982322] env[61957]: ERROR nova.compute.manager [instance: 43942e78-7703-444d-91e3-bc7913299654] self[:] = self._gt.wait() [ 582.982322] env[61957]: ERROR nova.compute.manager [instance: 43942e78-7703-444d-91e3-bc7913299654] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 582.982322] env[61957]: ERROR nova.compute.manager [instance: 43942e78-7703-444d-91e3-bc7913299654] return self._exit_event.wait() [ 582.982322] env[61957]: ERROR nova.compute.manager [instance: 43942e78-7703-444d-91e3-bc7913299654] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 582.982736] env[61957]: ERROR nova.compute.manager [instance: 43942e78-7703-444d-91e3-bc7913299654] result = hub.switch() [ 582.982736] env[61957]: ERROR nova.compute.manager [instance: 43942e78-7703-444d-91e3-bc7913299654] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 582.982736] env[61957]: ERROR nova.compute.manager [instance: 43942e78-7703-444d-91e3-bc7913299654] return self.greenlet.switch() [ 582.982736] env[61957]: ERROR nova.compute.manager [instance: 43942e78-7703-444d-91e3-bc7913299654] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 582.982736] env[61957]: ERROR nova.compute.manager [instance: 43942e78-7703-444d-91e3-bc7913299654] result = function(*args, **kwargs) [ 582.982736] env[61957]: ERROR nova.compute.manager [instance: 43942e78-7703-444d-91e3-bc7913299654] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 582.982736] env[61957]: ERROR nova.compute.manager [instance: 43942e78-7703-444d-91e3-bc7913299654] return func(*args, **kwargs) [ 582.982736] env[61957]: ERROR nova.compute.manager [instance: 43942e78-7703-444d-91e3-bc7913299654] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 582.982736] env[61957]: ERROR nova.compute.manager [instance: 43942e78-7703-444d-91e3-bc7913299654] raise e [ 582.982736] env[61957]: ERROR nova.compute.manager [instance: 43942e78-7703-444d-91e3-bc7913299654] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 582.982736] env[61957]: ERROR nova.compute.manager [instance: 43942e78-7703-444d-91e3-bc7913299654] nwinfo = self.network_api.allocate_for_instance( [ 582.982736] env[61957]: ERROR nova.compute.manager [instance: 43942e78-7703-444d-91e3-bc7913299654] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 582.982736] env[61957]: ERROR nova.compute.manager [instance: 43942e78-7703-444d-91e3-bc7913299654] created_port_ids = self._update_ports_for_instance( [ 582.983140] env[61957]: ERROR nova.compute.manager [instance: 43942e78-7703-444d-91e3-bc7913299654] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 582.983140] env[61957]: ERROR nova.compute.manager [instance: 43942e78-7703-444d-91e3-bc7913299654] with excutils.save_and_reraise_exception(): [ 582.983140] env[61957]: ERROR nova.compute.manager [instance: 43942e78-7703-444d-91e3-bc7913299654] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 582.983140] env[61957]: ERROR nova.compute.manager [instance: 43942e78-7703-444d-91e3-bc7913299654] self.force_reraise() [ 582.983140] env[61957]: ERROR nova.compute.manager [instance: 43942e78-7703-444d-91e3-bc7913299654] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 582.983140] env[61957]: ERROR nova.compute.manager [instance: 43942e78-7703-444d-91e3-bc7913299654] raise self.value [ 582.983140] env[61957]: ERROR nova.compute.manager [instance: 43942e78-7703-444d-91e3-bc7913299654] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 582.983140] env[61957]: ERROR nova.compute.manager [instance: 43942e78-7703-444d-91e3-bc7913299654] updated_port = self._update_port( [ 582.983140] env[61957]: ERROR nova.compute.manager [instance: 43942e78-7703-444d-91e3-bc7913299654] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 582.983140] env[61957]: ERROR nova.compute.manager [instance: 43942e78-7703-444d-91e3-bc7913299654] _ensure_no_port_binding_failure(port) [ 582.983140] env[61957]: ERROR nova.compute.manager [instance: 43942e78-7703-444d-91e3-bc7913299654] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 582.983140] env[61957]: ERROR nova.compute.manager [instance: 43942e78-7703-444d-91e3-bc7913299654] raise exception.PortBindingFailed(port_id=port['id']) [ 582.983532] env[61957]: ERROR nova.compute.manager [instance: 43942e78-7703-444d-91e3-bc7913299654] nova.exception.PortBindingFailed: Binding failed for port 021f5ba1-78f4-4aa6-b030-243c371db8ac, please check neutron logs for more information. [ 582.983532] env[61957]: ERROR nova.compute.manager [instance: 43942e78-7703-444d-91e3-bc7913299654] [ 582.983532] env[61957]: INFO nova.compute.manager [None req-837238f2-cf29-45f6-8541-231ac85b718c tempest-ServerDiagnosticsTest-1677027478 tempest-ServerDiagnosticsTest-1677027478-project-member] [instance: 43942e78-7703-444d-91e3-bc7913299654] Terminating instance [ 582.983532] env[61957]: DEBUG oslo_concurrency.lockutils [None req-837238f2-cf29-45f6-8541-231ac85b718c tempest-ServerDiagnosticsTest-1677027478 tempest-ServerDiagnosticsTest-1677027478-project-member] Acquiring lock "refresh_cache-43942e78-7703-444d-91e3-bc7913299654" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 582.983532] env[61957]: DEBUG oslo_concurrency.lockutils [None req-837238f2-cf29-45f6-8541-231ac85b718c tempest-ServerDiagnosticsTest-1677027478 tempest-ServerDiagnosticsTest-1677027478-project-member] Acquired lock "refresh_cache-43942e78-7703-444d-91e3-bc7913299654" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 582.983532] env[61957]: DEBUG nova.network.neutron [None req-837238f2-cf29-45f6-8541-231ac85b718c tempest-ServerDiagnosticsTest-1677027478 tempest-ServerDiagnosticsTest-1677027478-project-member] [instance: 43942e78-7703-444d-91e3-bc7913299654] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 583.066312] env[61957]: DEBUG oslo_concurrency.lockutils [None req-abb0f6e8-eaec-4ec9-a6d4-89648c99b1d8 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.848s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 583.066943] env[61957]: ERROR nova.compute.manager [None req-abb0f6e8-eaec-4ec9-a6d4-89648c99b1d8 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 6bf15563-9c78-4c42-9a95-c660f6c0e3b5, please check neutron logs for more information. [ 583.066943] env[61957]: ERROR nova.compute.manager [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] Traceback (most recent call last): [ 583.066943] env[61957]: ERROR nova.compute.manager [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 583.066943] env[61957]: ERROR nova.compute.manager [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] self.driver.spawn(context, instance, image_meta, [ 583.066943] env[61957]: ERROR nova.compute.manager [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 583.066943] env[61957]: ERROR nova.compute.manager [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] self._vmops.spawn(context, instance, image_meta, injected_files, [ 583.066943] env[61957]: ERROR nova.compute.manager [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 583.066943] env[61957]: ERROR nova.compute.manager [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] vm_ref = self.build_virtual_machine(instance, [ 583.066943] env[61957]: ERROR nova.compute.manager [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 583.066943] env[61957]: ERROR nova.compute.manager [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] vif_infos = vmwarevif.get_vif_info(self._session, [ 583.066943] env[61957]: ERROR nova.compute.manager [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 583.067314] env[61957]: ERROR nova.compute.manager [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] for vif in network_info: [ 583.067314] env[61957]: ERROR nova.compute.manager [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 583.067314] env[61957]: ERROR nova.compute.manager [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] return self._sync_wrapper(fn, *args, **kwargs) [ 583.067314] env[61957]: ERROR nova.compute.manager [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 583.067314] env[61957]: ERROR nova.compute.manager [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] self.wait() [ 583.067314] env[61957]: ERROR nova.compute.manager [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 583.067314] env[61957]: ERROR nova.compute.manager [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] self[:] = self._gt.wait() [ 583.067314] env[61957]: ERROR nova.compute.manager [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 583.067314] env[61957]: ERROR nova.compute.manager [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] return self._exit_event.wait() [ 583.067314] env[61957]: ERROR nova.compute.manager [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 583.067314] env[61957]: ERROR nova.compute.manager [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] result = hub.switch() [ 583.067314] env[61957]: ERROR nova.compute.manager [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 583.067314] env[61957]: ERROR nova.compute.manager [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] return self.greenlet.switch() [ 583.067657] env[61957]: ERROR nova.compute.manager [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 583.067657] env[61957]: ERROR nova.compute.manager [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] result = function(*args, **kwargs) [ 583.067657] env[61957]: ERROR nova.compute.manager [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 583.067657] env[61957]: ERROR nova.compute.manager [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] return func(*args, **kwargs) [ 583.067657] env[61957]: ERROR nova.compute.manager [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 583.067657] env[61957]: ERROR nova.compute.manager [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] raise e [ 583.067657] env[61957]: ERROR nova.compute.manager [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 583.067657] env[61957]: ERROR nova.compute.manager [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] nwinfo = self.network_api.allocate_for_instance( [ 583.067657] env[61957]: ERROR nova.compute.manager [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 583.067657] env[61957]: ERROR nova.compute.manager [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] created_port_ids = self._update_ports_for_instance( [ 583.067657] env[61957]: ERROR nova.compute.manager [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 583.067657] env[61957]: ERROR nova.compute.manager [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] with excutils.save_and_reraise_exception(): [ 583.067657] env[61957]: ERROR nova.compute.manager [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 583.068052] env[61957]: ERROR nova.compute.manager [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] self.force_reraise() [ 583.068052] env[61957]: ERROR nova.compute.manager [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 583.068052] env[61957]: ERROR nova.compute.manager [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] raise self.value [ 583.068052] env[61957]: ERROR nova.compute.manager [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 583.068052] env[61957]: ERROR nova.compute.manager [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] updated_port = self._update_port( [ 583.068052] env[61957]: ERROR nova.compute.manager [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 583.068052] env[61957]: ERROR nova.compute.manager [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] _ensure_no_port_binding_failure(port) [ 583.068052] env[61957]: ERROR nova.compute.manager [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 583.068052] env[61957]: ERROR nova.compute.manager [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] raise exception.PortBindingFailed(port_id=port['id']) [ 583.068052] env[61957]: ERROR nova.compute.manager [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] nova.exception.PortBindingFailed: Binding failed for port 6bf15563-9c78-4c42-9a95-c660f6c0e3b5, please check neutron logs for more information. [ 583.068052] env[61957]: ERROR nova.compute.manager [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] [ 583.068343] env[61957]: DEBUG nova.compute.utils [None req-abb0f6e8-eaec-4ec9-a6d4-89648c99b1d8 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] Binding failed for port 6bf15563-9c78-4c42-9a95-c660f6c0e3b5, please check neutron logs for more information. {{(pid=61957) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 583.068895] env[61957]: DEBUG oslo_concurrency.lockutils [None req-4b8bfc5c-9cfa-4b72-9c9f-71ae820b9856 tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.659s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 583.070382] env[61957]: INFO nova.compute.claims [None req-4b8bfc5c-9cfa-4b72-9c9f-71ae820b9856 tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] [instance: c5af147e-3526-4014-98bc-7ad163ae89d9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 583.077567] env[61957]: DEBUG nova.compute.manager [None req-abb0f6e8-eaec-4ec9-a6d4-89648c99b1d8 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] Build of instance 1a0e117e-aa56-44fa-8818-71f0fd0b6183 was re-scheduled: Binding failed for port 6bf15563-9c78-4c42-9a95-c660f6c0e3b5, please check neutron logs for more information. {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 583.078077] env[61957]: DEBUG nova.compute.manager [None req-abb0f6e8-eaec-4ec9-a6d4-89648c99b1d8 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] Unplugging VIFs for instance {{(pid=61957) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 583.078366] env[61957]: DEBUG oslo_concurrency.lockutils [None req-abb0f6e8-eaec-4ec9-a6d4-89648c99b1d8 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] Acquiring lock "refresh_cache-1a0e117e-aa56-44fa-8818-71f0fd0b6183" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 583.078465] env[61957]: DEBUG oslo_concurrency.lockutils [None req-abb0f6e8-eaec-4ec9-a6d4-89648c99b1d8 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] Acquired lock "refresh_cache-1a0e117e-aa56-44fa-8818-71f0fd0b6183" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 583.078586] env[61957]: DEBUG nova.network.neutron [None req-abb0f6e8-eaec-4ec9-a6d4-89648c99b1d8 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 583.245256] env[61957]: DEBUG nova.compute.manager [None req-423a0985-b17f-45fa-b4ad-54e09005ad46 tempest-TenantUsagesTestJSON-2011428287 tempest-TenantUsagesTestJSON-2011428287-project-member] [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] Start spawning the instance on the hypervisor. {{(pid=61957) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 583.283603] env[61957]: DEBUG nova.virt.hardware [None req-423a0985-b17f-45fa-b4ad-54e09005ad46 tempest-TenantUsagesTestJSON-2011428287 tempest-TenantUsagesTestJSON-2011428287-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T17:22:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T17:21:50Z,direct_url=,disk_format='vmdk',id=11c76a2c-f705-470a-ba9d-4657858bab38,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='81d3096d0f094373913fc3dc8f5c3c6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T17:21:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 583.283903] env[61957]: DEBUG nova.virt.hardware [None req-423a0985-b17f-45fa-b4ad-54e09005ad46 tempest-TenantUsagesTestJSON-2011428287 tempest-TenantUsagesTestJSON-2011428287-project-member] Flavor limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 583.283991] env[61957]: DEBUG nova.virt.hardware [None req-423a0985-b17f-45fa-b4ad-54e09005ad46 tempest-TenantUsagesTestJSON-2011428287 tempest-TenantUsagesTestJSON-2011428287-project-member] Image limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 583.284193] env[61957]: DEBUG nova.virt.hardware [None req-423a0985-b17f-45fa-b4ad-54e09005ad46 tempest-TenantUsagesTestJSON-2011428287 tempest-TenantUsagesTestJSON-2011428287-project-member] Flavor pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 583.284339] env[61957]: DEBUG nova.virt.hardware [None req-423a0985-b17f-45fa-b4ad-54e09005ad46 tempest-TenantUsagesTestJSON-2011428287 tempest-TenantUsagesTestJSON-2011428287-project-member] Image pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 583.284478] env[61957]: DEBUG nova.virt.hardware [None req-423a0985-b17f-45fa-b4ad-54e09005ad46 tempest-TenantUsagesTestJSON-2011428287 tempest-TenantUsagesTestJSON-2011428287-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 583.284882] env[61957]: DEBUG nova.virt.hardware [None req-423a0985-b17f-45fa-b4ad-54e09005ad46 tempest-TenantUsagesTestJSON-2011428287 tempest-TenantUsagesTestJSON-2011428287-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 583.284882] env[61957]: DEBUG nova.virt.hardware [None req-423a0985-b17f-45fa-b4ad-54e09005ad46 tempest-TenantUsagesTestJSON-2011428287 tempest-TenantUsagesTestJSON-2011428287-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 583.285140] env[61957]: DEBUG nova.virt.hardware [None req-423a0985-b17f-45fa-b4ad-54e09005ad46 tempest-TenantUsagesTestJSON-2011428287 tempest-TenantUsagesTestJSON-2011428287-project-member] Got 1 possible topologies {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 583.285359] env[61957]: DEBUG nova.virt.hardware [None req-423a0985-b17f-45fa-b4ad-54e09005ad46 tempest-TenantUsagesTestJSON-2011428287 tempest-TenantUsagesTestJSON-2011428287-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 583.285556] env[61957]: DEBUG nova.virt.hardware [None req-423a0985-b17f-45fa-b4ad-54e09005ad46 tempest-TenantUsagesTestJSON-2011428287 tempest-TenantUsagesTestJSON-2011428287-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 583.286568] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c722936b-3672-43ef-a5e6-3a6b2e29e2c6 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.297402] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b124c17-4da1-42e7-97d0-60d0070170a5 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.355919] env[61957]: DEBUG nova.compute.manager [req-4690bcb9-cece-4c16-9f7d-932dbdbb1531 req-ccace116-45b0-445f-bd39-2eebd84de03e service nova] [instance: 43942e78-7703-444d-91e3-bc7913299654] Received event network-changed-021f5ba1-78f4-4aa6-b030-243c371db8ac {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 583.356230] env[61957]: DEBUG nova.compute.manager [req-4690bcb9-cece-4c16-9f7d-932dbdbb1531 req-ccace116-45b0-445f-bd39-2eebd84de03e service nova] [instance: 43942e78-7703-444d-91e3-bc7913299654] Refreshing instance network info cache due to event network-changed-021f5ba1-78f4-4aa6-b030-243c371db8ac. {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 583.356829] env[61957]: DEBUG oslo_concurrency.lockutils [req-4690bcb9-cece-4c16-9f7d-932dbdbb1531 req-ccace116-45b0-445f-bd39-2eebd84de03e service nova] Acquiring lock "refresh_cache-43942e78-7703-444d-91e3-bc7913299654" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 583.515727] env[61957]: DEBUG nova.network.neutron [None req-837238f2-cf29-45f6-8541-231ac85b718c tempest-ServerDiagnosticsTest-1677027478 tempest-ServerDiagnosticsTest-1677027478-project-member] [instance: 43942e78-7703-444d-91e3-bc7913299654] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 583.620623] env[61957]: DEBUG nova.network.neutron [None req-abb0f6e8-eaec-4ec9-a6d4-89648c99b1d8 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 583.688351] env[61957]: DEBUG nova.network.neutron [None req-837238f2-cf29-45f6-8541-231ac85b718c tempest-ServerDiagnosticsTest-1677027478 tempest-ServerDiagnosticsTest-1677027478-project-member] [instance: 43942e78-7703-444d-91e3-bc7913299654] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 583.755967] env[61957]: DEBUG nova.network.neutron [None req-abb0f6e8-eaec-4ec9-a6d4-89648c99b1d8 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 584.195967] env[61957]: DEBUG oslo_concurrency.lockutils [None req-837238f2-cf29-45f6-8541-231ac85b718c tempest-ServerDiagnosticsTest-1677027478 tempest-ServerDiagnosticsTest-1677027478-project-member] Releasing lock "refresh_cache-43942e78-7703-444d-91e3-bc7913299654" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 584.196515] env[61957]: DEBUG nova.compute.manager [None req-837238f2-cf29-45f6-8541-231ac85b718c tempest-ServerDiagnosticsTest-1677027478 tempest-ServerDiagnosticsTest-1677027478-project-member] [instance: 43942e78-7703-444d-91e3-bc7913299654] Start destroying the instance on the hypervisor. {{(pid=61957) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 584.196788] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-837238f2-cf29-45f6-8541-231ac85b718c tempest-ServerDiagnosticsTest-1677027478 tempest-ServerDiagnosticsTest-1677027478-project-member] [instance: 43942e78-7703-444d-91e3-bc7913299654] Destroying instance {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 584.197381] env[61957]: DEBUG oslo_concurrency.lockutils [req-4690bcb9-cece-4c16-9f7d-932dbdbb1531 req-ccace116-45b0-445f-bd39-2eebd84de03e service nova] Acquired lock "refresh_cache-43942e78-7703-444d-91e3-bc7913299654" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 584.197590] env[61957]: DEBUG nova.network.neutron [req-4690bcb9-cece-4c16-9f7d-932dbdbb1531 req-ccace116-45b0-445f-bd39-2eebd84de03e service nova] [instance: 43942e78-7703-444d-91e3-bc7913299654] Refreshing network info cache for port 021f5ba1-78f4-4aa6-b030-243c371db8ac {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 584.199817] env[61957]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-27ac5227-c374-4ae3-b727-146fc2d272c5 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.212438] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd4b680b-da05-4926-aad8-bf2d910138a4 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.248016] env[61957]: WARNING nova.virt.vmwareapi.vmops [None req-837238f2-cf29-45f6-8541-231ac85b718c tempest-ServerDiagnosticsTest-1677027478 tempest-ServerDiagnosticsTest-1677027478-project-member] [instance: 43942e78-7703-444d-91e3-bc7913299654] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 43942e78-7703-444d-91e3-bc7913299654 could not be found. [ 584.248503] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-837238f2-cf29-45f6-8541-231ac85b718c tempest-ServerDiagnosticsTest-1677027478 tempest-ServerDiagnosticsTest-1677027478-project-member] [instance: 43942e78-7703-444d-91e3-bc7913299654] Instance destroyed {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 584.248965] env[61957]: INFO nova.compute.manager [None req-837238f2-cf29-45f6-8541-231ac85b718c tempest-ServerDiagnosticsTest-1677027478 tempest-ServerDiagnosticsTest-1677027478-project-member] [instance: 43942e78-7703-444d-91e3-bc7913299654] Took 0.05 seconds to destroy the instance on the hypervisor. [ 584.249203] env[61957]: DEBUG oslo.service.loopingcall [None req-837238f2-cf29-45f6-8541-231ac85b718c tempest-ServerDiagnosticsTest-1677027478 tempest-ServerDiagnosticsTest-1677027478-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 584.252150] env[61957]: DEBUG nova.compute.manager [-] [instance: 43942e78-7703-444d-91e3-bc7913299654] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 584.252150] env[61957]: DEBUG nova.network.neutron [-] [instance: 43942e78-7703-444d-91e3-bc7913299654] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 584.259480] env[61957]: DEBUG oslo_concurrency.lockutils [None req-abb0f6e8-eaec-4ec9-a6d4-89648c99b1d8 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] Releasing lock "refresh_cache-1a0e117e-aa56-44fa-8818-71f0fd0b6183" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 584.259480] env[61957]: DEBUG nova.compute.manager [None req-abb0f6e8-eaec-4ec9-a6d4-89648c99b1d8 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61957) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 584.259480] env[61957]: DEBUG nova.compute.manager [None req-abb0f6e8-eaec-4ec9-a6d4-89648c99b1d8 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 584.259480] env[61957]: DEBUG nova.network.neutron [None req-abb0f6e8-eaec-4ec9-a6d4-89648c99b1d8 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 584.326195] env[61957]: DEBUG nova.network.neutron [-] [instance: 43942e78-7703-444d-91e3-bc7913299654] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 584.363347] env[61957]: DEBUG nova.network.neutron [None req-abb0f6e8-eaec-4ec9-a6d4-89648c99b1d8 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 584.387754] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd4b9902-58ab-427f-a856-4b40aaf6bf2a {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.399716] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f6a6023-6583-4c28-ae18-15e911631b17 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.442090] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba63e48d-c6a6-44e9-8040-66a0824649d6 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.452130] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3827a26-4761-4ded-9912-af3c20b1e149 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.469501] env[61957]: DEBUG nova.compute.provider_tree [None req-4b8bfc5c-9cfa-4b72-9c9f-71ae820b9856 tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 584.480896] env[61957]: DEBUG oslo_concurrency.lockutils [None req-260de5ae-9540-404d-a23e-9facd81096ee tempest-AttachInterfacesUnderV243Test-1229304814 tempest-AttachInterfacesUnderV243Test-1229304814-project-member] Acquiring lock "02d215d8-dd42-49cc-9509-08335c7bafe3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 584.481176] env[61957]: DEBUG oslo_concurrency.lockutils [None req-260de5ae-9540-404d-a23e-9facd81096ee tempest-AttachInterfacesUnderV243Test-1229304814 tempest-AttachInterfacesUnderV243Test-1229304814-project-member] Lock "02d215d8-dd42-49cc-9509-08335c7bafe3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 584.749697] env[61957]: DEBUG nova.network.neutron [req-4690bcb9-cece-4c16-9f7d-932dbdbb1531 req-ccace116-45b0-445f-bd39-2eebd84de03e service nova] [instance: 43942e78-7703-444d-91e3-bc7913299654] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 584.829973] env[61957]: DEBUG nova.network.neutron [-] [instance: 43942e78-7703-444d-91e3-bc7913299654] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 584.870988] env[61957]: DEBUG nova.network.neutron [None req-abb0f6e8-eaec-4ec9-a6d4-89648c99b1d8 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 584.949033] env[61957]: DEBUG nova.network.neutron [req-4690bcb9-cece-4c16-9f7d-932dbdbb1531 req-ccace116-45b0-445f-bd39-2eebd84de03e service nova] [instance: 43942e78-7703-444d-91e3-bc7913299654] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 584.973100] env[61957]: DEBUG nova.scheduler.client.report [None req-4b8bfc5c-9cfa-4b72-9c9f-71ae820b9856 tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 585.335673] env[61957]: INFO nova.compute.manager [-] [instance: 43942e78-7703-444d-91e3-bc7913299654] Took 1.08 seconds to deallocate network for instance. [ 585.337707] env[61957]: DEBUG nova.compute.claims [None req-837238f2-cf29-45f6-8541-231ac85b718c tempest-ServerDiagnosticsTest-1677027478 tempest-ServerDiagnosticsTest-1677027478-project-member] [instance: 43942e78-7703-444d-91e3-bc7913299654] Aborting claim: {{(pid=61957) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 585.338361] env[61957]: DEBUG oslo_concurrency.lockutils [None req-837238f2-cf29-45f6-8541-231ac85b718c tempest-ServerDiagnosticsTest-1677027478 tempest-ServerDiagnosticsTest-1677027478-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 585.376548] env[61957]: INFO nova.compute.manager [None req-abb0f6e8-eaec-4ec9-a6d4-89648c99b1d8 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] [instance: 1a0e117e-aa56-44fa-8818-71f0fd0b6183] Took 1.12 seconds to deallocate network for instance. [ 585.452486] env[61957]: DEBUG oslo_concurrency.lockutils [req-4690bcb9-cece-4c16-9f7d-932dbdbb1531 req-ccace116-45b0-445f-bd39-2eebd84de03e service nova] Releasing lock "refresh_cache-43942e78-7703-444d-91e3-bc7913299654" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 585.479605] env[61957]: DEBUG oslo_concurrency.lockutils [None req-4b8bfc5c-9cfa-4b72-9c9f-71ae820b9856 tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.411s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 585.480211] env[61957]: DEBUG nova.compute.manager [None req-4b8bfc5c-9cfa-4b72-9c9f-71ae820b9856 tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] [instance: c5af147e-3526-4014-98bc-7ad163ae89d9] Start building networks asynchronously for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 585.489158] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d3a5f917-ae8e-4fd7-b4eb-cbb6e49d67de tempest-ServerDiagnosticsNegativeTest-1621277291 tempest-ServerDiagnosticsNegativeTest-1621277291-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 8.904s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 585.951418] env[61957]: ERROR nova.compute.manager [None req-423a0985-b17f-45fa-b4ad-54e09005ad46 tempest-TenantUsagesTestJSON-2011428287 tempest-TenantUsagesTestJSON-2011428287-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port edd470af-41f9-45fb-a01b-74445e4d2f66, please check neutron logs for more information. [ 585.951418] env[61957]: ERROR nova.compute.manager Traceback (most recent call last): [ 585.951418] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 585.951418] env[61957]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 585.951418] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 585.951418] env[61957]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 585.951418] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 585.951418] env[61957]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 585.951418] env[61957]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 585.951418] env[61957]: ERROR nova.compute.manager self.force_reraise() [ 585.951418] env[61957]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 585.951418] env[61957]: ERROR nova.compute.manager raise self.value [ 585.951418] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 585.951418] env[61957]: ERROR nova.compute.manager updated_port = self._update_port( [ 585.951418] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 585.951418] env[61957]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 585.953290] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 585.953290] env[61957]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 585.953290] env[61957]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port edd470af-41f9-45fb-a01b-74445e4d2f66, please check neutron logs for more information. [ 585.953290] env[61957]: ERROR nova.compute.manager [ 585.953290] env[61957]: Traceback (most recent call last): [ 585.953290] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 585.953290] env[61957]: listener.cb(fileno) [ 585.953290] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 585.953290] env[61957]: result = function(*args, **kwargs) [ 585.953290] env[61957]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 585.953290] env[61957]: return func(*args, **kwargs) [ 585.953290] env[61957]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 585.953290] env[61957]: raise e [ 585.953290] env[61957]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 585.953290] env[61957]: nwinfo = self.network_api.allocate_for_instance( [ 585.953290] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 585.953290] env[61957]: created_port_ids = self._update_ports_for_instance( [ 585.953290] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 585.953290] env[61957]: with excutils.save_and_reraise_exception(): [ 585.953290] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 585.953290] env[61957]: self.force_reraise() [ 585.953290] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 585.953290] env[61957]: raise self.value [ 585.953290] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 585.953290] env[61957]: updated_port = self._update_port( [ 585.953290] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 585.953290] env[61957]: _ensure_no_port_binding_failure(port) [ 585.953290] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 585.953290] env[61957]: raise exception.PortBindingFailed(port_id=port['id']) [ 585.955064] env[61957]: nova.exception.PortBindingFailed: Binding failed for port edd470af-41f9-45fb-a01b-74445e4d2f66, please check neutron logs for more information. [ 585.955064] env[61957]: Removing descriptor: 15 [ 585.955064] env[61957]: ERROR nova.compute.manager [None req-423a0985-b17f-45fa-b4ad-54e09005ad46 tempest-TenantUsagesTestJSON-2011428287 tempest-TenantUsagesTestJSON-2011428287-project-member] [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port edd470af-41f9-45fb-a01b-74445e4d2f66, please check neutron logs for more information. [ 585.955064] env[61957]: ERROR nova.compute.manager [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] Traceback (most recent call last): [ 585.955064] env[61957]: ERROR nova.compute.manager [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 585.955064] env[61957]: ERROR nova.compute.manager [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] yield resources [ 585.955064] env[61957]: ERROR nova.compute.manager [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 585.955064] env[61957]: ERROR nova.compute.manager [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] self.driver.spawn(context, instance, image_meta, [ 585.955064] env[61957]: ERROR nova.compute.manager [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 585.955064] env[61957]: ERROR nova.compute.manager [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 585.955064] env[61957]: ERROR nova.compute.manager [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 585.955064] env[61957]: ERROR nova.compute.manager [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] vm_ref = self.build_virtual_machine(instance, [ 585.955756] env[61957]: ERROR nova.compute.manager [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 585.955756] env[61957]: ERROR nova.compute.manager [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] vif_infos = vmwarevif.get_vif_info(self._session, [ 585.955756] env[61957]: ERROR nova.compute.manager [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 585.955756] env[61957]: ERROR nova.compute.manager [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] for vif in network_info: [ 585.955756] env[61957]: ERROR nova.compute.manager [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 585.955756] env[61957]: ERROR nova.compute.manager [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] return self._sync_wrapper(fn, *args, **kwargs) [ 585.955756] env[61957]: ERROR nova.compute.manager [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 585.955756] env[61957]: ERROR nova.compute.manager [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] self.wait() [ 585.955756] env[61957]: ERROR nova.compute.manager [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 585.955756] env[61957]: ERROR nova.compute.manager [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] self[:] = self._gt.wait() [ 585.955756] env[61957]: ERROR nova.compute.manager [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 585.955756] env[61957]: ERROR nova.compute.manager [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] return self._exit_event.wait() [ 585.955756] env[61957]: ERROR nova.compute.manager [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 585.957804] env[61957]: ERROR nova.compute.manager [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] result = hub.switch() [ 585.957804] env[61957]: ERROR nova.compute.manager [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 585.957804] env[61957]: ERROR nova.compute.manager [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] return self.greenlet.switch() [ 585.957804] env[61957]: ERROR nova.compute.manager [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 585.957804] env[61957]: ERROR nova.compute.manager [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] result = function(*args, **kwargs) [ 585.957804] env[61957]: ERROR nova.compute.manager [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 585.957804] env[61957]: ERROR nova.compute.manager [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] return func(*args, **kwargs) [ 585.957804] env[61957]: ERROR nova.compute.manager [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 585.957804] env[61957]: ERROR nova.compute.manager [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] raise e [ 585.957804] env[61957]: ERROR nova.compute.manager [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 585.957804] env[61957]: ERROR nova.compute.manager [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] nwinfo = self.network_api.allocate_for_instance( [ 585.957804] env[61957]: ERROR nova.compute.manager [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 585.957804] env[61957]: ERROR nova.compute.manager [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] created_port_ids = self._update_ports_for_instance( [ 585.958230] env[61957]: ERROR nova.compute.manager [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 585.958230] env[61957]: ERROR nova.compute.manager [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] with excutils.save_and_reraise_exception(): [ 585.958230] env[61957]: ERROR nova.compute.manager [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 585.958230] env[61957]: ERROR nova.compute.manager [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] self.force_reraise() [ 585.958230] env[61957]: ERROR nova.compute.manager [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 585.958230] env[61957]: ERROR nova.compute.manager [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] raise self.value [ 585.958230] env[61957]: ERROR nova.compute.manager [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 585.958230] env[61957]: ERROR nova.compute.manager [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] updated_port = self._update_port( [ 585.958230] env[61957]: ERROR nova.compute.manager [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 585.958230] env[61957]: ERROR nova.compute.manager [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] _ensure_no_port_binding_failure(port) [ 585.958230] env[61957]: ERROR nova.compute.manager [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 585.958230] env[61957]: ERROR nova.compute.manager [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] raise exception.PortBindingFailed(port_id=port['id']) [ 585.958534] env[61957]: ERROR nova.compute.manager [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] nova.exception.PortBindingFailed: Binding failed for port edd470af-41f9-45fb-a01b-74445e4d2f66, please check neutron logs for more information. [ 585.958534] env[61957]: ERROR nova.compute.manager [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] [ 585.958534] env[61957]: INFO nova.compute.manager [None req-423a0985-b17f-45fa-b4ad-54e09005ad46 tempest-TenantUsagesTestJSON-2011428287 tempest-TenantUsagesTestJSON-2011428287-project-member] [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] Terminating instance [ 585.959603] env[61957]: DEBUG oslo_concurrency.lockutils [None req-423a0985-b17f-45fa-b4ad-54e09005ad46 tempest-TenantUsagesTestJSON-2011428287 tempest-TenantUsagesTestJSON-2011428287-project-member] Acquiring lock "refresh_cache-ed456ff1-249c-45d0-a007-3fef96ae8a2d" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 585.961289] env[61957]: DEBUG oslo_concurrency.lockutils [None req-423a0985-b17f-45fa-b4ad-54e09005ad46 tempest-TenantUsagesTestJSON-2011428287 tempest-TenantUsagesTestJSON-2011428287-project-member] Acquired lock "refresh_cache-ed456ff1-249c-45d0-a007-3fef96ae8a2d" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 585.961289] env[61957]: DEBUG nova.network.neutron [None req-423a0985-b17f-45fa-b4ad-54e09005ad46 tempest-TenantUsagesTestJSON-2011428287 tempest-TenantUsagesTestJSON-2011428287-project-member] [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 585.995739] env[61957]: DEBUG nova.compute.utils [None req-4b8bfc5c-9cfa-4b72-9c9f-71ae820b9856 tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] Using /dev/sd instead of None {{(pid=61957) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 585.999244] env[61957]: DEBUG nova.compute.manager [None req-4b8bfc5c-9cfa-4b72-9c9f-71ae820b9856 tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] [instance: c5af147e-3526-4014-98bc-7ad163ae89d9] Not allocating networking since 'none' was specified. {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 586.236083] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d45990c2-96e9-4e99-829a-351c842338bd {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.245137] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53d2e871-48e8-4d63-8649-17d8a7606595 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.281530] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e51a9d3-be37-4e83-afa4-8c48ec31b512 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.290101] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb8900df-1e3e-4e63-a04d-662187a86dd5 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.307595] env[61957]: DEBUG nova.compute.provider_tree [None req-d3a5f917-ae8e-4fd7-b4eb-cbb6e49d67de tempest-ServerDiagnosticsNegativeTest-1621277291 tempest-ServerDiagnosticsNegativeTest-1621277291-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 586.359140] env[61957]: DEBUG nova.compute.manager [req-b10f6cc4-a460-49bd-a2f0-bf77afd98222 req-54f4be1c-012a-4bac-a197-5ed99c8d8274 service nova] [instance: 43942e78-7703-444d-91e3-bc7913299654] Received event network-vif-deleted-021f5ba1-78f4-4aa6-b030-243c371db8ac {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 586.359340] env[61957]: DEBUG nova.compute.manager [req-b10f6cc4-a460-49bd-a2f0-bf77afd98222 req-54f4be1c-012a-4bac-a197-5ed99c8d8274 service nova] [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] Received event network-changed-edd470af-41f9-45fb-a01b-74445e4d2f66 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 586.359492] env[61957]: DEBUG nova.compute.manager [req-b10f6cc4-a460-49bd-a2f0-bf77afd98222 req-54f4be1c-012a-4bac-a197-5ed99c8d8274 service nova] [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] Refreshing instance network info cache due to event network-changed-edd470af-41f9-45fb-a01b-74445e4d2f66. {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 586.359669] env[61957]: DEBUG oslo_concurrency.lockutils [req-b10f6cc4-a460-49bd-a2f0-bf77afd98222 req-54f4be1c-012a-4bac-a197-5ed99c8d8274 service nova] Acquiring lock "refresh_cache-ed456ff1-249c-45d0-a007-3fef96ae8a2d" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 586.425307] env[61957]: INFO nova.scheduler.client.report [None req-abb0f6e8-eaec-4ec9-a6d4-89648c99b1d8 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] Deleted allocations for instance 1a0e117e-aa56-44fa-8818-71f0fd0b6183 [ 586.502148] env[61957]: DEBUG nova.compute.manager [None req-4b8bfc5c-9cfa-4b72-9c9f-71ae820b9856 tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] [instance: c5af147e-3526-4014-98bc-7ad163ae89d9] Start building block device mappings for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 586.511095] env[61957]: DEBUG nova.network.neutron [None req-423a0985-b17f-45fa-b4ad-54e09005ad46 tempest-TenantUsagesTestJSON-2011428287 tempest-TenantUsagesTestJSON-2011428287-project-member] [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 586.716927] env[61957]: DEBUG nova.network.neutron [None req-423a0985-b17f-45fa-b4ad-54e09005ad46 tempest-TenantUsagesTestJSON-2011428287 tempest-TenantUsagesTestJSON-2011428287-project-member] [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 586.811938] env[61957]: DEBUG nova.scheduler.client.report [None req-d3a5f917-ae8e-4fd7-b4eb-cbb6e49d67de tempest-ServerDiagnosticsNegativeTest-1621277291 tempest-ServerDiagnosticsNegativeTest-1621277291-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 586.938347] env[61957]: DEBUG oslo_concurrency.lockutils [None req-abb0f6e8-eaec-4ec9-a6d4-89648c99b1d8 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] Lock "1a0e117e-aa56-44fa-8818-71f0fd0b6183" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 23.806s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 587.220294] env[61957]: DEBUG oslo_concurrency.lockutils [None req-423a0985-b17f-45fa-b4ad-54e09005ad46 tempest-TenantUsagesTestJSON-2011428287 tempest-TenantUsagesTestJSON-2011428287-project-member] Releasing lock "refresh_cache-ed456ff1-249c-45d0-a007-3fef96ae8a2d" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 587.221334] env[61957]: DEBUG nova.compute.manager [None req-423a0985-b17f-45fa-b4ad-54e09005ad46 tempest-TenantUsagesTestJSON-2011428287 tempest-TenantUsagesTestJSON-2011428287-project-member] [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] Start destroying the instance on the hypervisor. {{(pid=61957) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 587.221947] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-423a0985-b17f-45fa-b4ad-54e09005ad46 tempest-TenantUsagesTestJSON-2011428287 tempest-TenantUsagesTestJSON-2011428287-project-member] [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] Destroying instance {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 587.222125] env[61957]: DEBUG oslo_concurrency.lockutils [req-b10f6cc4-a460-49bd-a2f0-bf77afd98222 req-54f4be1c-012a-4bac-a197-5ed99c8d8274 service nova] Acquired lock "refresh_cache-ed456ff1-249c-45d0-a007-3fef96ae8a2d" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 587.222435] env[61957]: DEBUG nova.network.neutron [req-b10f6cc4-a460-49bd-a2f0-bf77afd98222 req-54f4be1c-012a-4bac-a197-5ed99c8d8274 service nova] [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] Refreshing network info cache for port edd470af-41f9-45fb-a01b-74445e4d2f66 {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 587.223462] env[61957]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-622820e8-7498-460e-89eb-45a02bc64e01 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.235955] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91d5862b-2387-485e-bd7b-d8b86efcf9d9 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.270117] env[61957]: WARNING nova.virt.vmwareapi.vmops [None req-423a0985-b17f-45fa-b4ad-54e09005ad46 tempest-TenantUsagesTestJSON-2011428287 tempest-TenantUsagesTestJSON-2011428287-project-member] [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance ed456ff1-249c-45d0-a007-3fef96ae8a2d could not be found. [ 587.270851] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-423a0985-b17f-45fa-b4ad-54e09005ad46 tempest-TenantUsagesTestJSON-2011428287 tempest-TenantUsagesTestJSON-2011428287-project-member] [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] Instance destroyed {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 587.270851] env[61957]: INFO nova.compute.manager [None req-423a0985-b17f-45fa-b4ad-54e09005ad46 tempest-TenantUsagesTestJSON-2011428287 tempest-TenantUsagesTestJSON-2011428287-project-member] [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] Took 0.05 seconds to destroy the instance on the hypervisor. [ 587.272470] env[61957]: DEBUG oslo.service.loopingcall [None req-423a0985-b17f-45fa-b4ad-54e09005ad46 tempest-TenantUsagesTestJSON-2011428287 tempest-TenantUsagesTestJSON-2011428287-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 587.272470] env[61957]: DEBUG nova.compute.manager [-] [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 587.272470] env[61957]: DEBUG nova.network.neutron [-] [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 587.317113] env[61957]: DEBUG nova.network.neutron [-] [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 587.321134] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d3a5f917-ae8e-4fd7-b4eb-cbb6e49d67de tempest-ServerDiagnosticsNegativeTest-1621277291 tempest-ServerDiagnosticsNegativeTest-1621277291-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.834s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 587.322022] env[61957]: ERROR nova.compute.manager [None req-d3a5f917-ae8e-4fd7-b4eb-cbb6e49d67de tempest-ServerDiagnosticsNegativeTest-1621277291 tempest-ServerDiagnosticsNegativeTest-1621277291-project-member] [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port c759a216-a6b5-4bff-97b2-bdfda7cbc7d5, please check neutron logs for more information. [ 587.322022] env[61957]: ERROR nova.compute.manager [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] Traceback (most recent call last): [ 587.322022] env[61957]: ERROR nova.compute.manager [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 587.322022] env[61957]: ERROR nova.compute.manager [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] self.driver.spawn(context, instance, image_meta, [ 587.322022] env[61957]: ERROR nova.compute.manager [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 587.322022] env[61957]: ERROR nova.compute.manager [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] self._vmops.spawn(context, instance, image_meta, injected_files, [ 587.322022] env[61957]: ERROR nova.compute.manager [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 587.322022] env[61957]: ERROR nova.compute.manager [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] vm_ref = self.build_virtual_machine(instance, [ 587.322022] env[61957]: ERROR nova.compute.manager [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 587.322022] env[61957]: ERROR nova.compute.manager [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] vif_infos = vmwarevif.get_vif_info(self._session, [ 587.322022] env[61957]: ERROR nova.compute.manager [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 587.322567] env[61957]: ERROR nova.compute.manager [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] for vif in network_info: [ 587.322567] env[61957]: ERROR nova.compute.manager [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 587.322567] env[61957]: ERROR nova.compute.manager [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] return self._sync_wrapper(fn, *args, **kwargs) [ 587.322567] env[61957]: ERROR nova.compute.manager [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 587.322567] env[61957]: ERROR nova.compute.manager [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] self.wait() [ 587.322567] env[61957]: ERROR nova.compute.manager [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 587.322567] env[61957]: ERROR nova.compute.manager [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] self[:] = self._gt.wait() [ 587.322567] env[61957]: ERROR nova.compute.manager [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 587.322567] env[61957]: ERROR nova.compute.manager [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] return self._exit_event.wait() [ 587.322567] env[61957]: ERROR nova.compute.manager [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 587.322567] env[61957]: ERROR nova.compute.manager [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] result = hub.switch() [ 587.322567] env[61957]: ERROR nova.compute.manager [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 587.322567] env[61957]: ERROR nova.compute.manager [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] return self.greenlet.switch() [ 587.323123] env[61957]: ERROR nova.compute.manager [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 587.323123] env[61957]: ERROR nova.compute.manager [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] result = function(*args, **kwargs) [ 587.323123] env[61957]: ERROR nova.compute.manager [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 587.323123] env[61957]: ERROR nova.compute.manager [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] return func(*args, **kwargs) [ 587.323123] env[61957]: ERROR nova.compute.manager [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 587.323123] env[61957]: ERROR nova.compute.manager [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] raise e [ 587.323123] env[61957]: ERROR nova.compute.manager [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 587.323123] env[61957]: ERROR nova.compute.manager [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] nwinfo = self.network_api.allocate_for_instance( [ 587.323123] env[61957]: ERROR nova.compute.manager [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 587.323123] env[61957]: ERROR nova.compute.manager [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] created_port_ids = self._update_ports_for_instance( [ 587.323123] env[61957]: ERROR nova.compute.manager [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 587.323123] env[61957]: ERROR nova.compute.manager [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] with excutils.save_and_reraise_exception(): [ 587.323123] env[61957]: ERROR nova.compute.manager [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 587.323684] env[61957]: ERROR nova.compute.manager [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] self.force_reraise() [ 587.323684] env[61957]: ERROR nova.compute.manager [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 587.323684] env[61957]: ERROR nova.compute.manager [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] raise self.value [ 587.323684] env[61957]: ERROR nova.compute.manager [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 587.323684] env[61957]: ERROR nova.compute.manager [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] updated_port = self._update_port( [ 587.323684] env[61957]: ERROR nova.compute.manager [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 587.323684] env[61957]: ERROR nova.compute.manager [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] _ensure_no_port_binding_failure(port) [ 587.323684] env[61957]: ERROR nova.compute.manager [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 587.323684] env[61957]: ERROR nova.compute.manager [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] raise exception.PortBindingFailed(port_id=port['id']) [ 587.323684] env[61957]: ERROR nova.compute.manager [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] nova.exception.PortBindingFailed: Binding failed for port c759a216-a6b5-4bff-97b2-bdfda7cbc7d5, please check neutron logs for more information. [ 587.323684] env[61957]: ERROR nova.compute.manager [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] [ 587.324642] env[61957]: DEBUG nova.compute.utils [None req-d3a5f917-ae8e-4fd7-b4eb-cbb6e49d67de tempest-ServerDiagnosticsNegativeTest-1621277291 tempest-ServerDiagnosticsNegativeTest-1621277291-project-member] [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] Binding failed for port c759a216-a6b5-4bff-97b2-bdfda7cbc7d5, please check neutron logs for more information. {{(pid=61957) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 587.324642] env[61957]: DEBUG oslo_concurrency.lockutils [None req-840b2d8a-f610-4c89-a367-da388d81d5aa tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.903s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 587.325499] env[61957]: INFO nova.compute.claims [None req-840b2d8a-f610-4c89-a367-da388d81d5aa tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 587.331031] env[61957]: DEBUG nova.compute.manager [None req-d3a5f917-ae8e-4fd7-b4eb-cbb6e49d67de tempest-ServerDiagnosticsNegativeTest-1621277291 tempest-ServerDiagnosticsNegativeTest-1621277291-project-member] [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] Build of instance f28611df-9528-4ec0-b575-ad54ebd4a958 was re-scheduled: Binding failed for port c759a216-a6b5-4bff-97b2-bdfda7cbc7d5, please check neutron logs for more information. {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 587.331031] env[61957]: DEBUG nova.compute.manager [None req-d3a5f917-ae8e-4fd7-b4eb-cbb6e49d67de tempest-ServerDiagnosticsNegativeTest-1621277291 tempest-ServerDiagnosticsNegativeTest-1621277291-project-member] [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] Unplugging VIFs for instance {{(pid=61957) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 587.331031] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d3a5f917-ae8e-4fd7-b4eb-cbb6e49d67de tempest-ServerDiagnosticsNegativeTest-1621277291 tempest-ServerDiagnosticsNegativeTest-1621277291-project-member] Acquiring lock "refresh_cache-f28611df-9528-4ec0-b575-ad54ebd4a958" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 587.331031] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d3a5f917-ae8e-4fd7-b4eb-cbb6e49d67de tempest-ServerDiagnosticsNegativeTest-1621277291 tempest-ServerDiagnosticsNegativeTest-1621277291-project-member] Acquired lock "refresh_cache-f28611df-9528-4ec0-b575-ad54ebd4a958" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 587.331251] env[61957]: DEBUG nova.network.neutron [None req-d3a5f917-ae8e-4fd7-b4eb-cbb6e49d67de tempest-ServerDiagnosticsNegativeTest-1621277291 tempest-ServerDiagnosticsNegativeTest-1621277291-project-member] [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 587.442285] env[61957]: DEBUG nova.compute.manager [None req-61bcf42d-e42b-40a8-98f8-af8cb211ebed tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] Starting instance... {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 587.522713] env[61957]: DEBUG nova.compute.manager [None req-4b8bfc5c-9cfa-4b72-9c9f-71ae820b9856 tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] [instance: c5af147e-3526-4014-98bc-7ad163ae89d9] Start spawning the instance on the hypervisor. {{(pid=61957) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 587.560653] env[61957]: DEBUG nova.virt.hardware [None req-4b8bfc5c-9cfa-4b72-9c9f-71ae820b9856 tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T17:22:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T17:21:50Z,direct_url=,disk_format='vmdk',id=11c76a2c-f705-470a-ba9d-4657858bab38,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='81d3096d0f094373913fc3dc8f5c3c6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T17:21:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 587.561073] env[61957]: DEBUG nova.virt.hardware [None req-4b8bfc5c-9cfa-4b72-9c9f-71ae820b9856 tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] Flavor limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 587.561073] env[61957]: DEBUG nova.virt.hardware [None req-4b8bfc5c-9cfa-4b72-9c9f-71ae820b9856 tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] Image limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 587.561250] env[61957]: DEBUG nova.virt.hardware [None req-4b8bfc5c-9cfa-4b72-9c9f-71ae820b9856 tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] Flavor pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 587.561403] env[61957]: DEBUG nova.virt.hardware [None req-4b8bfc5c-9cfa-4b72-9c9f-71ae820b9856 tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] Image pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 587.562800] env[61957]: DEBUG nova.virt.hardware [None req-4b8bfc5c-9cfa-4b72-9c9f-71ae820b9856 tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 587.562800] env[61957]: DEBUG nova.virt.hardware [None req-4b8bfc5c-9cfa-4b72-9c9f-71ae820b9856 tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 587.562800] env[61957]: DEBUG nova.virt.hardware [None req-4b8bfc5c-9cfa-4b72-9c9f-71ae820b9856 tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 587.562800] env[61957]: DEBUG nova.virt.hardware [None req-4b8bfc5c-9cfa-4b72-9c9f-71ae820b9856 tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] Got 1 possible topologies {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 587.562800] env[61957]: DEBUG nova.virt.hardware [None req-4b8bfc5c-9cfa-4b72-9c9f-71ae820b9856 tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 587.562971] env[61957]: DEBUG nova.virt.hardware [None req-4b8bfc5c-9cfa-4b72-9c9f-71ae820b9856 tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 587.563609] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92d85535-9371-4661-8a4c-24b7deb536b1 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.571989] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21641ccf-b284-4588-9e21-46ea0d7d93e0 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.587867] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-4b8bfc5c-9cfa-4b72-9c9f-71ae820b9856 tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] [instance: c5af147e-3526-4014-98bc-7ad163ae89d9] Instance VIF info [] {{(pid=61957) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 587.597451] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-4b8bfc5c-9cfa-4b72-9c9f-71ae820b9856 tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] Creating folder: OpenStack. Parent ref: group-v4. {{(pid=61957) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 587.597739] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a644c202-1712-4e99-a4ae-ba3907a376ea {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.613066] env[61957]: WARNING suds.client [-] Web service reported a SOAP processing fault using an unexpected HTTP status code 200. Reporting as an internal server error. [ 587.613173] env[61957]: DEBUG oslo_vmware.api [-] Fault list: [DuplicateName] {{(pid=61957) _invoke_api /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:337}} [ 587.613963] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-4b8bfc5c-9cfa-4b72-9c9f-71ae820b9856 tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] Folder already exists: OpenStack. Parent ref: group-v4. {{(pid=61957) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 587.613963] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-4b8bfc5c-9cfa-4b72-9c9f-71ae820b9856 tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] Creating folder: Project (0177f13e3d9e4695aeaa2aee835f21e8). Parent ref: group-v274445. {{(pid=61957) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 587.613963] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e5fbf164-ce36-4f16-b624-22298d6bb97f {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.631889] env[61957]: INFO nova.virt.vmwareapi.vm_util [None req-4b8bfc5c-9cfa-4b72-9c9f-71ae820b9856 tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] Created folder: Project (0177f13e3d9e4695aeaa2aee835f21e8) in parent group-v274445. [ 587.631889] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-4b8bfc5c-9cfa-4b72-9c9f-71ae820b9856 tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] Creating folder: Instances. Parent ref: group-v274450. {{(pid=61957) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 587.631889] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d07d72cb-93a9-4e16-8d06-aab34a04422a {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.645093] env[61957]: INFO nova.virt.vmwareapi.vm_util [None req-4b8bfc5c-9cfa-4b72-9c9f-71ae820b9856 tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] Created folder: Instances in parent group-v274450. [ 587.647986] env[61957]: DEBUG oslo.service.loopingcall [None req-4b8bfc5c-9cfa-4b72-9c9f-71ae820b9856 tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 587.648132] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c5af147e-3526-4014-98bc-7ad163ae89d9] Creating VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 587.648395] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f81cb6c0-ac48-4c67-b989-ac00afe7200c {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.671163] env[61957]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 587.671163] env[61957]: value = "task-1277299" [ 587.671163] env[61957]: _type = "Task" [ 587.671163] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 587.680090] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277299, 'name': CreateVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 587.824640] env[61957]: DEBUG nova.network.neutron [-] [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 587.900609] env[61957]: DEBUG nova.network.neutron [req-b10f6cc4-a460-49bd-a2f0-bf77afd98222 req-54f4be1c-012a-4bac-a197-5ed99c8d8274 service nova] [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 587.905186] env[61957]: DEBUG nova.network.neutron [None req-d3a5f917-ae8e-4fd7-b4eb-cbb6e49d67de tempest-ServerDiagnosticsNegativeTest-1621277291 tempest-ServerDiagnosticsNegativeTest-1621277291-project-member] [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 587.973713] env[61957]: DEBUG oslo_concurrency.lockutils [None req-61bcf42d-e42b-40a8-98f8-af8cb211ebed tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 588.033866] env[61957]: DEBUG oslo_concurrency.lockutils [None req-adf5e3c1-3919-48f3-8d1a-68a60974255e tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] Acquiring lock "3c099888-d5dc-4cee-9729-311808694625" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 588.033866] env[61957]: DEBUG oslo_concurrency.lockutils [None req-adf5e3c1-3919-48f3-8d1a-68a60974255e tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] Lock "3c099888-d5dc-4cee-9729-311808694625" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 588.085923] env[61957]: DEBUG nova.network.neutron [req-b10f6cc4-a460-49bd-a2f0-bf77afd98222 req-54f4be1c-012a-4bac-a197-5ed99c8d8274 service nova] [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 588.093549] env[61957]: DEBUG nova.network.neutron [None req-d3a5f917-ae8e-4fd7-b4eb-cbb6e49d67de tempest-ServerDiagnosticsNegativeTest-1621277291 tempest-ServerDiagnosticsNegativeTest-1621277291-project-member] [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 588.184132] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277299, 'name': CreateVM_Task, 'duration_secs': 0.333983} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 588.184132] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c5af147e-3526-4014-98bc-7ad163ae89d9] Created VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 588.184325] env[61957]: DEBUG oslo_vmware.service [None req-4b8bfc5c-9cfa-4b72-9c9f-71ae820b9856 tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a94948e2-19fe-4618-aef9-0985add1de47 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.190959] env[61957]: DEBUG oslo_concurrency.lockutils [None req-4b8bfc5c-9cfa-4b72-9c9f-71ae820b9856 tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 588.191192] env[61957]: DEBUG oslo_concurrency.lockutils [None req-4b8bfc5c-9cfa-4b72-9c9f-71ae820b9856 tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 588.191845] env[61957]: DEBUG oslo_concurrency.lockutils [None req-4b8bfc5c-9cfa-4b72-9c9f-71ae820b9856 tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 588.192853] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4121e439-8c2e-4354-a076-ba4eeb06a4cb {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.197435] env[61957]: DEBUG oslo_vmware.api [None req-4b8bfc5c-9cfa-4b72-9c9f-71ae820b9856 tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] Waiting for the task: (returnval){ [ 588.197435] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52fcf15d-6784-86e7-e40b-9c036a1aaaff" [ 588.197435] env[61957]: _type = "Task" [ 588.197435] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 588.206633] env[61957]: DEBUG oslo_vmware.api [None req-4b8bfc5c-9cfa-4b72-9c9f-71ae820b9856 tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52fcf15d-6784-86e7-e40b-9c036a1aaaff, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 588.330738] env[61957]: INFO nova.compute.manager [-] [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] Took 1.06 seconds to deallocate network for instance. [ 588.334754] env[61957]: DEBUG nova.compute.claims [None req-423a0985-b17f-45fa-b4ad-54e09005ad46 tempest-TenantUsagesTestJSON-2011428287 tempest-TenantUsagesTestJSON-2011428287-project-member] [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] Aborting claim: {{(pid=61957) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 588.335030] env[61957]: DEBUG oslo_concurrency.lockutils [None req-423a0985-b17f-45fa-b4ad-54e09005ad46 tempest-TenantUsagesTestJSON-2011428287 tempest-TenantUsagesTestJSON-2011428287-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 588.589629] env[61957]: DEBUG oslo_concurrency.lockutils [req-b10f6cc4-a460-49bd-a2f0-bf77afd98222 req-54f4be1c-012a-4bac-a197-5ed99c8d8274 service nova] Releasing lock "refresh_cache-ed456ff1-249c-45d0-a007-3fef96ae8a2d" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 588.590779] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45f3f4c4-a1ca-45f6-8fc5-a5988a7e268f {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.598626] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d3a5f917-ae8e-4fd7-b4eb-cbb6e49d67de tempest-ServerDiagnosticsNegativeTest-1621277291 tempest-ServerDiagnosticsNegativeTest-1621277291-project-member] Releasing lock "refresh_cache-f28611df-9528-4ec0-b575-ad54ebd4a958" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 588.601940] env[61957]: DEBUG nova.compute.manager [None req-d3a5f917-ae8e-4fd7-b4eb-cbb6e49d67de tempest-ServerDiagnosticsNegativeTest-1621277291 tempest-ServerDiagnosticsNegativeTest-1621277291-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61957) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 588.601940] env[61957]: DEBUG nova.compute.manager [None req-d3a5f917-ae8e-4fd7-b4eb-cbb6e49d67de tempest-ServerDiagnosticsNegativeTest-1621277291 tempest-ServerDiagnosticsNegativeTest-1621277291-project-member] [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 588.601940] env[61957]: DEBUG nova.network.neutron [None req-d3a5f917-ae8e-4fd7-b4eb-cbb6e49d67de tempest-ServerDiagnosticsNegativeTest-1621277291 tempest-ServerDiagnosticsNegativeTest-1621277291-project-member] [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 588.602316] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2824ac8-0f1b-4f72-b1de-47fb071d1a52 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.643190] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-554a10b1-7a44-465d-aeba-1360d33d2722 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.649660] env[61957]: DEBUG nova.network.neutron [None req-d3a5f917-ae8e-4fd7-b4eb-cbb6e49d67de tempest-ServerDiagnosticsNegativeTest-1621277291 tempest-ServerDiagnosticsNegativeTest-1621277291-project-member] [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 588.652434] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f25d2ac-bfb1-4b4b-940a-7da9475ff2e9 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.669817] env[61957]: DEBUG nova.compute.provider_tree [None req-840b2d8a-f610-4c89-a367-da388d81d5aa tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 588.711137] env[61957]: DEBUG oslo_concurrency.lockutils [None req-4b8bfc5c-9cfa-4b72-9c9f-71ae820b9856 tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 588.711409] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-4b8bfc5c-9cfa-4b72-9c9f-71ae820b9856 tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] [instance: c5af147e-3526-4014-98bc-7ad163ae89d9] Processing image 11c76a2c-f705-470a-ba9d-4657858bab38 {{(pid=61957) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 588.711643] env[61957]: DEBUG oslo_concurrency.lockutils [None req-4b8bfc5c-9cfa-4b72-9c9f-71ae820b9856 tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 588.711821] env[61957]: DEBUG oslo_concurrency.lockutils [None req-4b8bfc5c-9cfa-4b72-9c9f-71ae820b9856 tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 588.712563] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-4b8bfc5c-9cfa-4b72-9c9f-71ae820b9856 tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 588.712616] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4a9c1a5c-a870-4b59-b8b8-fe4bccba8c99 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.733805] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-4b8bfc5c-9cfa-4b72-9c9f-71ae820b9856 tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 588.734065] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-4b8bfc5c-9cfa-4b72-9c9f-71ae820b9856 tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61957) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 588.734911] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5a9efea-265a-4d57-85df-55009148bebb {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.743415] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d2669de8-2051-4cdd-b0c3-edfde6950c26 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.752199] env[61957]: DEBUG oslo_vmware.api [None req-4b8bfc5c-9cfa-4b72-9c9f-71ae820b9856 tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] Waiting for the task: (returnval){ [ 588.752199] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52922ebc-901e-315b-721c-58984f1ba814" [ 588.752199] env[61957]: _type = "Task" [ 588.752199] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 588.764400] env[61957]: DEBUG oslo_vmware.api [None req-4b8bfc5c-9cfa-4b72-9c9f-71ae820b9856 tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52922ebc-901e-315b-721c-58984f1ba814, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 589.158331] env[61957]: DEBUG nova.network.neutron [None req-d3a5f917-ae8e-4fd7-b4eb-cbb6e49d67de tempest-ServerDiagnosticsNegativeTest-1621277291 tempest-ServerDiagnosticsNegativeTest-1621277291-project-member] [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 589.172869] env[61957]: DEBUG nova.scheduler.client.report [None req-840b2d8a-f610-4c89-a367-da388d81d5aa tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 589.268888] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-4b8bfc5c-9cfa-4b72-9c9f-71ae820b9856 tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] [instance: c5af147e-3526-4014-98bc-7ad163ae89d9] Preparing fetch location {{(pid=61957) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 589.268888] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-4b8bfc5c-9cfa-4b72-9c9f-71ae820b9856 tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] Creating directory with path [datastore2] vmware_temp/ac3f241c-a427-4dcd-b860-d9939dd5b05e/11c76a2c-f705-470a-ba9d-4657858bab38 {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 589.268888] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-fb9f1a71-57a6-4e33-8922-abc30691db72 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.304763] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-4b8bfc5c-9cfa-4b72-9c9f-71ae820b9856 tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] Created directory with path [datastore2] vmware_temp/ac3f241c-a427-4dcd-b860-d9939dd5b05e/11c76a2c-f705-470a-ba9d-4657858bab38 {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 589.305011] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-4b8bfc5c-9cfa-4b72-9c9f-71ae820b9856 tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] [instance: c5af147e-3526-4014-98bc-7ad163ae89d9] Fetch image to [datastore2] vmware_temp/ac3f241c-a427-4dcd-b860-d9939dd5b05e/11c76a2c-f705-470a-ba9d-4657858bab38/tmp-sparse.vmdk {{(pid=61957) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 589.305156] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-4b8bfc5c-9cfa-4b72-9c9f-71ae820b9856 tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] [instance: c5af147e-3526-4014-98bc-7ad163ae89d9] Downloading image file data 11c76a2c-f705-470a-ba9d-4657858bab38 to [datastore2] vmware_temp/ac3f241c-a427-4dcd-b860-d9939dd5b05e/11c76a2c-f705-470a-ba9d-4657858bab38/tmp-sparse.vmdk on the data store datastore2 {{(pid=61957) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 589.306023] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c14e126d-08af-4fe7-b540-4dc069f691da {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.316135] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ef4435d-9d59-4c89-bc0e-95972a19c7c1 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.328839] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a101a4d-8d56-4428-8d91-fa650655567e {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.372727] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d4def87-e4c7-40a3-a5e7-fdc2baf92768 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.377689] env[61957]: DEBUG nova.compute.manager [req-a06fa753-f6e4-432e-a0dc-f3ca9c529600 req-64f2f97d-2821-4e18-aa6a-e6f8fbe63bd8 service nova] [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] Received event network-vif-deleted-edd470af-41f9-45fb-a01b-74445e4d2f66 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 589.383882] env[61957]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-d72f2c6a-aa62-46e9-8495-b7cdd6eece67 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.420845] env[61957]: DEBUG nova.virt.vmwareapi.images [None req-4b8bfc5c-9cfa-4b72-9c9f-71ae820b9856 tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] [instance: c5af147e-3526-4014-98bc-7ad163ae89d9] Downloading image file data 11c76a2c-f705-470a-ba9d-4657858bab38 to the data store datastore2 {{(pid=61957) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 589.494463] env[61957]: DEBUG oslo_vmware.rw_handles [None req-4b8bfc5c-9cfa-4b72-9c9f-71ae820b9856 tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c2n2.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/ac3f241c-a427-4dcd-b860-d9939dd5b05e/11c76a2c-f705-470a-ba9d-4657858bab38/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore2. {{(pid=61957) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 589.624715] env[61957]: DEBUG oslo_concurrency.lockutils [None req-4e7f7a28-6560-4b73-affe-356aa8d11b0b tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] Acquiring lock "763d87bb-652c-43e0-ba39-135bae2cc368" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 589.624715] env[61957]: DEBUG oslo_concurrency.lockutils [None req-4e7f7a28-6560-4b73-affe-356aa8d11b0b tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] Lock "763d87bb-652c-43e0-ba39-135bae2cc368" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 589.661111] env[61957]: INFO nova.compute.manager [None req-d3a5f917-ae8e-4fd7-b4eb-cbb6e49d67de tempest-ServerDiagnosticsNegativeTest-1621277291 tempest-ServerDiagnosticsNegativeTest-1621277291-project-member] [instance: f28611df-9528-4ec0-b575-ad54ebd4a958] Took 1.06 seconds to deallocate network for instance. [ 589.679301] env[61957]: DEBUG oslo_concurrency.lockutils [None req-840b2d8a-f610-4c89-a367-da388d81d5aa tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.355s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 589.680860] env[61957]: DEBUG nova.compute.manager [None req-840b2d8a-f610-4c89-a367-da388d81d5aa tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] Start building networks asynchronously for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 589.682669] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d102ca19-392e-4051-87d1-b58bed2da29d tempest-ServersAdminNegativeTestJSON-555579 tempest-ServersAdminNegativeTestJSON-555579-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 11.575s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 590.189888] env[61957]: DEBUG nova.compute.utils [None req-840b2d8a-f610-4c89-a367-da388d81d5aa tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] Using /dev/sd instead of None {{(pid=61957) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 590.189888] env[61957]: DEBUG nova.compute.manager [None req-840b2d8a-f610-4c89-a367-da388d81d5aa tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] Allocating IP information in the background. {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 590.189888] env[61957]: DEBUG nova.network.neutron [None req-840b2d8a-f610-4c89-a367-da388d81d5aa tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] allocate_for_instance() {{(pid=61957) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 590.200034] env[61957]: DEBUG oslo_vmware.rw_handles [None req-4b8bfc5c-9cfa-4b72-9c9f-71ae820b9856 tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] Completed reading data from the image iterator. {{(pid=61957) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 590.200208] env[61957]: DEBUG oslo_vmware.rw_handles [None req-4b8bfc5c-9cfa-4b72-9c9f-71ae820b9856 tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] Closing write handle for https://esx7c2n2.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/ac3f241c-a427-4dcd-b860-d9939dd5b05e/11c76a2c-f705-470a-ba9d-4657858bab38/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore2. {{(pid=61957) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 590.355675] env[61957]: DEBUG nova.policy [None req-840b2d8a-f610-4c89-a367-da388d81d5aa tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6871e6657be047708fb735ffb71dc19d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f1243661c0ff4c33b901de6aacfd5987', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61957) authorize /opt/stack/nova/nova/policy.py:203}} [ 590.358402] env[61957]: DEBUG nova.virt.vmwareapi.images [None req-4b8bfc5c-9cfa-4b72-9c9f-71ae820b9856 tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] [instance: c5af147e-3526-4014-98bc-7ad163ae89d9] Downloaded image file data 11c76a2c-f705-470a-ba9d-4657858bab38 to vmware_temp/ac3f241c-a427-4dcd-b860-d9939dd5b05e/11c76a2c-f705-470a-ba9d-4657858bab38/tmp-sparse.vmdk on the data store datastore2 {{(pid=61957) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 590.359846] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-4b8bfc5c-9cfa-4b72-9c9f-71ae820b9856 tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] [instance: c5af147e-3526-4014-98bc-7ad163ae89d9] Caching image {{(pid=61957) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 590.359846] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-4b8bfc5c-9cfa-4b72-9c9f-71ae820b9856 tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] Copying Virtual Disk [datastore2] vmware_temp/ac3f241c-a427-4dcd-b860-d9939dd5b05e/11c76a2c-f705-470a-ba9d-4657858bab38/tmp-sparse.vmdk to [datastore2] vmware_temp/ac3f241c-a427-4dcd-b860-d9939dd5b05e/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 590.360254] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-faa6c6d3-d712-4445-8541-0466836d784b {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.373180] env[61957]: DEBUG oslo_vmware.api [None req-4b8bfc5c-9cfa-4b72-9c9f-71ae820b9856 tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] Waiting for the task: (returnval){ [ 590.373180] env[61957]: value = "task-1277300" [ 590.373180] env[61957]: _type = "Task" [ 590.373180] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 590.384887] env[61957]: DEBUG oslo_vmware.api [None req-4b8bfc5c-9cfa-4b72-9c9f-71ae820b9856 tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] Task: {'id': task-1277300, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 590.467858] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1063f81-5c5b-42a4-8c4a-2b34033b1dca {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.480970] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e760632-e114-4d01-847f-47ae601f18b3 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.523505] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4894bbbd-3442-44c1-a9df-c3466f0ee1b6 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.535687] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93bf12e1-0935-4654-96b6-a6e7c4a675c5 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.554039] env[61957]: DEBUG nova.compute.provider_tree [None req-d102ca19-392e-4051-87d1-b58bed2da29d tempest-ServersAdminNegativeTestJSON-555579 tempest-ServersAdminNegativeTestJSON-555579-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 590.698659] env[61957]: DEBUG nova.compute.manager [None req-840b2d8a-f610-4c89-a367-da388d81d5aa tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] Start building block device mappings for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 590.703294] env[61957]: INFO nova.scheduler.client.report [None req-d3a5f917-ae8e-4fd7-b4eb-cbb6e49d67de tempest-ServerDiagnosticsNegativeTest-1621277291 tempest-ServerDiagnosticsNegativeTest-1621277291-project-member] Deleted allocations for instance f28611df-9528-4ec0-b575-ad54ebd4a958 [ 590.889199] env[61957]: DEBUG oslo_vmware.api [None req-4b8bfc5c-9cfa-4b72-9c9f-71ae820b9856 tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] Task: {'id': task-1277300, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 591.057234] env[61957]: DEBUG nova.scheduler.client.report [None req-d102ca19-392e-4051-87d1-b58bed2da29d tempest-ServersAdminNegativeTestJSON-555579 tempest-ServersAdminNegativeTestJSON-555579-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 591.213400] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d3a5f917-ae8e-4fd7-b4eb-cbb6e49d67de tempest-ServerDiagnosticsNegativeTest-1621277291 tempest-ServerDiagnosticsNegativeTest-1621277291-project-member] Lock "f28611df-9528-4ec0-b575-ad54ebd4a958" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 27.441s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 591.265977] env[61957]: DEBUG nova.network.neutron [None req-840b2d8a-f610-4c89-a367-da388d81d5aa tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] Successfully created port: 25e905db-2ea8-4412-bfad-1406d998af37 {{(pid=61957) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 591.393114] env[61957]: DEBUG oslo_vmware.api [None req-4b8bfc5c-9cfa-4b72-9c9f-71ae820b9856 tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] Task: {'id': task-1277300, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.738481} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 591.393436] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-4b8bfc5c-9cfa-4b72-9c9f-71ae820b9856 tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] Copied Virtual Disk [datastore2] vmware_temp/ac3f241c-a427-4dcd-b860-d9939dd5b05e/11c76a2c-f705-470a-ba9d-4657858bab38/tmp-sparse.vmdk to [datastore2] vmware_temp/ac3f241c-a427-4dcd-b860-d9939dd5b05e/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 591.393926] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-4b8bfc5c-9cfa-4b72-9c9f-71ae820b9856 tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] Deleting the datastore file [datastore2] vmware_temp/ac3f241c-a427-4dcd-b860-d9939dd5b05e/11c76a2c-f705-470a-ba9d-4657858bab38/tmp-sparse.vmdk {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 591.394526] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8bc9068a-163a-4fc8-8060-99df16e42c6c {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.406152] env[61957]: DEBUG oslo_vmware.api [None req-4b8bfc5c-9cfa-4b72-9c9f-71ae820b9856 tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] Waiting for the task: (returnval){ [ 591.406152] env[61957]: value = "task-1277301" [ 591.406152] env[61957]: _type = "Task" [ 591.406152] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 591.417136] env[61957]: DEBUG oslo_vmware.api [None req-4b8bfc5c-9cfa-4b72-9c9f-71ae820b9856 tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] Task: {'id': task-1277301, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 591.435780] env[61957]: DEBUG oslo_concurrency.lockutils [None req-9c3c6101-e50c-439a-91f6-f293e059d0d2 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] Acquiring lock "4be3eabd-4e58-48b8-b14a-798ba419655e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 591.436067] env[61957]: DEBUG oslo_concurrency.lockutils [None req-9c3c6101-e50c-439a-91f6-f293e059d0d2 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] Lock "4be3eabd-4e58-48b8-b14a-798ba419655e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 591.565668] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d102ca19-392e-4051-87d1-b58bed2da29d tempest-ServersAdminNegativeTestJSON-555579 tempest-ServersAdminNegativeTestJSON-555579-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.883s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 591.566557] env[61957]: ERROR nova.compute.manager [None req-d102ca19-392e-4051-87d1-b58bed2da29d tempest-ServersAdminNegativeTestJSON-555579 tempest-ServersAdminNegativeTestJSON-555579-project-member] [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 178d7697-f823-4565-9fc6-706777f5e312, please check neutron logs for more information. [ 591.566557] env[61957]: ERROR nova.compute.manager [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] Traceback (most recent call last): [ 591.566557] env[61957]: ERROR nova.compute.manager [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 591.566557] env[61957]: ERROR nova.compute.manager [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] self.driver.spawn(context, instance, image_meta, [ 591.566557] env[61957]: ERROR nova.compute.manager [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 591.566557] env[61957]: ERROR nova.compute.manager [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] self._vmops.spawn(context, instance, image_meta, injected_files, [ 591.566557] env[61957]: ERROR nova.compute.manager [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 591.566557] env[61957]: ERROR nova.compute.manager [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] vm_ref = self.build_virtual_machine(instance, [ 591.566557] env[61957]: ERROR nova.compute.manager [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 591.566557] env[61957]: ERROR nova.compute.manager [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] vif_infos = vmwarevif.get_vif_info(self._session, [ 591.566557] env[61957]: ERROR nova.compute.manager [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 591.566930] env[61957]: ERROR nova.compute.manager [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] for vif in network_info: [ 591.566930] env[61957]: ERROR nova.compute.manager [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 591.566930] env[61957]: ERROR nova.compute.manager [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] return self._sync_wrapper(fn, *args, **kwargs) [ 591.566930] env[61957]: ERROR nova.compute.manager [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 591.566930] env[61957]: ERROR nova.compute.manager [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] self.wait() [ 591.566930] env[61957]: ERROR nova.compute.manager [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 591.566930] env[61957]: ERROR nova.compute.manager [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] self[:] = self._gt.wait() [ 591.566930] env[61957]: ERROR nova.compute.manager [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 591.566930] env[61957]: ERROR nova.compute.manager [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] return self._exit_event.wait() [ 591.566930] env[61957]: ERROR nova.compute.manager [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 591.566930] env[61957]: ERROR nova.compute.manager [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] result = hub.switch() [ 591.566930] env[61957]: ERROR nova.compute.manager [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 591.566930] env[61957]: ERROR nova.compute.manager [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] return self.greenlet.switch() [ 591.567407] env[61957]: ERROR nova.compute.manager [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 591.567407] env[61957]: ERROR nova.compute.manager [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] result = function(*args, **kwargs) [ 591.567407] env[61957]: ERROR nova.compute.manager [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 591.567407] env[61957]: ERROR nova.compute.manager [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] return func(*args, **kwargs) [ 591.567407] env[61957]: ERROR nova.compute.manager [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 591.567407] env[61957]: ERROR nova.compute.manager [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] raise e [ 591.567407] env[61957]: ERROR nova.compute.manager [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 591.567407] env[61957]: ERROR nova.compute.manager [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] nwinfo = self.network_api.allocate_for_instance( [ 591.567407] env[61957]: ERROR nova.compute.manager [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 591.567407] env[61957]: ERROR nova.compute.manager [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] created_port_ids = self._update_ports_for_instance( [ 591.567407] env[61957]: ERROR nova.compute.manager [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 591.567407] env[61957]: ERROR nova.compute.manager [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] with excutils.save_and_reraise_exception(): [ 591.567407] env[61957]: ERROR nova.compute.manager [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 591.568071] env[61957]: ERROR nova.compute.manager [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] self.force_reraise() [ 591.568071] env[61957]: ERROR nova.compute.manager [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 591.568071] env[61957]: ERROR nova.compute.manager [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] raise self.value [ 591.568071] env[61957]: ERROR nova.compute.manager [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 591.568071] env[61957]: ERROR nova.compute.manager [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] updated_port = self._update_port( [ 591.568071] env[61957]: ERROR nova.compute.manager [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 591.568071] env[61957]: ERROR nova.compute.manager [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] _ensure_no_port_binding_failure(port) [ 591.568071] env[61957]: ERROR nova.compute.manager [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 591.568071] env[61957]: ERROR nova.compute.manager [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] raise exception.PortBindingFailed(port_id=port['id']) [ 591.568071] env[61957]: ERROR nova.compute.manager [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] nova.exception.PortBindingFailed: Binding failed for port 178d7697-f823-4565-9fc6-706777f5e312, please check neutron logs for more information. [ 591.568071] env[61957]: ERROR nova.compute.manager [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] [ 591.568661] env[61957]: DEBUG nova.compute.utils [None req-d102ca19-392e-4051-87d1-b58bed2da29d tempest-ServersAdminNegativeTestJSON-555579 tempest-ServersAdminNegativeTestJSON-555579-project-member] [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] Binding failed for port 178d7697-f823-4565-9fc6-706777f5e312, please check neutron logs for more information. {{(pid=61957) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 591.569473] env[61957]: DEBUG nova.compute.manager [None req-d102ca19-392e-4051-87d1-b58bed2da29d tempest-ServersAdminNegativeTestJSON-555579 tempest-ServersAdminNegativeTestJSON-555579-project-member] [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] Build of instance 37837096-5ad0-4b6f-b32d-f2899d88b6dd was re-scheduled: Binding failed for port 178d7697-f823-4565-9fc6-706777f5e312, please check neutron logs for more information. {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 591.569881] env[61957]: DEBUG nova.compute.manager [None req-d102ca19-392e-4051-87d1-b58bed2da29d tempest-ServersAdminNegativeTestJSON-555579 tempest-ServersAdminNegativeTestJSON-555579-project-member] [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] Unplugging VIFs for instance {{(pid=61957) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 591.570119] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d102ca19-392e-4051-87d1-b58bed2da29d tempest-ServersAdminNegativeTestJSON-555579 tempest-ServersAdminNegativeTestJSON-555579-project-member] Acquiring lock "refresh_cache-37837096-5ad0-4b6f-b32d-f2899d88b6dd" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 591.571051] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d102ca19-392e-4051-87d1-b58bed2da29d tempest-ServersAdminNegativeTestJSON-555579 tempest-ServersAdminNegativeTestJSON-555579-project-member] Acquired lock "refresh_cache-37837096-5ad0-4b6f-b32d-f2899d88b6dd" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 591.571051] env[61957]: DEBUG nova.network.neutron [None req-d102ca19-392e-4051-87d1-b58bed2da29d tempest-ServersAdminNegativeTestJSON-555579 tempest-ServersAdminNegativeTestJSON-555579-project-member] [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 591.572734] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d83719b8-6bd9-4ae7-99ea-7d0970401ba5 tempest-ServerExternalEventsTest-1886420864 tempest-ServerExternalEventsTest-1886420864-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 12.267s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 591.715417] env[61957]: DEBUG nova.compute.manager [None req-840b2d8a-f610-4c89-a367-da388d81d5aa tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] Start spawning the instance on the hypervisor. {{(pid=61957) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 591.718296] env[61957]: DEBUG nova.compute.manager [None req-7527ee8e-e60c-46dd-bb63-3e8733859a46 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] [instance: 4ad746d0-4076-4c97-8ea9-20e49fda59ca] Starting instance... {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 591.741586] env[61957]: DEBUG nova.virt.hardware [None req-840b2d8a-f610-4c89-a367-da388d81d5aa tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T17:22:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T17:21:50Z,direct_url=,disk_format='vmdk',id=11c76a2c-f705-470a-ba9d-4657858bab38,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='81d3096d0f094373913fc3dc8f5c3c6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T17:21:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 591.741826] env[61957]: DEBUG nova.virt.hardware [None req-840b2d8a-f610-4c89-a367-da388d81d5aa tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] Flavor limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 591.741980] env[61957]: DEBUG nova.virt.hardware [None req-840b2d8a-f610-4c89-a367-da388d81d5aa tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] Image limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 591.743276] env[61957]: DEBUG nova.virt.hardware [None req-840b2d8a-f610-4c89-a367-da388d81d5aa tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] Flavor pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 591.743276] env[61957]: DEBUG nova.virt.hardware [None req-840b2d8a-f610-4c89-a367-da388d81d5aa tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] Image pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 591.743723] env[61957]: DEBUG nova.virt.hardware [None req-840b2d8a-f610-4c89-a367-da388d81d5aa tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 591.744226] env[61957]: DEBUG nova.virt.hardware [None req-840b2d8a-f610-4c89-a367-da388d81d5aa tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 591.744481] env[61957]: DEBUG nova.virt.hardware [None req-840b2d8a-f610-4c89-a367-da388d81d5aa tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 591.745072] env[61957]: DEBUG nova.virt.hardware [None req-840b2d8a-f610-4c89-a367-da388d81d5aa tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] Got 1 possible topologies {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 591.745281] env[61957]: DEBUG nova.virt.hardware [None req-840b2d8a-f610-4c89-a367-da388d81d5aa tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 591.745622] env[61957]: DEBUG nova.virt.hardware [None req-840b2d8a-f610-4c89-a367-da388d81d5aa tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 591.747304] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d9f79e0-ffa5-411a-acf3-c10833001758 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.758218] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b997e57d-cd41-46d7-ac28-2e6ba353284d {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.916064] env[61957]: DEBUG oslo_vmware.api [None req-4b8bfc5c-9cfa-4b72-9c9f-71ae820b9856 tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] Task: {'id': task-1277301, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.024558} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 591.916639] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-4b8bfc5c-9cfa-4b72-9c9f-71ae820b9856 tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] Deleted the datastore file {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 591.916910] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-4b8bfc5c-9cfa-4b72-9c9f-71ae820b9856 tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] Moving file from [datastore2] vmware_temp/ac3f241c-a427-4dcd-b860-d9939dd5b05e/11c76a2c-f705-470a-ba9d-4657858bab38 to [datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38. {{(pid=61957) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 591.917252] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-e16de0e1-5e35-4a26-9f47-523f1ba86473 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.925210] env[61957]: DEBUG oslo_vmware.api [None req-4b8bfc5c-9cfa-4b72-9c9f-71ae820b9856 tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] Waiting for the task: (returnval){ [ 591.925210] env[61957]: value = "task-1277302" [ 591.925210] env[61957]: _type = "Task" [ 591.925210] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 591.933843] env[61957]: DEBUG oslo_vmware.api [None req-4b8bfc5c-9cfa-4b72-9c9f-71ae820b9856 tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] Task: {'id': task-1277302, 'name': MoveDatastoreFile_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 592.136274] env[61957]: DEBUG nova.network.neutron [None req-d102ca19-392e-4051-87d1-b58bed2da29d tempest-ServersAdminNegativeTestJSON-555579 tempest-ServersAdminNegativeTestJSON-555579-project-member] [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 592.250737] env[61957]: DEBUG oslo_concurrency.lockutils [None req-7527ee8e-e60c-46dd-bb63-3e8733859a46 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 592.305428] env[61957]: DEBUG nova.network.neutron [None req-d102ca19-392e-4051-87d1-b58bed2da29d tempest-ServersAdminNegativeTestJSON-555579 tempest-ServersAdminNegativeTestJSON-555579-project-member] [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 592.415560] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f72eb853-5dbb-4b75-81ea-76d7f59719f1 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.426209] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab6aa4ad-e815-4070-bce8-1fbc97975897 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.438638] env[61957]: DEBUG oslo_vmware.api [None req-4b8bfc5c-9cfa-4b72-9c9f-71ae820b9856 tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] Task: {'id': task-1277302, 'name': MoveDatastoreFile_Task, 'duration_secs': 0.030864} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 592.477857] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-4b8bfc5c-9cfa-4b72-9c9f-71ae820b9856 tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] File moved {{(pid=61957) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 592.477857] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-4b8bfc5c-9cfa-4b72-9c9f-71ae820b9856 tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] [instance: c5af147e-3526-4014-98bc-7ad163ae89d9] Cleaning up location [datastore2] vmware_temp/ac3f241c-a427-4dcd-b860-d9939dd5b05e {{(pid=61957) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 592.477857] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-4b8bfc5c-9cfa-4b72-9c9f-71ae820b9856 tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] Deleting the datastore file [datastore2] vmware_temp/ac3f241c-a427-4dcd-b860-d9939dd5b05e {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 592.481513] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0242378a-2dd0-464b-bfe4-59301eecd6a9 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.483842] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3551b5b6-9b36-460f-810a-fa177b7aece7 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.496863] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ec01259-a645-4b70-a311-357b32599697 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.501833] env[61957]: DEBUG oslo_vmware.api [None req-4b8bfc5c-9cfa-4b72-9c9f-71ae820b9856 tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] Waiting for the task: (returnval){ [ 592.501833] env[61957]: value = "task-1277303" [ 592.501833] env[61957]: _type = "Task" [ 592.501833] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 592.517418] env[61957]: DEBUG nova.compute.provider_tree [None req-d83719b8-6bd9-4ae7-99ea-7d0970401ba5 tempest-ServerExternalEventsTest-1886420864 tempest-ServerExternalEventsTest-1886420864-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 592.524335] env[61957]: DEBUG oslo_vmware.api [None req-4b8bfc5c-9cfa-4b72-9c9f-71ae820b9856 tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] Task: {'id': task-1277303, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.027198} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 592.524843] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-4b8bfc5c-9cfa-4b72-9c9f-71ae820b9856 tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] Deleted the datastore file {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 592.528382] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9a98c063-b02d-446e-b620-21f861571714 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.536031] env[61957]: DEBUG oslo_vmware.api [None req-4b8bfc5c-9cfa-4b72-9c9f-71ae820b9856 tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] Waiting for the task: (returnval){ [ 592.536031] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52ccc7e0-705c-bf17-1c11-f5b975b20a09" [ 592.536031] env[61957]: _type = "Task" [ 592.536031] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 592.551690] env[61957]: DEBUG oslo_vmware.api [None req-4b8bfc5c-9cfa-4b72-9c9f-71ae820b9856 tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52ccc7e0-705c-bf17-1c11-f5b975b20a09, 'name': SearchDatastore_Task, 'duration_secs': 0.009211} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 592.551690] env[61957]: DEBUG oslo_concurrency.lockutils [None req-4b8bfc5c-9cfa-4b72-9c9f-71ae820b9856 tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 592.551690] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-4b8bfc5c-9cfa-4b72-9c9f-71ae820b9856 tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore2] c5af147e-3526-4014-98bc-7ad163ae89d9/c5af147e-3526-4014-98bc-7ad163ae89d9.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 592.551690] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c831ad01-6fdc-4505-ba2d-6d872e05534b {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.559968] env[61957]: DEBUG oslo_vmware.api [None req-4b8bfc5c-9cfa-4b72-9c9f-71ae820b9856 tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] Waiting for the task: (returnval){ [ 592.559968] env[61957]: value = "task-1277304" [ 592.559968] env[61957]: _type = "Task" [ 592.559968] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 592.571129] env[61957]: DEBUG oslo_vmware.api [None req-4b8bfc5c-9cfa-4b72-9c9f-71ae820b9856 tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] Task: {'id': task-1277304, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 592.808921] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d102ca19-392e-4051-87d1-b58bed2da29d tempest-ServersAdminNegativeTestJSON-555579 tempest-ServersAdminNegativeTestJSON-555579-project-member] Releasing lock "refresh_cache-37837096-5ad0-4b6f-b32d-f2899d88b6dd" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 592.812440] env[61957]: DEBUG nova.compute.manager [None req-d102ca19-392e-4051-87d1-b58bed2da29d tempest-ServersAdminNegativeTestJSON-555579 tempest-ServersAdminNegativeTestJSON-555579-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61957) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 592.812684] env[61957]: DEBUG nova.compute.manager [None req-d102ca19-392e-4051-87d1-b58bed2da29d tempest-ServersAdminNegativeTestJSON-555579 tempest-ServersAdminNegativeTestJSON-555579-project-member] [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 592.812855] env[61957]: DEBUG nova.network.neutron [None req-d102ca19-392e-4051-87d1-b58bed2da29d tempest-ServersAdminNegativeTestJSON-555579 tempest-ServersAdminNegativeTestJSON-555579-project-member] [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 592.992785] env[61957]: DEBUG nova.network.neutron [None req-d102ca19-392e-4051-87d1-b58bed2da29d tempest-ServersAdminNegativeTestJSON-555579 tempest-ServersAdminNegativeTestJSON-555579-project-member] [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 593.019675] env[61957]: DEBUG nova.scheduler.client.report [None req-d83719b8-6bd9-4ae7-99ea-7d0970401ba5 tempest-ServerExternalEventsTest-1886420864 tempest-ServerExternalEventsTest-1886420864-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 593.078999] env[61957]: DEBUG oslo_vmware.api [None req-4b8bfc5c-9cfa-4b72-9c9f-71ae820b9856 tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] Task: {'id': task-1277304, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.446535} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 593.079273] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-4b8bfc5c-9cfa-4b72-9c9f-71ae820b9856 tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore2] c5af147e-3526-4014-98bc-7ad163ae89d9/c5af147e-3526-4014-98bc-7ad163ae89d9.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 593.079503] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-4b8bfc5c-9cfa-4b72-9c9f-71ae820b9856 tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] [instance: c5af147e-3526-4014-98bc-7ad163ae89d9] Extending root virtual disk to 1048576 {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 593.079752] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5d2cf73c-9499-4b9f-b8a6-fc7c370ce3ad {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.093849] env[61957]: DEBUG oslo_vmware.api [None req-4b8bfc5c-9cfa-4b72-9c9f-71ae820b9856 tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] Waiting for the task: (returnval){ [ 593.093849] env[61957]: value = "task-1277305" [ 593.093849] env[61957]: _type = "Task" [ 593.093849] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 593.105783] env[61957]: DEBUG oslo_vmware.api [None req-4b8bfc5c-9cfa-4b72-9c9f-71ae820b9856 tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] Task: {'id': task-1277305, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 593.253144] env[61957]: DEBUG oslo_concurrency.lockutils [None req-af971b81-6348-4338-b0a9-b06675d34db5 tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] Acquiring lock "967056dd-1679-421c-b4fd-9d663b1411e2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 593.253601] env[61957]: DEBUG oslo_concurrency.lockutils [None req-af971b81-6348-4338-b0a9-b06675d34db5 tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] Lock "967056dd-1679-421c-b4fd-9d663b1411e2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 593.495831] env[61957]: DEBUG nova.network.neutron [None req-d102ca19-392e-4051-87d1-b58bed2da29d tempest-ServersAdminNegativeTestJSON-555579 tempest-ServersAdminNegativeTestJSON-555579-project-member] [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 593.532624] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d83719b8-6bd9-4ae7-99ea-7d0970401ba5 tempest-ServerExternalEventsTest-1886420864 tempest-ServerExternalEventsTest-1886420864-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.959s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 593.533655] env[61957]: ERROR nova.compute.manager [None req-d83719b8-6bd9-4ae7-99ea-7d0970401ba5 tempest-ServerExternalEventsTest-1886420864 tempest-ServerExternalEventsTest-1886420864-project-member] [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 3d42e5dd-f713-445a-8855-3e0e765e2c4a, please check neutron logs for more information. [ 593.533655] env[61957]: ERROR nova.compute.manager [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] Traceback (most recent call last): [ 593.533655] env[61957]: ERROR nova.compute.manager [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 593.533655] env[61957]: ERROR nova.compute.manager [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] self.driver.spawn(context, instance, image_meta, [ 593.533655] env[61957]: ERROR nova.compute.manager [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 593.533655] env[61957]: ERROR nova.compute.manager [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] self._vmops.spawn(context, instance, image_meta, injected_files, [ 593.533655] env[61957]: ERROR nova.compute.manager [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 593.533655] env[61957]: ERROR nova.compute.manager [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] vm_ref = self.build_virtual_machine(instance, [ 593.533655] env[61957]: ERROR nova.compute.manager [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 593.533655] env[61957]: ERROR nova.compute.manager [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] vif_infos = vmwarevif.get_vif_info(self._session, [ 593.533655] env[61957]: ERROR nova.compute.manager [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 593.535893] env[61957]: ERROR nova.compute.manager [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] for vif in network_info: [ 593.535893] env[61957]: ERROR nova.compute.manager [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 593.535893] env[61957]: ERROR nova.compute.manager [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] return self._sync_wrapper(fn, *args, **kwargs) [ 593.535893] env[61957]: ERROR nova.compute.manager [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 593.535893] env[61957]: ERROR nova.compute.manager [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] self.wait() [ 593.535893] env[61957]: ERROR nova.compute.manager [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 593.535893] env[61957]: ERROR nova.compute.manager [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] self[:] = self._gt.wait() [ 593.535893] env[61957]: ERROR nova.compute.manager [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 593.535893] env[61957]: ERROR nova.compute.manager [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] return self._exit_event.wait() [ 593.535893] env[61957]: ERROR nova.compute.manager [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 593.535893] env[61957]: ERROR nova.compute.manager [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] result = hub.switch() [ 593.535893] env[61957]: ERROR nova.compute.manager [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 593.535893] env[61957]: ERROR nova.compute.manager [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] return self.greenlet.switch() [ 593.536297] env[61957]: ERROR nova.compute.manager [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 593.536297] env[61957]: ERROR nova.compute.manager [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] result = function(*args, **kwargs) [ 593.536297] env[61957]: ERROR nova.compute.manager [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 593.536297] env[61957]: ERROR nova.compute.manager [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] return func(*args, **kwargs) [ 593.536297] env[61957]: ERROR nova.compute.manager [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 593.536297] env[61957]: ERROR nova.compute.manager [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] raise e [ 593.536297] env[61957]: ERROR nova.compute.manager [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 593.536297] env[61957]: ERROR nova.compute.manager [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] nwinfo = self.network_api.allocate_for_instance( [ 593.536297] env[61957]: ERROR nova.compute.manager [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 593.536297] env[61957]: ERROR nova.compute.manager [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] created_port_ids = self._update_ports_for_instance( [ 593.536297] env[61957]: ERROR nova.compute.manager [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 593.536297] env[61957]: ERROR nova.compute.manager [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] with excutils.save_and_reraise_exception(): [ 593.536297] env[61957]: ERROR nova.compute.manager [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 593.536672] env[61957]: ERROR nova.compute.manager [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] self.force_reraise() [ 593.536672] env[61957]: ERROR nova.compute.manager [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 593.536672] env[61957]: ERROR nova.compute.manager [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] raise self.value [ 593.536672] env[61957]: ERROR nova.compute.manager [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 593.536672] env[61957]: ERROR nova.compute.manager [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] updated_port = self._update_port( [ 593.536672] env[61957]: ERROR nova.compute.manager [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 593.536672] env[61957]: ERROR nova.compute.manager [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] _ensure_no_port_binding_failure(port) [ 593.536672] env[61957]: ERROR nova.compute.manager [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 593.536672] env[61957]: ERROR nova.compute.manager [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] raise exception.PortBindingFailed(port_id=port['id']) [ 593.536672] env[61957]: ERROR nova.compute.manager [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] nova.exception.PortBindingFailed: Binding failed for port 3d42e5dd-f713-445a-8855-3e0e765e2c4a, please check neutron logs for more information. [ 593.536672] env[61957]: ERROR nova.compute.manager [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] [ 593.536969] env[61957]: DEBUG nova.compute.utils [None req-d83719b8-6bd9-4ae7-99ea-7d0970401ba5 tempest-ServerExternalEventsTest-1886420864 tempest-ServerExternalEventsTest-1886420864-project-member] [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] Binding failed for port 3d42e5dd-f713-445a-8855-3e0e765e2c4a, please check neutron logs for more information. {{(pid=61957) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 593.538105] env[61957]: DEBUG nova.compute.manager [None req-d83719b8-6bd9-4ae7-99ea-7d0970401ba5 tempest-ServerExternalEventsTest-1886420864 tempest-ServerExternalEventsTest-1886420864-project-member] [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] Build of instance abdb0ba9-abef-4528-96fb-ae7d7d41f8fc was re-scheduled: Binding failed for port 3d42e5dd-f713-445a-8855-3e0e765e2c4a, please check neutron logs for more information. {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 593.539290] env[61957]: DEBUG nova.compute.manager [None req-d83719b8-6bd9-4ae7-99ea-7d0970401ba5 tempest-ServerExternalEventsTest-1886420864 tempest-ServerExternalEventsTest-1886420864-project-member] [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] Unplugging VIFs for instance {{(pid=61957) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 593.539290] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d83719b8-6bd9-4ae7-99ea-7d0970401ba5 tempest-ServerExternalEventsTest-1886420864 tempest-ServerExternalEventsTest-1886420864-project-member] Acquiring lock "refresh_cache-abdb0ba9-abef-4528-96fb-ae7d7d41f8fc" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 593.539429] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d83719b8-6bd9-4ae7-99ea-7d0970401ba5 tempest-ServerExternalEventsTest-1886420864 tempest-ServerExternalEventsTest-1886420864-project-member] Acquired lock "refresh_cache-abdb0ba9-abef-4528-96fb-ae7d7d41f8fc" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 593.539612] env[61957]: DEBUG nova.network.neutron [None req-d83719b8-6bd9-4ae7-99ea-7d0970401ba5 tempest-ServerExternalEventsTest-1886420864 tempest-ServerExternalEventsTest-1886420864-project-member] [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 593.540850] env[61957]: DEBUG oslo_concurrency.lockutils [None req-e4765209-cb68-4173-9918-09c43244c91d tempest-ServerMetadataTestJSON-1199197451 tempest-ServerMetadataTestJSON-1199197451-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.769s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 593.545187] env[61957]: INFO nova.compute.claims [None req-e4765209-cb68-4173-9918-09c43244c91d tempest-ServerMetadataTestJSON-1199197451 tempest-ServerMetadataTestJSON-1199197451-project-member] [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 593.606032] env[61957]: DEBUG oslo_vmware.api [None req-4b8bfc5c-9cfa-4b72-9c9f-71ae820b9856 tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] Task: {'id': task-1277305, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.107776} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 593.609218] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-4b8bfc5c-9cfa-4b72-9c9f-71ae820b9856 tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] [instance: c5af147e-3526-4014-98bc-7ad163ae89d9] Extended root virtual disk {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 593.609218] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42ee5e8a-9bf4-4a42-a5dd-e4d2fe7ea8ba {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.635374] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-4b8bfc5c-9cfa-4b72-9c9f-71ae820b9856 tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] [instance: c5af147e-3526-4014-98bc-7ad163ae89d9] Reconfiguring VM instance instance-00000007 to attach disk [datastore2] c5af147e-3526-4014-98bc-7ad163ae89d9/c5af147e-3526-4014-98bc-7ad163ae89d9.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 593.635374] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f944fa37-7be1-45a8-944f-424df29fa130 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.656296] env[61957]: DEBUG oslo_vmware.api [None req-4b8bfc5c-9cfa-4b72-9c9f-71ae820b9856 tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] Waiting for the task: (returnval){ [ 593.656296] env[61957]: value = "task-1277306" [ 593.656296] env[61957]: _type = "Task" [ 593.656296] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 593.666213] env[61957]: DEBUG oslo_vmware.api [None req-4b8bfc5c-9cfa-4b72-9c9f-71ae820b9856 tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] Task: {'id': task-1277306, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 593.821612] env[61957]: DEBUG oslo_concurrency.lockutils [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] Acquiring lock "74d3de77-1272-449d-8b64-75e21fff8d7f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 593.821851] env[61957]: DEBUG oslo_concurrency.lockutils [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] Lock "74d3de77-1272-449d-8b64-75e21fff8d7f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 593.874137] env[61957]: DEBUG oslo_concurrency.lockutils [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] Acquiring lock "b541c437-93b7-4e4d-9ff8-bfc9cdf0e237" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 593.874371] env[61957]: DEBUG oslo_concurrency.lockutils [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] Lock "b541c437-93b7-4e4d-9ff8-bfc9cdf0e237" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 593.937414] env[61957]: DEBUG oslo_concurrency.lockutils [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] Acquiring lock "394b1955-c448-42c2-a718-28df7bd366e9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 593.937414] env[61957]: DEBUG oslo_concurrency.lockutils [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] Lock "394b1955-c448-42c2-a718-28df7bd366e9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 593.998329] env[61957]: ERROR nova.compute.manager [None req-840b2d8a-f610-4c89-a367-da388d81d5aa tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 25e905db-2ea8-4412-bfad-1406d998af37, please check neutron logs for more information. [ 593.998329] env[61957]: ERROR nova.compute.manager Traceback (most recent call last): [ 593.998329] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 593.998329] env[61957]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 593.998329] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 593.998329] env[61957]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 593.998329] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 593.998329] env[61957]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 593.998329] env[61957]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 593.998329] env[61957]: ERROR nova.compute.manager self.force_reraise() [ 593.998329] env[61957]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 593.998329] env[61957]: ERROR nova.compute.manager raise self.value [ 593.998329] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 593.998329] env[61957]: ERROR nova.compute.manager updated_port = self._update_port( [ 593.998329] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 593.998329] env[61957]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 593.998880] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 593.998880] env[61957]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 593.998880] env[61957]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 25e905db-2ea8-4412-bfad-1406d998af37, please check neutron logs for more information. [ 593.998880] env[61957]: ERROR nova.compute.manager [ 593.998880] env[61957]: Traceback (most recent call last): [ 593.998880] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 593.998880] env[61957]: listener.cb(fileno) [ 593.998880] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 593.998880] env[61957]: result = function(*args, **kwargs) [ 593.998880] env[61957]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 593.998880] env[61957]: return func(*args, **kwargs) [ 593.998880] env[61957]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 593.998880] env[61957]: raise e [ 593.998880] env[61957]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 593.998880] env[61957]: nwinfo = self.network_api.allocate_for_instance( [ 593.998880] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 593.998880] env[61957]: created_port_ids = self._update_ports_for_instance( [ 593.998880] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 593.998880] env[61957]: with excutils.save_and_reraise_exception(): [ 593.998880] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 593.998880] env[61957]: self.force_reraise() [ 593.998880] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 593.998880] env[61957]: raise self.value [ 593.998880] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 593.998880] env[61957]: updated_port = self._update_port( [ 593.998880] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 593.998880] env[61957]: _ensure_no_port_binding_failure(port) [ 593.998880] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 593.998880] env[61957]: raise exception.PortBindingFailed(port_id=port['id']) [ 593.999838] env[61957]: nova.exception.PortBindingFailed: Binding failed for port 25e905db-2ea8-4412-bfad-1406d998af37, please check neutron logs for more information. [ 593.999838] env[61957]: Removing descriptor: 17 [ 593.999838] env[61957]: ERROR nova.compute.manager [None req-840b2d8a-f610-4c89-a367-da388d81d5aa tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 25e905db-2ea8-4412-bfad-1406d998af37, please check neutron logs for more information. [ 593.999838] env[61957]: ERROR nova.compute.manager [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] Traceback (most recent call last): [ 593.999838] env[61957]: ERROR nova.compute.manager [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 593.999838] env[61957]: ERROR nova.compute.manager [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] yield resources [ 593.999838] env[61957]: ERROR nova.compute.manager [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 593.999838] env[61957]: ERROR nova.compute.manager [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] self.driver.spawn(context, instance, image_meta, [ 593.999838] env[61957]: ERROR nova.compute.manager [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 593.999838] env[61957]: ERROR nova.compute.manager [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] self._vmops.spawn(context, instance, image_meta, injected_files, [ 593.999838] env[61957]: ERROR nova.compute.manager [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 593.999838] env[61957]: ERROR nova.compute.manager [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] vm_ref = self.build_virtual_machine(instance, [ 594.000243] env[61957]: ERROR nova.compute.manager [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 594.000243] env[61957]: ERROR nova.compute.manager [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] vif_infos = vmwarevif.get_vif_info(self._session, [ 594.000243] env[61957]: ERROR nova.compute.manager [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 594.000243] env[61957]: ERROR nova.compute.manager [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] for vif in network_info: [ 594.000243] env[61957]: ERROR nova.compute.manager [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 594.000243] env[61957]: ERROR nova.compute.manager [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] return self._sync_wrapper(fn, *args, **kwargs) [ 594.000243] env[61957]: ERROR nova.compute.manager [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 594.000243] env[61957]: ERROR nova.compute.manager [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] self.wait() [ 594.000243] env[61957]: ERROR nova.compute.manager [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 594.000243] env[61957]: ERROR nova.compute.manager [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] self[:] = self._gt.wait() [ 594.000243] env[61957]: ERROR nova.compute.manager [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 594.000243] env[61957]: ERROR nova.compute.manager [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] return self._exit_event.wait() [ 594.000243] env[61957]: ERROR nova.compute.manager [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 594.000661] env[61957]: ERROR nova.compute.manager [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] result = hub.switch() [ 594.000661] env[61957]: ERROR nova.compute.manager [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 594.000661] env[61957]: ERROR nova.compute.manager [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] return self.greenlet.switch() [ 594.000661] env[61957]: ERROR nova.compute.manager [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 594.000661] env[61957]: ERROR nova.compute.manager [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] result = function(*args, **kwargs) [ 594.000661] env[61957]: ERROR nova.compute.manager [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 594.000661] env[61957]: ERROR nova.compute.manager [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] return func(*args, **kwargs) [ 594.000661] env[61957]: ERROR nova.compute.manager [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 594.000661] env[61957]: ERROR nova.compute.manager [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] raise e [ 594.000661] env[61957]: ERROR nova.compute.manager [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 594.000661] env[61957]: ERROR nova.compute.manager [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] nwinfo = self.network_api.allocate_for_instance( [ 594.000661] env[61957]: ERROR nova.compute.manager [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 594.000661] env[61957]: ERROR nova.compute.manager [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] created_port_ids = self._update_ports_for_instance( [ 594.001086] env[61957]: ERROR nova.compute.manager [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 594.001086] env[61957]: ERROR nova.compute.manager [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] with excutils.save_and_reraise_exception(): [ 594.001086] env[61957]: ERROR nova.compute.manager [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 594.001086] env[61957]: ERROR nova.compute.manager [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] self.force_reraise() [ 594.001086] env[61957]: ERROR nova.compute.manager [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 594.001086] env[61957]: ERROR nova.compute.manager [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] raise self.value [ 594.001086] env[61957]: ERROR nova.compute.manager [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 594.001086] env[61957]: ERROR nova.compute.manager [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] updated_port = self._update_port( [ 594.001086] env[61957]: ERROR nova.compute.manager [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 594.001086] env[61957]: ERROR nova.compute.manager [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] _ensure_no_port_binding_failure(port) [ 594.001086] env[61957]: ERROR nova.compute.manager [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 594.001086] env[61957]: ERROR nova.compute.manager [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] raise exception.PortBindingFailed(port_id=port['id']) [ 594.001448] env[61957]: ERROR nova.compute.manager [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] nova.exception.PortBindingFailed: Binding failed for port 25e905db-2ea8-4412-bfad-1406d998af37, please check neutron logs for more information. [ 594.001448] env[61957]: ERROR nova.compute.manager [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] [ 594.001448] env[61957]: INFO nova.compute.manager [None req-840b2d8a-f610-4c89-a367-da388d81d5aa tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] Terminating instance [ 594.004355] env[61957]: INFO nova.compute.manager [None req-d102ca19-392e-4051-87d1-b58bed2da29d tempest-ServersAdminNegativeTestJSON-555579 tempest-ServersAdminNegativeTestJSON-555579-project-member] [instance: 37837096-5ad0-4b6f-b32d-f2899d88b6dd] Took 1.19 seconds to deallocate network for instance. [ 594.004809] env[61957]: DEBUG oslo_concurrency.lockutils [None req-840b2d8a-f610-4c89-a367-da388d81d5aa tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] Acquiring lock "refresh_cache-f9e04dbd-d5ae-4f03-9406-14b2caeaf010" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 594.004809] env[61957]: DEBUG oslo_concurrency.lockutils [None req-840b2d8a-f610-4c89-a367-da388d81d5aa tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] Acquired lock "refresh_cache-f9e04dbd-d5ae-4f03-9406-14b2caeaf010" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 594.004809] env[61957]: DEBUG nova.network.neutron [None req-840b2d8a-f610-4c89-a367-da388d81d5aa tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 594.100754] env[61957]: DEBUG nova.network.neutron [None req-d83719b8-6bd9-4ae7-99ea-7d0970401ba5 tempest-ServerExternalEventsTest-1886420864 tempest-ServerExternalEventsTest-1886420864-project-member] [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 594.170818] env[61957]: DEBUG oslo_vmware.api [None req-4b8bfc5c-9cfa-4b72-9c9f-71ae820b9856 tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] Task: {'id': task-1277306, 'name': ReconfigVM_Task, 'duration_secs': 0.434153} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 594.170818] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-4b8bfc5c-9cfa-4b72-9c9f-71ae820b9856 tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] [instance: c5af147e-3526-4014-98bc-7ad163ae89d9] Reconfigured VM instance instance-00000007 to attach disk [datastore2] c5af147e-3526-4014-98bc-7ad163ae89d9/c5af147e-3526-4014-98bc-7ad163ae89d9.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 594.171340] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6817f9dd-cc75-476e-9107-8460ae7eac96 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.179221] env[61957]: DEBUG oslo_vmware.api [None req-4b8bfc5c-9cfa-4b72-9c9f-71ae820b9856 tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] Waiting for the task: (returnval){ [ 594.179221] env[61957]: value = "task-1277307" [ 594.179221] env[61957]: _type = "Task" [ 594.179221] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 594.196695] env[61957]: DEBUG oslo_vmware.api [None req-4b8bfc5c-9cfa-4b72-9c9f-71ae820b9856 tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] Task: {'id': task-1277307, 'name': Rename_Task} progress is 6%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 594.235223] env[61957]: DEBUG nova.network.neutron [None req-d83719b8-6bd9-4ae7-99ea-7d0970401ba5 tempest-ServerExternalEventsTest-1886420864 tempest-ServerExternalEventsTest-1886420864-project-member] [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 594.544071] env[61957]: DEBUG nova.network.neutron [None req-840b2d8a-f610-4c89-a367-da388d81d5aa tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 594.668242] env[61957]: DEBUG nova.network.neutron [None req-840b2d8a-f610-4c89-a367-da388d81d5aa tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 594.694224] env[61957]: DEBUG oslo_vmware.api [None req-4b8bfc5c-9cfa-4b72-9c9f-71ae820b9856 tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] Task: {'id': task-1277307, 'name': Rename_Task, 'duration_secs': 0.133586} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 594.694653] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-4b8bfc5c-9cfa-4b72-9c9f-71ae820b9856 tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] [instance: c5af147e-3526-4014-98bc-7ad163ae89d9] Powering on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 594.694755] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d0caee0e-6edf-4959-adeb-0dabdff120f8 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.702311] env[61957]: DEBUG oslo_vmware.api [None req-4b8bfc5c-9cfa-4b72-9c9f-71ae820b9856 tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] Waiting for the task: (returnval){ [ 594.702311] env[61957]: value = "task-1277308" [ 594.702311] env[61957]: _type = "Task" [ 594.702311] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 594.713732] env[61957]: DEBUG oslo_vmware.api [None req-4b8bfc5c-9cfa-4b72-9c9f-71ae820b9856 tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] Task: {'id': task-1277308, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 594.737589] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d83719b8-6bd9-4ae7-99ea-7d0970401ba5 tempest-ServerExternalEventsTest-1886420864 tempest-ServerExternalEventsTest-1886420864-project-member] Releasing lock "refresh_cache-abdb0ba9-abef-4528-96fb-ae7d7d41f8fc" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 594.737792] env[61957]: DEBUG nova.compute.manager [None req-d83719b8-6bd9-4ae7-99ea-7d0970401ba5 tempest-ServerExternalEventsTest-1886420864 tempest-ServerExternalEventsTest-1886420864-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61957) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 594.737976] env[61957]: DEBUG nova.compute.manager [None req-d83719b8-6bd9-4ae7-99ea-7d0970401ba5 tempest-ServerExternalEventsTest-1886420864 tempest-ServerExternalEventsTest-1886420864-project-member] [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 594.738897] env[61957]: DEBUG nova.network.neutron [None req-d83719b8-6bd9-4ae7-99ea-7d0970401ba5 tempest-ServerExternalEventsTest-1886420864 tempest-ServerExternalEventsTest-1886420864-project-member] [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 594.784021] env[61957]: DEBUG nova.network.neutron [None req-d83719b8-6bd9-4ae7-99ea-7d0970401ba5 tempest-ServerExternalEventsTest-1886420864 tempest-ServerExternalEventsTest-1886420864-project-member] [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 594.905971] env[61957]: DEBUG nova.compute.manager [req-515e4691-9335-4a1b-9325-48477a55d90a req-cdea48c2-373b-4713-bd55-eab92a2a40a6 service nova] [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] Received event network-changed-25e905db-2ea8-4412-bfad-1406d998af37 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 594.905971] env[61957]: DEBUG nova.compute.manager [req-515e4691-9335-4a1b-9325-48477a55d90a req-cdea48c2-373b-4713-bd55-eab92a2a40a6 service nova] [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] Refreshing instance network info cache due to event network-changed-25e905db-2ea8-4412-bfad-1406d998af37. {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 594.905971] env[61957]: DEBUG oslo_concurrency.lockutils [req-515e4691-9335-4a1b-9325-48477a55d90a req-cdea48c2-373b-4713-bd55-eab92a2a40a6 service nova] Acquiring lock "refresh_cache-f9e04dbd-d5ae-4f03-9406-14b2caeaf010" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 594.976391] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d294ff68-8875-4944-8ce3-919e40c94899 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.986118] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f185fa0-0573-4558-95f3-de861f004233 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.021805] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cdf9b6c-d098-48e1-b9d8-058b447e90aa {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.030459] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f5302b5-3a25-4397-8de3-58540e995c6f {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.045072] env[61957]: DEBUG nova.compute.provider_tree [None req-e4765209-cb68-4173-9918-09c43244c91d tempest-ServerMetadataTestJSON-1199197451 tempest-ServerMetadataTestJSON-1199197451-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 595.047979] env[61957]: INFO nova.scheduler.client.report [None req-d102ca19-392e-4051-87d1-b58bed2da29d tempest-ServersAdminNegativeTestJSON-555579 tempest-ServersAdminNegativeTestJSON-555579-project-member] Deleted allocations for instance 37837096-5ad0-4b6f-b32d-f2899d88b6dd [ 595.175238] env[61957]: DEBUG oslo_concurrency.lockutils [None req-840b2d8a-f610-4c89-a367-da388d81d5aa tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] Releasing lock "refresh_cache-f9e04dbd-d5ae-4f03-9406-14b2caeaf010" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 595.175238] env[61957]: DEBUG nova.compute.manager [None req-840b2d8a-f610-4c89-a367-da388d81d5aa tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] Start destroying the instance on the hypervisor. {{(pid=61957) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 595.175238] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-840b2d8a-f610-4c89-a367-da388d81d5aa tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] Destroying instance {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 595.175238] env[61957]: DEBUG oslo_concurrency.lockutils [req-515e4691-9335-4a1b-9325-48477a55d90a req-cdea48c2-373b-4713-bd55-eab92a2a40a6 service nova] Acquired lock "refresh_cache-f9e04dbd-d5ae-4f03-9406-14b2caeaf010" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 595.175238] env[61957]: DEBUG nova.network.neutron [req-515e4691-9335-4a1b-9325-48477a55d90a req-cdea48c2-373b-4713-bd55-eab92a2a40a6 service nova] [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] Refreshing network info cache for port 25e905db-2ea8-4412-bfad-1406d998af37 {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 595.176288] env[61957]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2cc3010c-eb3b-4cae-a43a-0461fd1676a0 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.185374] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38ce9801-ad46-4961-a174-d6b7a8c9ba7e {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.215388] env[61957]: WARNING nova.virt.vmwareapi.vmops [None req-840b2d8a-f610-4c89-a367-da388d81d5aa tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance f9e04dbd-d5ae-4f03-9406-14b2caeaf010 could not be found. [ 595.215791] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-840b2d8a-f610-4c89-a367-da388d81d5aa tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] Instance destroyed {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 595.216068] env[61957]: INFO nova.compute.manager [None req-840b2d8a-f610-4c89-a367-da388d81d5aa tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] Took 0.04 seconds to destroy the instance on the hypervisor. [ 595.216465] env[61957]: DEBUG oslo.service.loopingcall [None req-840b2d8a-f610-4c89-a367-da388d81d5aa tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 595.221960] env[61957]: DEBUG nova.compute.manager [-] [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 595.221960] env[61957]: DEBUG nova.network.neutron [-] [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 595.224267] env[61957]: DEBUG oslo_vmware.api [None req-4b8bfc5c-9cfa-4b72-9c9f-71ae820b9856 tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] Task: {'id': task-1277308, 'name': PowerOnVM_Task, 'duration_secs': 0.448146} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 595.225209] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-4b8bfc5c-9cfa-4b72-9c9f-71ae820b9856 tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] [instance: c5af147e-3526-4014-98bc-7ad163ae89d9] Powered on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 595.225419] env[61957]: INFO nova.compute.manager [None req-4b8bfc5c-9cfa-4b72-9c9f-71ae820b9856 tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] [instance: c5af147e-3526-4014-98bc-7ad163ae89d9] Took 7.70 seconds to spawn the instance on the hypervisor. [ 595.226400] env[61957]: DEBUG nova.compute.manager [None req-4b8bfc5c-9cfa-4b72-9c9f-71ae820b9856 tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] [instance: c5af147e-3526-4014-98bc-7ad163ae89d9] Checking state {{(pid=61957) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 595.227618] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6479108-118a-4661-a051-fc38e2e3206c {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.255888] env[61957]: DEBUG nova.network.neutron [-] [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 595.287032] env[61957]: DEBUG nova.network.neutron [None req-d83719b8-6bd9-4ae7-99ea-7d0970401ba5 tempest-ServerExternalEventsTest-1886420864 tempest-ServerExternalEventsTest-1886420864-project-member] [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 595.558182] env[61957]: DEBUG nova.scheduler.client.report [None req-e4765209-cb68-4173-9918-09c43244c91d tempest-ServerMetadataTestJSON-1199197451 tempest-ServerMetadataTestJSON-1199197451-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 595.564351] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d102ca19-392e-4051-87d1-b58bed2da29d tempest-ServersAdminNegativeTestJSON-555579 tempest-ServersAdminNegativeTestJSON-555579-project-member] Lock "37837096-5ad0-4b6f-b32d-f2899d88b6dd" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 31.476s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 595.706426] env[61957]: DEBUG nova.network.neutron [req-515e4691-9335-4a1b-9325-48477a55d90a req-cdea48c2-373b-4713-bd55-eab92a2a40a6 service nova] [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 595.752861] env[61957]: INFO nova.compute.manager [None req-4b8bfc5c-9cfa-4b72-9c9f-71ae820b9856 tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] [instance: c5af147e-3526-4014-98bc-7ad163ae89d9] Took 19.37 seconds to build instance. [ 595.762070] env[61957]: DEBUG nova.network.neutron [-] [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 595.790037] env[61957]: INFO nova.compute.manager [None req-d83719b8-6bd9-4ae7-99ea-7d0970401ba5 tempest-ServerExternalEventsTest-1886420864 tempest-ServerExternalEventsTest-1886420864-project-member] [instance: abdb0ba9-abef-4528-96fb-ae7d7d41f8fc] Took 1.05 seconds to deallocate network for instance. [ 595.894601] env[61957]: DEBUG nova.network.neutron [req-515e4691-9335-4a1b-9325-48477a55d90a req-cdea48c2-373b-4713-bd55-eab92a2a40a6 service nova] [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 596.066892] env[61957]: DEBUG oslo_concurrency.lockutils [None req-e4765209-cb68-4173-9918-09c43244c91d tempest-ServerMetadataTestJSON-1199197451 tempest-ServerMetadataTestJSON-1199197451-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.525s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 596.066892] env[61957]: DEBUG nova.compute.manager [None req-e4765209-cb68-4173-9918-09c43244c91d tempest-ServerMetadataTestJSON-1199197451 tempest-ServerMetadataTestJSON-1199197451-project-member] [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] Start building networks asynchronously for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 596.072146] env[61957]: DEBUG oslo_concurrency.lockutils [None req-1b94927c-57d2-4b99-982c-53a5e69214ea tempest-VolumesAssistedSnapshotsTest-190365324 tempest-VolumesAssistedSnapshotsTest-190365324-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.997s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 596.074155] env[61957]: INFO nova.compute.claims [None req-1b94927c-57d2-4b99-982c-53a5e69214ea tempest-VolumesAssistedSnapshotsTest-190365324 tempest-VolumesAssistedSnapshotsTest-190365324-project-member] [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 596.076880] env[61957]: DEBUG nova.compute.manager [None req-1b6e3bde-e95e-4682-9e04-c3f69f3be74f tempest-ServersWithSpecificFlavorTestJSON-568377525 tempest-ServersWithSpecificFlavorTestJSON-568377525-project-member] [instance: f660c725-fc06-42e4-8a34-5d80568b535e] Starting instance... {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 596.257199] env[61957]: DEBUG oslo_concurrency.lockutils [None req-4b8bfc5c-9cfa-4b72-9c9f-71ae820b9856 tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] Lock "c5af147e-3526-4014-98bc-7ad163ae89d9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.886s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 596.262762] env[61957]: INFO nova.compute.manager [-] [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] Took 1.05 seconds to deallocate network for instance. [ 596.264976] env[61957]: DEBUG nova.compute.claims [None req-840b2d8a-f610-4c89-a367-da388d81d5aa tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] Aborting claim: {{(pid=61957) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 596.265169] env[61957]: DEBUG oslo_concurrency.lockutils [None req-840b2d8a-f610-4c89-a367-da388d81d5aa tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 596.397439] env[61957]: DEBUG oslo_concurrency.lockutils [req-515e4691-9335-4a1b-9325-48477a55d90a req-cdea48c2-373b-4713-bd55-eab92a2a40a6 service nova] Releasing lock "refresh_cache-f9e04dbd-d5ae-4f03-9406-14b2caeaf010" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 596.586166] env[61957]: DEBUG nova.compute.utils [None req-e4765209-cb68-4173-9918-09c43244c91d tempest-ServerMetadataTestJSON-1199197451 tempest-ServerMetadataTestJSON-1199197451-project-member] Using /dev/sd instead of None {{(pid=61957) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 596.589127] env[61957]: DEBUG nova.compute.manager [None req-e4765209-cb68-4173-9918-09c43244c91d tempest-ServerMetadataTestJSON-1199197451 tempest-ServerMetadataTestJSON-1199197451-project-member] [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] Allocating IP information in the background. {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 596.589642] env[61957]: DEBUG nova.network.neutron [None req-e4765209-cb68-4173-9918-09c43244c91d tempest-ServerMetadataTestJSON-1199197451 tempest-ServerMetadataTestJSON-1199197451-project-member] [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] allocate_for_instance() {{(pid=61957) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 596.621492] env[61957]: DEBUG oslo_concurrency.lockutils [None req-1b6e3bde-e95e-4682-9e04-c3f69f3be74f tempest-ServersWithSpecificFlavorTestJSON-568377525 tempest-ServersWithSpecificFlavorTestJSON-568377525-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 596.696932] env[61957]: DEBUG nova.policy [None req-e4765209-cb68-4173-9918-09c43244c91d tempest-ServerMetadataTestJSON-1199197451 tempest-ServerMetadataTestJSON-1199197451-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '97ba7439ffb843bcb7506b467ee505a3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ee202d2182eb4a82a1fcc9c61f782f49', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61957) authorize /opt/stack/nova/nova/policy.py:203}} [ 596.761506] env[61957]: DEBUG nova.compute.manager [None req-42e67e16-6377-41f9-9501-426c4335c678 tempest-ServersTestBootFromVolume-482289819 tempest-ServersTestBootFromVolume-482289819-project-member] [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] Starting instance... {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 596.830233] env[61957]: INFO nova.scheduler.client.report [None req-d83719b8-6bd9-4ae7-99ea-7d0970401ba5 tempest-ServerExternalEventsTest-1886420864 tempest-ServerExternalEventsTest-1886420864-project-member] Deleted allocations for instance abdb0ba9-abef-4528-96fb-ae7d7d41f8fc [ 597.089733] env[61957]: DEBUG nova.compute.manager [None req-e4765209-cb68-4173-9918-09c43244c91d tempest-ServerMetadataTestJSON-1199197451 tempest-ServerMetadataTestJSON-1199197451-project-member] [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] Start building block device mappings for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 597.247081] env[61957]: DEBUG nova.network.neutron [None req-e4765209-cb68-4173-9918-09c43244c91d tempest-ServerMetadataTestJSON-1199197451 tempest-ServerMetadataTestJSON-1199197451-project-member] [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] Successfully created port: 8834628f-9270-4eee-9918-dfc555410c7c {{(pid=61957) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 597.296221] env[61957]: DEBUG oslo_concurrency.lockutils [None req-42e67e16-6377-41f9-9501-426c4335c678 tempest-ServersTestBootFromVolume-482289819 tempest-ServersTestBootFromVolume-482289819-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 597.347812] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d83719b8-6bd9-4ae7-99ea-7d0970401ba5 tempest-ServerExternalEventsTest-1886420864 tempest-ServerExternalEventsTest-1886420864-project-member] Lock "abdb0ba9-abef-4528-96fb-ae7d7d41f8fc" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 32.077s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 597.434671] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b94f54c1-4452-44de-b873-163d183b75fc {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.446219] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5288bafc-9c0c-4628-aa7d-1dadc8b2ac79 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.486790] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab6d38e6-e99a-48e7-8bd4-bcf86a9a9188 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.497782] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31a2251b-3fdf-4a2b-ae54-9814f5e6b14b {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.513332] env[61957]: DEBUG nova.compute.provider_tree [None req-1b94927c-57d2-4b99-982c-53a5e69214ea tempest-VolumesAssistedSnapshotsTest-190365324 tempest-VolumesAssistedSnapshotsTest-190365324-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 597.518579] env[61957]: DEBUG oslo_concurrency.lockutils [None req-b1f62379-e574-4175-a099-f938af5ddaa8 tempest-ServerPasswordTestJSON-1543821420 tempest-ServerPasswordTestJSON-1543821420-project-member] Acquiring lock "49e99297-ffb9-4104-bd06-911243908828" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 597.518685] env[61957]: DEBUG oslo_concurrency.lockutils [None req-b1f62379-e574-4175-a099-f938af5ddaa8 tempest-ServerPasswordTestJSON-1543821420 tempest-ServerPasswordTestJSON-1543821420-project-member] Lock "49e99297-ffb9-4104-bd06-911243908828" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 597.854946] env[61957]: DEBUG nova.compute.manager [None req-260de5ae-9540-404d-a23e-9facd81096ee tempest-AttachInterfacesUnderV243Test-1229304814 tempest-AttachInterfacesUnderV243Test-1229304814-project-member] [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] Starting instance... {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 598.020835] env[61957]: DEBUG nova.scheduler.client.report [None req-1b94927c-57d2-4b99-982c-53a5e69214ea tempest-VolumesAssistedSnapshotsTest-190365324 tempest-VolumesAssistedSnapshotsTest-190365324-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 598.105968] env[61957]: DEBUG nova.compute.manager [None req-e4765209-cb68-4173-9918-09c43244c91d tempest-ServerMetadataTestJSON-1199197451 tempest-ServerMetadataTestJSON-1199197451-project-member] [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] Start spawning the instance on the hypervisor. {{(pid=61957) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 598.141918] env[61957]: DEBUG nova.virt.hardware [None req-e4765209-cb68-4173-9918-09c43244c91d tempest-ServerMetadataTestJSON-1199197451 tempest-ServerMetadataTestJSON-1199197451-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T17:22:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T17:21:50Z,direct_url=,disk_format='vmdk',id=11c76a2c-f705-470a-ba9d-4657858bab38,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='81d3096d0f094373913fc3dc8f5c3c6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T17:21:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 598.143617] env[61957]: DEBUG nova.virt.hardware [None req-e4765209-cb68-4173-9918-09c43244c91d tempest-ServerMetadataTestJSON-1199197451 tempest-ServerMetadataTestJSON-1199197451-project-member] Flavor limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 598.143617] env[61957]: DEBUG nova.virt.hardware [None req-e4765209-cb68-4173-9918-09c43244c91d tempest-ServerMetadataTestJSON-1199197451 tempest-ServerMetadataTestJSON-1199197451-project-member] Image limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 598.143617] env[61957]: DEBUG nova.virt.hardware [None req-e4765209-cb68-4173-9918-09c43244c91d tempest-ServerMetadataTestJSON-1199197451 tempest-ServerMetadataTestJSON-1199197451-project-member] Flavor pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 598.143617] env[61957]: DEBUG nova.virt.hardware [None req-e4765209-cb68-4173-9918-09c43244c91d tempest-ServerMetadataTestJSON-1199197451 tempest-ServerMetadataTestJSON-1199197451-project-member] Image pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 598.143617] env[61957]: DEBUG nova.virt.hardware [None req-e4765209-cb68-4173-9918-09c43244c91d tempest-ServerMetadataTestJSON-1199197451 tempest-ServerMetadataTestJSON-1199197451-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 598.143850] env[61957]: DEBUG nova.virt.hardware [None req-e4765209-cb68-4173-9918-09c43244c91d tempest-ServerMetadataTestJSON-1199197451 tempest-ServerMetadataTestJSON-1199197451-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 598.143850] env[61957]: DEBUG nova.virt.hardware [None req-e4765209-cb68-4173-9918-09c43244c91d tempest-ServerMetadataTestJSON-1199197451 tempest-ServerMetadataTestJSON-1199197451-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 598.144046] env[61957]: DEBUG nova.virt.hardware [None req-e4765209-cb68-4173-9918-09c43244c91d tempest-ServerMetadataTestJSON-1199197451 tempest-ServerMetadataTestJSON-1199197451-project-member] Got 1 possible topologies {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 598.144209] env[61957]: DEBUG nova.virt.hardware [None req-e4765209-cb68-4173-9918-09c43244c91d tempest-ServerMetadataTestJSON-1199197451 tempest-ServerMetadataTestJSON-1199197451-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 598.144376] env[61957]: DEBUG nova.virt.hardware [None req-e4765209-cb68-4173-9918-09c43244c91d tempest-ServerMetadataTestJSON-1199197451 tempest-ServerMetadataTestJSON-1199197451-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 598.145371] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd8e7860-7a6d-4155-b908-64d7fe0bb86c {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.157161] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6377d17a-ce2f-4890-9f6f-a2c3c97b8ac6 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.387771] env[61957]: DEBUG oslo_concurrency.lockutils [None req-260de5ae-9540-404d-a23e-9facd81096ee tempest-AttachInterfacesUnderV243Test-1229304814 tempest-AttachInterfacesUnderV243Test-1229304814-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 598.528353] env[61957]: DEBUG oslo_concurrency.lockutils [None req-1b94927c-57d2-4b99-982c-53a5e69214ea tempest-VolumesAssistedSnapshotsTest-190365324 tempest-VolumesAssistedSnapshotsTest-190365324-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.456s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 598.528894] env[61957]: DEBUG nova.compute.manager [None req-1b94927c-57d2-4b99-982c-53a5e69214ea tempest-VolumesAssistedSnapshotsTest-190365324 tempest-VolumesAssistedSnapshotsTest-190365324-project-member] [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] Start building networks asynchronously for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 598.532419] env[61957]: DEBUG oslo_concurrency.lockutils [None req-837238f2-cf29-45f6-8541-231ac85b718c tempest-ServerDiagnosticsTest-1677027478 tempest-ServerDiagnosticsTest-1677027478-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.194s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 598.983460] env[61957]: DEBUG oslo_concurrency.lockutils [None req-fc6b77c8-b5cd-4608-beab-d2cb5c9f7440 tempest-ImagesNegativeTestJSON-1548152881 tempest-ImagesNegativeTestJSON-1548152881-project-member] Acquiring lock "9df88f23-0219-43e0-b28a-e78f30a473a7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 598.983780] env[61957]: DEBUG oslo_concurrency.lockutils [None req-fc6b77c8-b5cd-4608-beab-d2cb5c9f7440 tempest-ImagesNegativeTestJSON-1548152881 tempest-ImagesNegativeTestJSON-1548152881-project-member] Lock "9df88f23-0219-43e0-b28a-e78f30a473a7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 599.043530] env[61957]: DEBUG nova.compute.utils [None req-1b94927c-57d2-4b99-982c-53a5e69214ea tempest-VolumesAssistedSnapshotsTest-190365324 tempest-VolumesAssistedSnapshotsTest-190365324-project-member] Using /dev/sd instead of None {{(pid=61957) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 599.045670] env[61957]: DEBUG nova.compute.manager [None req-1b94927c-57d2-4b99-982c-53a5e69214ea tempest-VolumesAssistedSnapshotsTest-190365324 tempest-VolumesAssistedSnapshotsTest-190365324-project-member] [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] Allocating IP information in the background. {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 599.052795] env[61957]: DEBUG nova.network.neutron [None req-1b94927c-57d2-4b99-982c-53a5e69214ea tempest-VolumesAssistedSnapshotsTest-190365324 tempest-VolumesAssistedSnapshotsTest-190365324-project-member] [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] allocate_for_instance() {{(pid=61957) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 599.132126] env[61957]: DEBUG nova.policy [None req-1b94927c-57d2-4b99-982c-53a5e69214ea tempest-VolumesAssistedSnapshotsTest-190365324 tempest-VolumesAssistedSnapshotsTest-190365324-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'aba0c97e57c74f0e8ccb92692c89c0f7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '61701640bb6b449fb0c4a6a981b76d77', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61957) authorize /opt/stack/nova/nova/policy.py:203}} [ 599.394026] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0ba5719-b125-4ba5-b633-024387db7d25 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.403960] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5143be9c-07be-49a8-abb6-88557bbd5f1b {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.441689] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85aa8783-2a29-46a4-a94a-0ea27fe7cd82 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.450363] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-666cf204-81b2-4934-97ac-7d3310cf0f80 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.464423] env[61957]: DEBUG nova.compute.provider_tree [None req-837238f2-cf29-45f6-8541-231ac85b718c tempest-ServerDiagnosticsTest-1677027478 tempest-ServerDiagnosticsTest-1677027478-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 599.554581] env[61957]: DEBUG nova.compute.manager [None req-1b94927c-57d2-4b99-982c-53a5e69214ea tempest-VolumesAssistedSnapshotsTest-190365324 tempest-VolumesAssistedSnapshotsTest-190365324-project-member] [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] Start building block device mappings for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 599.967371] env[61957]: DEBUG nova.compute.manager [req-59b616e8-ea15-4ed9-a55e-c66c343a2745 req-f00adb0d-008a-417a-895e-9e29fedceebe service nova] [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] Received event network-vif-deleted-25e905db-2ea8-4412-bfad-1406d998af37 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 599.968603] env[61957]: DEBUG nova.scheduler.client.report [None req-837238f2-cf29-45f6-8541-231ac85b718c tempest-ServerDiagnosticsTest-1677027478 tempest-ServerDiagnosticsTest-1677027478-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 600.254159] env[61957]: DEBUG nova.network.neutron [None req-1b94927c-57d2-4b99-982c-53a5e69214ea tempest-VolumesAssistedSnapshotsTest-190365324 tempest-VolumesAssistedSnapshotsTest-190365324-project-member] [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] Successfully created port: 6af72034-3abb-47c4-94e2-1b47e7c69c0a {{(pid=61957) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 600.331239] env[61957]: ERROR nova.compute.manager [None req-e4765209-cb68-4173-9918-09c43244c91d tempest-ServerMetadataTestJSON-1199197451 tempest-ServerMetadataTestJSON-1199197451-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 8834628f-9270-4eee-9918-dfc555410c7c, please check neutron logs for more information. [ 600.331239] env[61957]: ERROR nova.compute.manager Traceback (most recent call last): [ 600.331239] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 600.331239] env[61957]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 600.331239] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 600.331239] env[61957]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 600.331239] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 600.331239] env[61957]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 600.331239] env[61957]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 600.331239] env[61957]: ERROR nova.compute.manager self.force_reraise() [ 600.331239] env[61957]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 600.331239] env[61957]: ERROR nova.compute.manager raise self.value [ 600.331239] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 600.331239] env[61957]: ERROR nova.compute.manager updated_port = self._update_port( [ 600.331239] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 600.331239] env[61957]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 600.331683] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 600.331683] env[61957]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 600.331683] env[61957]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 8834628f-9270-4eee-9918-dfc555410c7c, please check neutron logs for more information. [ 600.331683] env[61957]: ERROR nova.compute.manager [ 600.331683] env[61957]: Traceback (most recent call last): [ 600.331683] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 600.331683] env[61957]: listener.cb(fileno) [ 600.331683] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 600.331683] env[61957]: result = function(*args, **kwargs) [ 600.331683] env[61957]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 600.331683] env[61957]: return func(*args, **kwargs) [ 600.331683] env[61957]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 600.331683] env[61957]: raise e [ 600.331683] env[61957]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 600.331683] env[61957]: nwinfo = self.network_api.allocate_for_instance( [ 600.331683] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 600.331683] env[61957]: created_port_ids = self._update_ports_for_instance( [ 600.331683] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 600.331683] env[61957]: with excutils.save_and_reraise_exception(): [ 600.331683] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 600.331683] env[61957]: self.force_reraise() [ 600.331683] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 600.331683] env[61957]: raise self.value [ 600.331683] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 600.331683] env[61957]: updated_port = self._update_port( [ 600.331683] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 600.331683] env[61957]: _ensure_no_port_binding_failure(port) [ 600.331683] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 600.331683] env[61957]: raise exception.PortBindingFailed(port_id=port['id']) [ 600.332450] env[61957]: nova.exception.PortBindingFailed: Binding failed for port 8834628f-9270-4eee-9918-dfc555410c7c, please check neutron logs for more information. [ 600.332450] env[61957]: Removing descriptor: 17 [ 600.332450] env[61957]: ERROR nova.compute.manager [None req-e4765209-cb68-4173-9918-09c43244c91d tempest-ServerMetadataTestJSON-1199197451 tempest-ServerMetadataTestJSON-1199197451-project-member] [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 8834628f-9270-4eee-9918-dfc555410c7c, please check neutron logs for more information. [ 600.332450] env[61957]: ERROR nova.compute.manager [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] Traceback (most recent call last): [ 600.332450] env[61957]: ERROR nova.compute.manager [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 600.332450] env[61957]: ERROR nova.compute.manager [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] yield resources [ 600.332450] env[61957]: ERROR nova.compute.manager [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 600.332450] env[61957]: ERROR nova.compute.manager [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] self.driver.spawn(context, instance, image_meta, [ 600.332450] env[61957]: ERROR nova.compute.manager [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 600.332450] env[61957]: ERROR nova.compute.manager [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 600.332450] env[61957]: ERROR nova.compute.manager [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 600.332450] env[61957]: ERROR nova.compute.manager [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] vm_ref = self.build_virtual_machine(instance, [ 600.332808] env[61957]: ERROR nova.compute.manager [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 600.332808] env[61957]: ERROR nova.compute.manager [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] vif_infos = vmwarevif.get_vif_info(self._session, [ 600.332808] env[61957]: ERROR nova.compute.manager [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 600.332808] env[61957]: ERROR nova.compute.manager [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] for vif in network_info: [ 600.332808] env[61957]: ERROR nova.compute.manager [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 600.332808] env[61957]: ERROR nova.compute.manager [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] return self._sync_wrapper(fn, *args, **kwargs) [ 600.332808] env[61957]: ERROR nova.compute.manager [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 600.332808] env[61957]: ERROR nova.compute.manager [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] self.wait() [ 600.332808] env[61957]: ERROR nova.compute.manager [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 600.332808] env[61957]: ERROR nova.compute.manager [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] self[:] = self._gt.wait() [ 600.332808] env[61957]: ERROR nova.compute.manager [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 600.332808] env[61957]: ERROR nova.compute.manager [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] return self._exit_event.wait() [ 600.332808] env[61957]: ERROR nova.compute.manager [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 600.333172] env[61957]: ERROR nova.compute.manager [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] result = hub.switch() [ 600.333172] env[61957]: ERROR nova.compute.manager [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 600.333172] env[61957]: ERROR nova.compute.manager [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] return self.greenlet.switch() [ 600.333172] env[61957]: ERROR nova.compute.manager [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 600.333172] env[61957]: ERROR nova.compute.manager [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] result = function(*args, **kwargs) [ 600.333172] env[61957]: ERROR nova.compute.manager [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 600.333172] env[61957]: ERROR nova.compute.manager [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] return func(*args, **kwargs) [ 600.333172] env[61957]: ERROR nova.compute.manager [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 600.333172] env[61957]: ERROR nova.compute.manager [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] raise e [ 600.333172] env[61957]: ERROR nova.compute.manager [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 600.333172] env[61957]: ERROR nova.compute.manager [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] nwinfo = self.network_api.allocate_for_instance( [ 600.333172] env[61957]: ERROR nova.compute.manager [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 600.333172] env[61957]: ERROR nova.compute.manager [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] created_port_ids = self._update_ports_for_instance( [ 600.333487] env[61957]: ERROR nova.compute.manager [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 600.333487] env[61957]: ERROR nova.compute.manager [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] with excutils.save_and_reraise_exception(): [ 600.333487] env[61957]: ERROR nova.compute.manager [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 600.333487] env[61957]: ERROR nova.compute.manager [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] self.force_reraise() [ 600.333487] env[61957]: ERROR nova.compute.manager [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 600.333487] env[61957]: ERROR nova.compute.manager [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] raise self.value [ 600.333487] env[61957]: ERROR nova.compute.manager [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 600.333487] env[61957]: ERROR nova.compute.manager [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] updated_port = self._update_port( [ 600.333487] env[61957]: ERROR nova.compute.manager [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 600.333487] env[61957]: ERROR nova.compute.manager [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] _ensure_no_port_binding_failure(port) [ 600.333487] env[61957]: ERROR nova.compute.manager [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 600.333487] env[61957]: ERROR nova.compute.manager [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] raise exception.PortBindingFailed(port_id=port['id']) [ 600.333807] env[61957]: ERROR nova.compute.manager [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] nova.exception.PortBindingFailed: Binding failed for port 8834628f-9270-4eee-9918-dfc555410c7c, please check neutron logs for more information. [ 600.333807] env[61957]: ERROR nova.compute.manager [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] [ 600.333807] env[61957]: INFO nova.compute.manager [None req-e4765209-cb68-4173-9918-09c43244c91d tempest-ServerMetadataTestJSON-1199197451 tempest-ServerMetadataTestJSON-1199197451-project-member] [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] Terminating instance [ 600.333807] env[61957]: DEBUG oslo_concurrency.lockutils [None req-e4765209-cb68-4173-9918-09c43244c91d tempest-ServerMetadataTestJSON-1199197451 tempest-ServerMetadataTestJSON-1199197451-project-member] Acquiring lock "refresh_cache-d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 600.333926] env[61957]: DEBUG oslo_concurrency.lockutils [None req-e4765209-cb68-4173-9918-09c43244c91d tempest-ServerMetadataTestJSON-1199197451 tempest-ServerMetadataTestJSON-1199197451-project-member] Acquired lock "refresh_cache-d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 600.334666] env[61957]: DEBUG nova.network.neutron [None req-e4765209-cb68-4173-9918-09c43244c91d tempest-ServerMetadataTestJSON-1199197451 tempest-ServerMetadataTestJSON-1199197451-project-member] [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 600.478880] env[61957]: DEBUG oslo_concurrency.lockutils [None req-837238f2-cf29-45f6-8541-231ac85b718c tempest-ServerDiagnosticsTest-1677027478 tempest-ServerDiagnosticsTest-1677027478-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.945s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 600.478880] env[61957]: ERROR nova.compute.manager [None req-837238f2-cf29-45f6-8541-231ac85b718c tempest-ServerDiagnosticsTest-1677027478 tempest-ServerDiagnosticsTest-1677027478-project-member] [instance: 43942e78-7703-444d-91e3-bc7913299654] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 021f5ba1-78f4-4aa6-b030-243c371db8ac, please check neutron logs for more information. [ 600.478880] env[61957]: ERROR nova.compute.manager [instance: 43942e78-7703-444d-91e3-bc7913299654] Traceback (most recent call last): [ 600.478880] env[61957]: ERROR nova.compute.manager [instance: 43942e78-7703-444d-91e3-bc7913299654] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 600.478880] env[61957]: ERROR nova.compute.manager [instance: 43942e78-7703-444d-91e3-bc7913299654] self.driver.spawn(context, instance, image_meta, [ 600.478880] env[61957]: ERROR nova.compute.manager [instance: 43942e78-7703-444d-91e3-bc7913299654] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 600.478880] env[61957]: ERROR nova.compute.manager [instance: 43942e78-7703-444d-91e3-bc7913299654] self._vmops.spawn(context, instance, image_meta, injected_files, [ 600.478880] env[61957]: ERROR nova.compute.manager [instance: 43942e78-7703-444d-91e3-bc7913299654] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 600.478880] env[61957]: ERROR nova.compute.manager [instance: 43942e78-7703-444d-91e3-bc7913299654] vm_ref = self.build_virtual_machine(instance, [ 600.479179] env[61957]: ERROR nova.compute.manager [instance: 43942e78-7703-444d-91e3-bc7913299654] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 600.479179] env[61957]: ERROR nova.compute.manager [instance: 43942e78-7703-444d-91e3-bc7913299654] vif_infos = vmwarevif.get_vif_info(self._session, [ 600.479179] env[61957]: ERROR nova.compute.manager [instance: 43942e78-7703-444d-91e3-bc7913299654] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 600.479179] env[61957]: ERROR nova.compute.manager [instance: 43942e78-7703-444d-91e3-bc7913299654] for vif in network_info: [ 600.479179] env[61957]: ERROR nova.compute.manager [instance: 43942e78-7703-444d-91e3-bc7913299654] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 600.479179] env[61957]: ERROR nova.compute.manager [instance: 43942e78-7703-444d-91e3-bc7913299654] return self._sync_wrapper(fn, *args, **kwargs) [ 600.479179] env[61957]: ERROR nova.compute.manager [instance: 43942e78-7703-444d-91e3-bc7913299654] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 600.479179] env[61957]: ERROR nova.compute.manager [instance: 43942e78-7703-444d-91e3-bc7913299654] self.wait() [ 600.479179] env[61957]: ERROR nova.compute.manager [instance: 43942e78-7703-444d-91e3-bc7913299654] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 600.479179] env[61957]: ERROR nova.compute.manager [instance: 43942e78-7703-444d-91e3-bc7913299654] self[:] = self._gt.wait() [ 600.479179] env[61957]: ERROR nova.compute.manager [instance: 43942e78-7703-444d-91e3-bc7913299654] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 600.479179] env[61957]: ERROR nova.compute.manager [instance: 43942e78-7703-444d-91e3-bc7913299654] return self._exit_event.wait() [ 600.479179] env[61957]: ERROR nova.compute.manager [instance: 43942e78-7703-444d-91e3-bc7913299654] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 600.479528] env[61957]: ERROR nova.compute.manager [instance: 43942e78-7703-444d-91e3-bc7913299654] result = hub.switch() [ 600.479528] env[61957]: ERROR nova.compute.manager [instance: 43942e78-7703-444d-91e3-bc7913299654] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 600.479528] env[61957]: ERROR nova.compute.manager [instance: 43942e78-7703-444d-91e3-bc7913299654] return self.greenlet.switch() [ 600.479528] env[61957]: ERROR nova.compute.manager [instance: 43942e78-7703-444d-91e3-bc7913299654] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 600.479528] env[61957]: ERROR nova.compute.manager [instance: 43942e78-7703-444d-91e3-bc7913299654] result = function(*args, **kwargs) [ 600.479528] env[61957]: ERROR nova.compute.manager [instance: 43942e78-7703-444d-91e3-bc7913299654] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 600.479528] env[61957]: ERROR nova.compute.manager [instance: 43942e78-7703-444d-91e3-bc7913299654] return func(*args, **kwargs) [ 600.479528] env[61957]: ERROR nova.compute.manager [instance: 43942e78-7703-444d-91e3-bc7913299654] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 600.479528] env[61957]: ERROR nova.compute.manager [instance: 43942e78-7703-444d-91e3-bc7913299654] raise e [ 600.479528] env[61957]: ERROR nova.compute.manager [instance: 43942e78-7703-444d-91e3-bc7913299654] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 600.479528] env[61957]: ERROR nova.compute.manager [instance: 43942e78-7703-444d-91e3-bc7913299654] nwinfo = self.network_api.allocate_for_instance( [ 600.479528] env[61957]: ERROR nova.compute.manager [instance: 43942e78-7703-444d-91e3-bc7913299654] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 600.479528] env[61957]: ERROR nova.compute.manager [instance: 43942e78-7703-444d-91e3-bc7913299654] created_port_ids = self._update_ports_for_instance( [ 600.479950] env[61957]: ERROR nova.compute.manager [instance: 43942e78-7703-444d-91e3-bc7913299654] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 600.479950] env[61957]: ERROR nova.compute.manager [instance: 43942e78-7703-444d-91e3-bc7913299654] with excutils.save_and_reraise_exception(): [ 600.479950] env[61957]: ERROR nova.compute.manager [instance: 43942e78-7703-444d-91e3-bc7913299654] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 600.479950] env[61957]: ERROR nova.compute.manager [instance: 43942e78-7703-444d-91e3-bc7913299654] self.force_reraise() [ 600.479950] env[61957]: ERROR nova.compute.manager [instance: 43942e78-7703-444d-91e3-bc7913299654] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 600.479950] env[61957]: ERROR nova.compute.manager [instance: 43942e78-7703-444d-91e3-bc7913299654] raise self.value [ 600.479950] env[61957]: ERROR nova.compute.manager [instance: 43942e78-7703-444d-91e3-bc7913299654] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 600.479950] env[61957]: ERROR nova.compute.manager [instance: 43942e78-7703-444d-91e3-bc7913299654] updated_port = self._update_port( [ 600.479950] env[61957]: ERROR nova.compute.manager [instance: 43942e78-7703-444d-91e3-bc7913299654] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 600.479950] env[61957]: ERROR nova.compute.manager [instance: 43942e78-7703-444d-91e3-bc7913299654] _ensure_no_port_binding_failure(port) [ 600.479950] env[61957]: ERROR nova.compute.manager [instance: 43942e78-7703-444d-91e3-bc7913299654] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 600.479950] env[61957]: ERROR nova.compute.manager [instance: 43942e78-7703-444d-91e3-bc7913299654] raise exception.PortBindingFailed(port_id=port['id']) [ 600.480369] env[61957]: ERROR nova.compute.manager [instance: 43942e78-7703-444d-91e3-bc7913299654] nova.exception.PortBindingFailed: Binding failed for port 021f5ba1-78f4-4aa6-b030-243c371db8ac, please check neutron logs for more information. [ 600.480369] env[61957]: ERROR nova.compute.manager [instance: 43942e78-7703-444d-91e3-bc7913299654] [ 600.480844] env[61957]: DEBUG nova.compute.utils [None req-837238f2-cf29-45f6-8541-231ac85b718c tempest-ServerDiagnosticsTest-1677027478 tempest-ServerDiagnosticsTest-1677027478-project-member] [instance: 43942e78-7703-444d-91e3-bc7913299654] Binding failed for port 021f5ba1-78f4-4aa6-b030-243c371db8ac, please check neutron logs for more information. {{(pid=61957) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 600.486255] env[61957]: DEBUG oslo_concurrency.lockutils [None req-f19b0e78-2c52-4a40-8c7e-832f687868ae tempest-ServersListShow296Test-1066136130 tempest-ServersListShow296Test-1066136130-project-member] Acquiring lock "af0a4907-4206-467f-98b9-a13b99fdf432" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 600.486566] env[61957]: DEBUG oslo_concurrency.lockutils [None req-f19b0e78-2c52-4a40-8c7e-832f687868ae tempest-ServersListShow296Test-1066136130 tempest-ServersListShow296Test-1066136130-project-member] Lock "af0a4907-4206-467f-98b9-a13b99fdf432" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.004s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 600.487089] env[61957]: DEBUG nova.compute.manager [None req-837238f2-cf29-45f6-8541-231ac85b718c tempest-ServerDiagnosticsTest-1677027478 tempest-ServerDiagnosticsTest-1677027478-project-member] [instance: 43942e78-7703-444d-91e3-bc7913299654] Build of instance 43942e78-7703-444d-91e3-bc7913299654 was re-scheduled: Binding failed for port 021f5ba1-78f4-4aa6-b030-243c371db8ac, please check neutron logs for more information. {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 600.487612] env[61957]: DEBUG nova.compute.manager [None req-837238f2-cf29-45f6-8541-231ac85b718c tempest-ServerDiagnosticsTest-1677027478 tempest-ServerDiagnosticsTest-1677027478-project-member] [instance: 43942e78-7703-444d-91e3-bc7913299654] Unplugging VIFs for instance {{(pid=61957) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 600.487924] env[61957]: DEBUG oslo_concurrency.lockutils [None req-837238f2-cf29-45f6-8541-231ac85b718c tempest-ServerDiagnosticsTest-1677027478 tempest-ServerDiagnosticsTest-1677027478-project-member] Acquiring lock "refresh_cache-43942e78-7703-444d-91e3-bc7913299654" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 600.488053] env[61957]: DEBUG oslo_concurrency.lockutils [None req-837238f2-cf29-45f6-8541-231ac85b718c tempest-ServerDiagnosticsTest-1677027478 tempest-ServerDiagnosticsTest-1677027478-project-member] Acquired lock "refresh_cache-43942e78-7703-444d-91e3-bc7913299654" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 600.488313] env[61957]: DEBUG nova.network.neutron [None req-837238f2-cf29-45f6-8541-231ac85b718c tempest-ServerDiagnosticsTest-1677027478 tempest-ServerDiagnosticsTest-1677027478-project-member] [instance: 43942e78-7703-444d-91e3-bc7913299654] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 600.490186] env[61957]: DEBUG oslo_concurrency.lockutils [None req-61bcf42d-e42b-40a8-98f8-af8cb211ebed tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.518s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 600.494123] env[61957]: INFO nova.compute.claims [None req-61bcf42d-e42b-40a8-98f8-af8cb211ebed tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 600.549253] env[61957]: DEBUG nova.compute.manager [None req-c5dea730-9118-43a6-b65c-221dcce54c46 tempest-ServerDiagnosticsV248Test-248432757 tempest-ServerDiagnosticsV248Test-248432757-project-admin] [instance: c5af147e-3526-4014-98bc-7ad163ae89d9] Checking state {{(pid=61957) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 600.551605] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-945f3139-340f-42af-8638-bbfb5a0b8f5c {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.562962] env[61957]: INFO nova.compute.manager [None req-c5dea730-9118-43a6-b65c-221dcce54c46 tempest-ServerDiagnosticsV248Test-248432757 tempest-ServerDiagnosticsV248Test-248432757-project-admin] [instance: c5af147e-3526-4014-98bc-7ad163ae89d9] Retrieving diagnostics [ 600.563814] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c259ab6-f88d-4fd1-aaa0-232b05636129 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.571484] env[61957]: DEBUG nova.compute.manager [None req-1b94927c-57d2-4b99-982c-53a5e69214ea tempest-VolumesAssistedSnapshotsTest-190365324 tempest-VolumesAssistedSnapshotsTest-190365324-project-member] [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] Start spawning the instance on the hypervisor. {{(pid=61957) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 600.620191] env[61957]: DEBUG nova.virt.hardware [None req-1b94927c-57d2-4b99-982c-53a5e69214ea tempest-VolumesAssistedSnapshotsTest-190365324 tempest-VolumesAssistedSnapshotsTest-190365324-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T17:22:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T17:21:50Z,direct_url=,disk_format='vmdk',id=11c76a2c-f705-470a-ba9d-4657858bab38,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='81d3096d0f094373913fc3dc8f5c3c6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T17:21:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 600.620457] env[61957]: DEBUG nova.virt.hardware [None req-1b94927c-57d2-4b99-982c-53a5e69214ea tempest-VolumesAssistedSnapshotsTest-190365324 tempest-VolumesAssistedSnapshotsTest-190365324-project-member] Flavor limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 600.620808] env[61957]: DEBUG nova.virt.hardware [None req-1b94927c-57d2-4b99-982c-53a5e69214ea tempest-VolumesAssistedSnapshotsTest-190365324 tempest-VolumesAssistedSnapshotsTest-190365324-project-member] Image limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 600.621075] env[61957]: DEBUG nova.virt.hardware [None req-1b94927c-57d2-4b99-982c-53a5e69214ea tempest-VolumesAssistedSnapshotsTest-190365324 tempest-VolumesAssistedSnapshotsTest-190365324-project-member] Flavor pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 600.621298] env[61957]: DEBUG nova.virt.hardware [None req-1b94927c-57d2-4b99-982c-53a5e69214ea tempest-VolumesAssistedSnapshotsTest-190365324 tempest-VolumesAssistedSnapshotsTest-190365324-project-member] Image pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 600.621515] env[61957]: DEBUG nova.virt.hardware [None req-1b94927c-57d2-4b99-982c-53a5e69214ea tempest-VolumesAssistedSnapshotsTest-190365324 tempest-VolumesAssistedSnapshotsTest-190365324-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 600.621764] env[61957]: DEBUG nova.virt.hardware [None req-1b94927c-57d2-4b99-982c-53a5e69214ea tempest-VolumesAssistedSnapshotsTest-190365324 tempest-VolumesAssistedSnapshotsTest-190365324-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 600.622477] env[61957]: DEBUG nova.virt.hardware [None req-1b94927c-57d2-4b99-982c-53a5e69214ea tempest-VolumesAssistedSnapshotsTest-190365324 tempest-VolumesAssistedSnapshotsTest-190365324-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 600.622477] env[61957]: DEBUG nova.virt.hardware [None req-1b94927c-57d2-4b99-982c-53a5e69214ea tempest-VolumesAssistedSnapshotsTest-190365324 tempest-VolumesAssistedSnapshotsTest-190365324-project-member] Got 1 possible topologies {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 600.622477] env[61957]: DEBUG nova.virt.hardware [None req-1b94927c-57d2-4b99-982c-53a5e69214ea tempest-VolumesAssistedSnapshotsTest-190365324 tempest-VolumesAssistedSnapshotsTest-190365324-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 600.622477] env[61957]: DEBUG nova.virt.hardware [None req-1b94927c-57d2-4b99-982c-53a5e69214ea tempest-VolumesAssistedSnapshotsTest-190365324 tempest-VolumesAssistedSnapshotsTest-190365324-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 600.624362] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b2bbccd-2f9a-49a4-a248-c2832636c968 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.632525] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bd4b76b-323d-4ce9-8fe2-8e6e668a5b1c {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.878365] env[61957]: DEBUG nova.network.neutron [None req-e4765209-cb68-4173-9918-09c43244c91d tempest-ServerMetadataTestJSON-1199197451 tempest-ServerMetadataTestJSON-1199197451-project-member] [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 601.058115] env[61957]: DEBUG nova.network.neutron [None req-837238f2-cf29-45f6-8541-231ac85b718c tempest-ServerDiagnosticsTest-1677027478 tempest-ServerDiagnosticsTest-1677027478-project-member] [instance: 43942e78-7703-444d-91e3-bc7913299654] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 601.206211] env[61957]: DEBUG nova.network.neutron [None req-e4765209-cb68-4173-9918-09c43244c91d tempest-ServerMetadataTestJSON-1199197451 tempest-ServerMetadataTestJSON-1199197451-project-member] [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 601.266955] env[61957]: DEBUG nova.network.neutron [None req-837238f2-cf29-45f6-8541-231ac85b718c tempest-ServerDiagnosticsTest-1677027478 tempest-ServerDiagnosticsTest-1677027478-project-member] [instance: 43942e78-7703-444d-91e3-bc7913299654] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 601.708424] env[61957]: DEBUG oslo_concurrency.lockutils [None req-e4765209-cb68-4173-9918-09c43244c91d tempest-ServerMetadataTestJSON-1199197451 tempest-ServerMetadataTestJSON-1199197451-project-member] Releasing lock "refresh_cache-d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 601.708845] env[61957]: DEBUG nova.compute.manager [None req-e4765209-cb68-4173-9918-09c43244c91d tempest-ServerMetadataTestJSON-1199197451 tempest-ServerMetadataTestJSON-1199197451-project-member] [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] Start destroying the instance on the hypervisor. {{(pid=61957) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 601.709044] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-e4765209-cb68-4173-9918-09c43244c91d tempest-ServerMetadataTestJSON-1199197451 tempest-ServerMetadataTestJSON-1199197451-project-member] [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] Destroying instance {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 601.709545] env[61957]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f04c0b06-5aa7-4b16-b2ee-0ca82e9a5c15 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.720044] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9455674-a2fd-44bc-88ba-9b504167845f {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.750357] env[61957]: WARNING nova.virt.vmwareapi.vmops [None req-e4765209-cb68-4173-9918-09c43244c91d tempest-ServerMetadataTestJSON-1199197451 tempest-ServerMetadataTestJSON-1199197451-project-member] [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f could not be found. [ 601.751712] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-e4765209-cb68-4173-9918-09c43244c91d tempest-ServerMetadataTestJSON-1199197451 tempest-ServerMetadataTestJSON-1199197451-project-member] [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] Instance destroyed {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 601.751712] env[61957]: INFO nova.compute.manager [None req-e4765209-cb68-4173-9918-09c43244c91d tempest-ServerMetadataTestJSON-1199197451 tempest-ServerMetadataTestJSON-1199197451-project-member] [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] Took 0.04 seconds to destroy the instance on the hypervisor. [ 601.752519] env[61957]: DEBUG oslo.service.loopingcall [None req-e4765209-cb68-4173-9918-09c43244c91d tempest-ServerMetadataTestJSON-1199197451 tempest-ServerMetadataTestJSON-1199197451-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 601.755118] env[61957]: DEBUG nova.compute.manager [-] [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 601.755455] env[61957]: DEBUG nova.network.neutron [-] [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 601.771049] env[61957]: DEBUG oslo_concurrency.lockutils [None req-837238f2-cf29-45f6-8541-231ac85b718c tempest-ServerDiagnosticsTest-1677027478 tempest-ServerDiagnosticsTest-1677027478-project-member] Releasing lock "refresh_cache-43942e78-7703-444d-91e3-bc7913299654" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 601.771285] env[61957]: DEBUG nova.compute.manager [None req-837238f2-cf29-45f6-8541-231ac85b718c tempest-ServerDiagnosticsTest-1677027478 tempest-ServerDiagnosticsTest-1677027478-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61957) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 601.771451] env[61957]: DEBUG nova.compute.manager [None req-837238f2-cf29-45f6-8541-231ac85b718c tempest-ServerDiagnosticsTest-1677027478 tempest-ServerDiagnosticsTest-1677027478-project-member] [instance: 43942e78-7703-444d-91e3-bc7913299654] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 601.771615] env[61957]: DEBUG nova.network.neutron [None req-837238f2-cf29-45f6-8541-231ac85b718c tempest-ServerDiagnosticsTest-1677027478 tempest-ServerDiagnosticsTest-1677027478-project-member] [instance: 43942e78-7703-444d-91e3-bc7913299654] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 601.900562] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9aa3a880-82e8-4cbc-a7c1-b1d96f641f0e {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.910825] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc788bed-d25b-4e07-92d0-c8556fc0d1b6 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.949032] env[61957]: DEBUG nova.network.neutron [None req-837238f2-cf29-45f6-8541-231ac85b718c tempest-ServerDiagnosticsTest-1677027478 tempest-ServerDiagnosticsTest-1677027478-project-member] [instance: 43942e78-7703-444d-91e3-bc7913299654] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 601.951570] env[61957]: DEBUG nova.network.neutron [-] [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 601.953766] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fde16a0-3959-4fb8-a40e-56bdd7b3871d {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.963654] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7316231-b983-499a-a5e6-5553ecacd403 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.979728] env[61957]: DEBUG nova.compute.provider_tree [None req-61bcf42d-e42b-40a8-98f8-af8cb211ebed tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 602.452333] env[61957]: DEBUG nova.network.neutron [None req-837238f2-cf29-45f6-8541-231ac85b718c tempest-ServerDiagnosticsTest-1677027478 tempest-ServerDiagnosticsTest-1677027478-project-member] [instance: 43942e78-7703-444d-91e3-bc7913299654] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 602.454325] env[61957]: DEBUG nova.network.neutron [-] [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 602.485804] env[61957]: DEBUG nova.scheduler.client.report [None req-61bcf42d-e42b-40a8-98f8-af8cb211ebed tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 602.957781] env[61957]: INFO nova.compute.manager [None req-837238f2-cf29-45f6-8541-231ac85b718c tempest-ServerDiagnosticsTest-1677027478 tempest-ServerDiagnosticsTest-1677027478-project-member] [instance: 43942e78-7703-444d-91e3-bc7913299654] Took 1.19 seconds to deallocate network for instance. [ 602.961894] env[61957]: INFO nova.compute.manager [-] [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] Took 1.21 seconds to deallocate network for instance. [ 602.964258] env[61957]: DEBUG nova.compute.claims [None req-e4765209-cb68-4173-9918-09c43244c91d tempest-ServerMetadataTestJSON-1199197451 tempest-ServerMetadataTestJSON-1199197451-project-member] [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] Aborting claim: {{(pid=61957) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 602.964647] env[61957]: DEBUG oslo_concurrency.lockutils [None req-e4765209-cb68-4173-9918-09c43244c91d tempest-ServerMetadataTestJSON-1199197451 tempest-ServerMetadataTestJSON-1199197451-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 602.995578] env[61957]: DEBUG oslo_concurrency.lockutils [None req-61bcf42d-e42b-40a8-98f8-af8cb211ebed tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.506s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 602.998034] env[61957]: DEBUG nova.compute.manager [None req-61bcf42d-e42b-40a8-98f8-af8cb211ebed tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] Start building networks asynchronously for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 603.005114] env[61957]: DEBUG oslo_concurrency.lockutils [None req-423a0985-b17f-45fa-b4ad-54e09005ad46 tempest-TenantUsagesTestJSON-2011428287 tempest-TenantUsagesTestJSON-2011428287-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.669s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 603.232186] env[61957]: DEBUG nova.compute.manager [req-876b8062-f79f-45f7-a7a2-76e9f47d7c16 req-62c6cd3a-0434-4c03-9a8e-54c28a72da1b service nova] [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] Received event network-changed-8834628f-9270-4eee-9918-dfc555410c7c {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 603.232186] env[61957]: DEBUG nova.compute.manager [req-876b8062-f79f-45f7-a7a2-76e9f47d7c16 req-62c6cd3a-0434-4c03-9a8e-54c28a72da1b service nova] [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] Refreshing instance network info cache due to event network-changed-8834628f-9270-4eee-9918-dfc555410c7c. {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 603.232186] env[61957]: DEBUG oslo_concurrency.lockutils [req-876b8062-f79f-45f7-a7a2-76e9f47d7c16 req-62c6cd3a-0434-4c03-9a8e-54c28a72da1b service nova] Acquiring lock "refresh_cache-d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 603.232364] env[61957]: DEBUG oslo_concurrency.lockutils [req-876b8062-f79f-45f7-a7a2-76e9f47d7c16 req-62c6cd3a-0434-4c03-9a8e-54c28a72da1b service nova] Acquired lock "refresh_cache-d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 603.232590] env[61957]: DEBUG nova.network.neutron [req-876b8062-f79f-45f7-a7a2-76e9f47d7c16 req-62c6cd3a-0434-4c03-9a8e-54c28a72da1b service nova] [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] Refreshing network info cache for port 8834628f-9270-4eee-9918-dfc555410c7c {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 603.378929] env[61957]: ERROR nova.compute.manager [None req-1b94927c-57d2-4b99-982c-53a5e69214ea tempest-VolumesAssistedSnapshotsTest-190365324 tempest-VolumesAssistedSnapshotsTest-190365324-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 6af72034-3abb-47c4-94e2-1b47e7c69c0a, please check neutron logs for more information. [ 603.378929] env[61957]: ERROR nova.compute.manager Traceback (most recent call last): [ 603.378929] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 603.378929] env[61957]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 603.378929] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 603.378929] env[61957]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 603.378929] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 603.378929] env[61957]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 603.378929] env[61957]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 603.378929] env[61957]: ERROR nova.compute.manager self.force_reraise() [ 603.378929] env[61957]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 603.378929] env[61957]: ERROR nova.compute.manager raise self.value [ 603.378929] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 603.378929] env[61957]: ERROR nova.compute.manager updated_port = self._update_port( [ 603.378929] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 603.378929] env[61957]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 603.379503] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 603.379503] env[61957]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 603.379503] env[61957]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 6af72034-3abb-47c4-94e2-1b47e7c69c0a, please check neutron logs for more information. [ 603.379503] env[61957]: ERROR nova.compute.manager [ 603.379503] env[61957]: Traceback (most recent call last): [ 603.379503] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 603.379503] env[61957]: listener.cb(fileno) [ 603.379503] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 603.379503] env[61957]: result = function(*args, **kwargs) [ 603.379503] env[61957]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 603.379503] env[61957]: return func(*args, **kwargs) [ 603.379503] env[61957]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 603.379503] env[61957]: raise e [ 603.379503] env[61957]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 603.379503] env[61957]: nwinfo = self.network_api.allocate_for_instance( [ 603.379503] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 603.379503] env[61957]: created_port_ids = self._update_ports_for_instance( [ 603.379503] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 603.379503] env[61957]: with excutils.save_and_reraise_exception(): [ 603.379503] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 603.379503] env[61957]: self.force_reraise() [ 603.379503] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 603.379503] env[61957]: raise self.value [ 603.379503] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 603.379503] env[61957]: updated_port = self._update_port( [ 603.379503] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 603.379503] env[61957]: _ensure_no_port_binding_failure(port) [ 603.379503] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 603.379503] env[61957]: raise exception.PortBindingFailed(port_id=port['id']) [ 603.381146] env[61957]: nova.exception.PortBindingFailed: Binding failed for port 6af72034-3abb-47c4-94e2-1b47e7c69c0a, please check neutron logs for more information. [ 603.381146] env[61957]: Removing descriptor: 15 [ 603.381146] env[61957]: ERROR nova.compute.manager [None req-1b94927c-57d2-4b99-982c-53a5e69214ea tempest-VolumesAssistedSnapshotsTest-190365324 tempest-VolumesAssistedSnapshotsTest-190365324-project-member] [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 6af72034-3abb-47c4-94e2-1b47e7c69c0a, please check neutron logs for more information. [ 603.381146] env[61957]: ERROR nova.compute.manager [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] Traceback (most recent call last): [ 603.381146] env[61957]: ERROR nova.compute.manager [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 603.381146] env[61957]: ERROR nova.compute.manager [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] yield resources [ 603.381146] env[61957]: ERROR nova.compute.manager [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 603.381146] env[61957]: ERROR nova.compute.manager [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] self.driver.spawn(context, instance, image_meta, [ 603.381146] env[61957]: ERROR nova.compute.manager [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 603.381146] env[61957]: ERROR nova.compute.manager [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] self._vmops.spawn(context, instance, image_meta, injected_files, [ 603.381146] env[61957]: ERROR nova.compute.manager [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 603.381146] env[61957]: ERROR nova.compute.manager [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] vm_ref = self.build_virtual_machine(instance, [ 603.382853] env[61957]: ERROR nova.compute.manager [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 603.382853] env[61957]: ERROR nova.compute.manager [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] vif_infos = vmwarevif.get_vif_info(self._session, [ 603.382853] env[61957]: ERROR nova.compute.manager [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 603.382853] env[61957]: ERROR nova.compute.manager [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] for vif in network_info: [ 603.382853] env[61957]: ERROR nova.compute.manager [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 603.382853] env[61957]: ERROR nova.compute.manager [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] return self._sync_wrapper(fn, *args, **kwargs) [ 603.382853] env[61957]: ERROR nova.compute.manager [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 603.382853] env[61957]: ERROR nova.compute.manager [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] self.wait() [ 603.382853] env[61957]: ERROR nova.compute.manager [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 603.382853] env[61957]: ERROR nova.compute.manager [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] self[:] = self._gt.wait() [ 603.382853] env[61957]: ERROR nova.compute.manager [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 603.382853] env[61957]: ERROR nova.compute.manager [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] return self._exit_event.wait() [ 603.382853] env[61957]: ERROR nova.compute.manager [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 603.384031] env[61957]: ERROR nova.compute.manager [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] result = hub.switch() [ 603.384031] env[61957]: ERROR nova.compute.manager [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 603.384031] env[61957]: ERROR nova.compute.manager [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] return self.greenlet.switch() [ 603.384031] env[61957]: ERROR nova.compute.manager [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 603.384031] env[61957]: ERROR nova.compute.manager [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] result = function(*args, **kwargs) [ 603.384031] env[61957]: ERROR nova.compute.manager [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 603.384031] env[61957]: ERROR nova.compute.manager [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] return func(*args, **kwargs) [ 603.384031] env[61957]: ERROR nova.compute.manager [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 603.384031] env[61957]: ERROR nova.compute.manager [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] raise e [ 603.384031] env[61957]: ERROR nova.compute.manager [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 603.384031] env[61957]: ERROR nova.compute.manager [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] nwinfo = self.network_api.allocate_for_instance( [ 603.384031] env[61957]: ERROR nova.compute.manager [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 603.384031] env[61957]: ERROR nova.compute.manager [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] created_port_ids = self._update_ports_for_instance( [ 603.384630] env[61957]: ERROR nova.compute.manager [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 603.384630] env[61957]: ERROR nova.compute.manager [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] with excutils.save_and_reraise_exception(): [ 603.384630] env[61957]: ERROR nova.compute.manager [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 603.384630] env[61957]: ERROR nova.compute.manager [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] self.force_reraise() [ 603.384630] env[61957]: ERROR nova.compute.manager [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 603.384630] env[61957]: ERROR nova.compute.manager [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] raise self.value [ 603.384630] env[61957]: ERROR nova.compute.manager [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 603.384630] env[61957]: ERROR nova.compute.manager [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] updated_port = self._update_port( [ 603.384630] env[61957]: ERROR nova.compute.manager [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 603.384630] env[61957]: ERROR nova.compute.manager [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] _ensure_no_port_binding_failure(port) [ 603.384630] env[61957]: ERROR nova.compute.manager [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 603.384630] env[61957]: ERROR nova.compute.manager [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] raise exception.PortBindingFailed(port_id=port['id']) [ 603.384963] env[61957]: ERROR nova.compute.manager [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] nova.exception.PortBindingFailed: Binding failed for port 6af72034-3abb-47c4-94e2-1b47e7c69c0a, please check neutron logs for more information. [ 603.384963] env[61957]: ERROR nova.compute.manager [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] [ 603.384963] env[61957]: INFO nova.compute.manager [None req-1b94927c-57d2-4b99-982c-53a5e69214ea tempest-VolumesAssistedSnapshotsTest-190365324 tempest-VolumesAssistedSnapshotsTest-190365324-project-member] [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] Terminating instance [ 603.384963] env[61957]: DEBUG oslo_concurrency.lockutils [None req-1b94927c-57d2-4b99-982c-53a5e69214ea tempest-VolumesAssistedSnapshotsTest-190365324 tempest-VolumesAssistedSnapshotsTest-190365324-project-member] Acquiring lock "refresh_cache-ee479d3a-239f-426c-8bdb-484fdf8dcd76" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 603.384963] env[61957]: DEBUG oslo_concurrency.lockutils [None req-1b94927c-57d2-4b99-982c-53a5e69214ea tempest-VolumesAssistedSnapshotsTest-190365324 tempest-VolumesAssistedSnapshotsTest-190365324-project-member] Acquired lock "refresh_cache-ee479d3a-239f-426c-8bdb-484fdf8dcd76" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 603.384963] env[61957]: DEBUG nova.network.neutron [None req-1b94927c-57d2-4b99-982c-53a5e69214ea tempest-VolumesAssistedSnapshotsTest-190365324 tempest-VolumesAssistedSnapshotsTest-190365324-project-member] [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 603.511179] env[61957]: DEBUG nova.compute.utils [None req-61bcf42d-e42b-40a8-98f8-af8cb211ebed tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] Using /dev/sd instead of None {{(pid=61957) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 603.517989] env[61957]: DEBUG nova.compute.manager [None req-61bcf42d-e42b-40a8-98f8-af8cb211ebed tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] Allocating IP information in the background. {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 603.518252] env[61957]: DEBUG nova.network.neutron [None req-61bcf42d-e42b-40a8-98f8-af8cb211ebed tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] allocate_for_instance() {{(pid=61957) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 603.724597] env[61957]: DEBUG nova.policy [None req-61bcf42d-e42b-40a8-98f8-af8cb211ebed tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '167bf8ddcb234f65bc90bca1b3cd4bab', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'cd3e0ac57aa342ec98fc6a5fd06f2a30', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61957) authorize /opt/stack/nova/nova/policy.py:203}} [ 603.799091] env[61957]: DEBUG nova.network.neutron [req-876b8062-f79f-45f7-a7a2-76e9f47d7c16 req-62c6cd3a-0434-4c03-9a8e-54c28a72da1b service nova] [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 603.929566] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac95361d-9db8-4cda-9da2-f888bc5467ae {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.937872] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9350f70-52fc-4cab-94d1-f54bfc2cebab {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.946592] env[61957]: DEBUG nova.network.neutron [None req-1b94927c-57d2-4b99-982c-53a5e69214ea tempest-VolumesAssistedSnapshotsTest-190365324 tempest-VolumesAssistedSnapshotsTest-190365324-project-member] [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 603.989121] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1a26d0f-6bf1-46fc-b80c-41356b84072b {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.991998] env[61957]: DEBUG oslo_concurrency.lockutils [None req-3d56c405-76b3-4bdb-9787-ecf239aeeb0d tempest-ImagesOneServerNegativeTestJSON-1581220397 tempest-ImagesOneServerNegativeTestJSON-1581220397-project-member] Acquiring lock "2b151571-348f-4543-a0c5-afe6458b1973" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 603.992231] env[61957]: DEBUG oslo_concurrency.lockutils [None req-3d56c405-76b3-4bdb-9787-ecf239aeeb0d tempest-ImagesOneServerNegativeTestJSON-1581220397 tempest-ImagesOneServerNegativeTestJSON-1581220397-project-member] Lock "2b151571-348f-4543-a0c5-afe6458b1973" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 603.999609] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4430faed-1d3b-4b8e-afc2-06057d768bb6 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.018510] env[61957]: DEBUG nova.compute.provider_tree [None req-423a0985-b17f-45fa-b4ad-54e09005ad46 tempest-TenantUsagesTestJSON-2011428287 tempest-TenantUsagesTestJSON-2011428287-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 604.020605] env[61957]: INFO nova.scheduler.client.report [None req-837238f2-cf29-45f6-8541-231ac85b718c tempest-ServerDiagnosticsTest-1677027478 tempest-ServerDiagnosticsTest-1677027478-project-member] Deleted allocations for instance 43942e78-7703-444d-91e3-bc7913299654 [ 604.025667] env[61957]: DEBUG nova.compute.manager [None req-61bcf42d-e42b-40a8-98f8-af8cb211ebed tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] Start building block device mappings for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 604.204238] env[61957]: DEBUG nova.network.neutron [req-876b8062-f79f-45f7-a7a2-76e9f47d7c16 req-62c6cd3a-0434-4c03-9a8e-54c28a72da1b service nova] [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 604.276193] env[61957]: DEBUG nova.network.neutron [None req-1b94927c-57d2-4b99-982c-53a5e69214ea tempest-VolumesAssistedSnapshotsTest-190365324 tempest-VolumesAssistedSnapshotsTest-190365324-project-member] [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 604.530548] env[61957]: DEBUG nova.scheduler.client.report [None req-423a0985-b17f-45fa-b4ad-54e09005ad46 tempest-TenantUsagesTestJSON-2011428287 tempest-TenantUsagesTestJSON-2011428287-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 604.536965] env[61957]: DEBUG oslo_concurrency.lockutils [None req-837238f2-cf29-45f6-8541-231ac85b718c tempest-ServerDiagnosticsTest-1677027478 tempest-ServerDiagnosticsTest-1677027478-project-member] Lock "43942e78-7703-444d-91e3-bc7913299654" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 30.608s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 604.641207] env[61957]: DEBUG nova.network.neutron [None req-61bcf42d-e42b-40a8-98f8-af8cb211ebed tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] Successfully created port: 6a3b4c95-304d-4644-ba91-67780e2d82ae {{(pid=61957) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 604.708475] env[61957]: DEBUG oslo_concurrency.lockutils [req-876b8062-f79f-45f7-a7a2-76e9f47d7c16 req-62c6cd3a-0434-4c03-9a8e-54c28a72da1b service nova] Releasing lock "refresh_cache-d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 604.708807] env[61957]: DEBUG nova.compute.manager [req-876b8062-f79f-45f7-a7a2-76e9f47d7c16 req-62c6cd3a-0434-4c03-9a8e-54c28a72da1b service nova] [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] Received event network-vif-deleted-8834628f-9270-4eee-9918-dfc555410c7c {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 604.778645] env[61957]: DEBUG oslo_concurrency.lockutils [None req-1b94927c-57d2-4b99-982c-53a5e69214ea tempest-VolumesAssistedSnapshotsTest-190365324 tempest-VolumesAssistedSnapshotsTest-190365324-project-member] Releasing lock "refresh_cache-ee479d3a-239f-426c-8bdb-484fdf8dcd76" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 604.778645] env[61957]: DEBUG nova.compute.manager [None req-1b94927c-57d2-4b99-982c-53a5e69214ea tempest-VolumesAssistedSnapshotsTest-190365324 tempest-VolumesAssistedSnapshotsTest-190365324-project-member] [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] Start destroying the instance on the hypervisor. {{(pid=61957) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 604.778930] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-1b94927c-57d2-4b99-982c-53a5e69214ea tempest-VolumesAssistedSnapshotsTest-190365324 tempest-VolumesAssistedSnapshotsTest-190365324-project-member] [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] Destroying instance {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 604.779162] env[61957]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-063f9322-81c3-449f-ad2d-4860669970f8 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.789647] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f944367-175e-47a5-ad15-750cf8e1beca {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.822307] env[61957]: WARNING nova.virt.vmwareapi.vmops [None req-1b94927c-57d2-4b99-982c-53a5e69214ea tempest-VolumesAssistedSnapshotsTest-190365324 tempest-VolumesAssistedSnapshotsTest-190365324-project-member] [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance ee479d3a-239f-426c-8bdb-484fdf8dcd76 could not be found. [ 604.822466] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-1b94927c-57d2-4b99-982c-53a5e69214ea tempest-VolumesAssistedSnapshotsTest-190365324 tempest-VolumesAssistedSnapshotsTest-190365324-project-member] [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] Instance destroyed {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 604.822641] env[61957]: INFO nova.compute.manager [None req-1b94927c-57d2-4b99-982c-53a5e69214ea tempest-VolumesAssistedSnapshotsTest-190365324 tempest-VolumesAssistedSnapshotsTest-190365324-project-member] [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] Took 0.04 seconds to destroy the instance on the hypervisor. [ 604.822873] env[61957]: DEBUG oslo.service.loopingcall [None req-1b94927c-57d2-4b99-982c-53a5e69214ea tempest-VolumesAssistedSnapshotsTest-190365324 tempest-VolumesAssistedSnapshotsTest-190365324-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 604.823170] env[61957]: DEBUG nova.compute.manager [-] [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 604.823273] env[61957]: DEBUG nova.network.neutron [-] [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 604.845760] env[61957]: DEBUG nova.network.neutron [-] [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 605.038181] env[61957]: DEBUG oslo_concurrency.lockutils [None req-423a0985-b17f-45fa-b4ad-54e09005ad46 tempest-TenantUsagesTestJSON-2011428287 tempest-TenantUsagesTestJSON-2011428287-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.033s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 605.038902] env[61957]: ERROR nova.compute.manager [None req-423a0985-b17f-45fa-b4ad-54e09005ad46 tempest-TenantUsagesTestJSON-2011428287 tempest-TenantUsagesTestJSON-2011428287-project-member] [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port edd470af-41f9-45fb-a01b-74445e4d2f66, please check neutron logs for more information. [ 605.038902] env[61957]: ERROR nova.compute.manager [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] Traceback (most recent call last): [ 605.038902] env[61957]: ERROR nova.compute.manager [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 605.038902] env[61957]: ERROR nova.compute.manager [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] self.driver.spawn(context, instance, image_meta, [ 605.038902] env[61957]: ERROR nova.compute.manager [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 605.038902] env[61957]: ERROR nova.compute.manager [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 605.038902] env[61957]: ERROR nova.compute.manager [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 605.038902] env[61957]: ERROR nova.compute.manager [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] vm_ref = self.build_virtual_machine(instance, [ 605.038902] env[61957]: ERROR nova.compute.manager [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 605.038902] env[61957]: ERROR nova.compute.manager [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] vif_infos = vmwarevif.get_vif_info(self._session, [ 605.038902] env[61957]: ERROR nova.compute.manager [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 605.039314] env[61957]: ERROR nova.compute.manager [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] for vif in network_info: [ 605.039314] env[61957]: ERROR nova.compute.manager [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 605.039314] env[61957]: ERROR nova.compute.manager [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] return self._sync_wrapper(fn, *args, **kwargs) [ 605.039314] env[61957]: ERROR nova.compute.manager [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 605.039314] env[61957]: ERROR nova.compute.manager [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] self.wait() [ 605.039314] env[61957]: ERROR nova.compute.manager [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 605.039314] env[61957]: ERROR nova.compute.manager [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] self[:] = self._gt.wait() [ 605.039314] env[61957]: ERROR nova.compute.manager [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 605.039314] env[61957]: ERROR nova.compute.manager [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] return self._exit_event.wait() [ 605.039314] env[61957]: ERROR nova.compute.manager [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 605.039314] env[61957]: ERROR nova.compute.manager [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] result = hub.switch() [ 605.039314] env[61957]: ERROR nova.compute.manager [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 605.039314] env[61957]: ERROR nova.compute.manager [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] return self.greenlet.switch() [ 605.039797] env[61957]: ERROR nova.compute.manager [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 605.039797] env[61957]: ERROR nova.compute.manager [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] result = function(*args, **kwargs) [ 605.039797] env[61957]: ERROR nova.compute.manager [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 605.039797] env[61957]: ERROR nova.compute.manager [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] return func(*args, **kwargs) [ 605.039797] env[61957]: ERROR nova.compute.manager [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 605.039797] env[61957]: ERROR nova.compute.manager [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] raise e [ 605.039797] env[61957]: ERROR nova.compute.manager [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 605.039797] env[61957]: ERROR nova.compute.manager [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] nwinfo = self.network_api.allocate_for_instance( [ 605.039797] env[61957]: ERROR nova.compute.manager [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 605.039797] env[61957]: ERROR nova.compute.manager [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] created_port_ids = self._update_ports_for_instance( [ 605.039797] env[61957]: ERROR nova.compute.manager [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 605.039797] env[61957]: ERROR nova.compute.manager [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] with excutils.save_and_reraise_exception(): [ 605.039797] env[61957]: ERROR nova.compute.manager [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 605.040229] env[61957]: ERROR nova.compute.manager [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] self.force_reraise() [ 605.040229] env[61957]: ERROR nova.compute.manager [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 605.040229] env[61957]: ERROR nova.compute.manager [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] raise self.value [ 605.040229] env[61957]: ERROR nova.compute.manager [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 605.040229] env[61957]: ERROR nova.compute.manager [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] updated_port = self._update_port( [ 605.040229] env[61957]: ERROR nova.compute.manager [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 605.040229] env[61957]: ERROR nova.compute.manager [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] _ensure_no_port_binding_failure(port) [ 605.040229] env[61957]: ERROR nova.compute.manager [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 605.040229] env[61957]: ERROR nova.compute.manager [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] raise exception.PortBindingFailed(port_id=port['id']) [ 605.040229] env[61957]: ERROR nova.compute.manager [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] nova.exception.PortBindingFailed: Binding failed for port edd470af-41f9-45fb-a01b-74445e4d2f66, please check neutron logs for more information. [ 605.040229] env[61957]: ERROR nova.compute.manager [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] [ 605.040568] env[61957]: DEBUG nova.compute.utils [None req-423a0985-b17f-45fa-b4ad-54e09005ad46 tempest-TenantUsagesTestJSON-2011428287 tempest-TenantUsagesTestJSON-2011428287-project-member] [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] Binding failed for port edd470af-41f9-45fb-a01b-74445e4d2f66, please check neutron logs for more information. {{(pid=61957) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 605.040787] env[61957]: DEBUG oslo_concurrency.lockutils [None req-7527ee8e-e60c-46dd-bb63-3e8733859a46 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.791s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 605.043918] env[61957]: INFO nova.compute.claims [None req-7527ee8e-e60c-46dd-bb63-3e8733859a46 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] [instance: 4ad746d0-4076-4c97-8ea9-20e49fda59ca] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 605.049661] env[61957]: DEBUG nova.compute.manager [None req-423a0985-b17f-45fa-b4ad-54e09005ad46 tempest-TenantUsagesTestJSON-2011428287 tempest-TenantUsagesTestJSON-2011428287-project-member] [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] Build of instance ed456ff1-249c-45d0-a007-3fef96ae8a2d was re-scheduled: Binding failed for port edd470af-41f9-45fb-a01b-74445e4d2f66, please check neutron logs for more information. {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 605.049661] env[61957]: DEBUG nova.compute.manager [None req-423a0985-b17f-45fa-b4ad-54e09005ad46 tempest-TenantUsagesTestJSON-2011428287 tempest-TenantUsagesTestJSON-2011428287-project-member] [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] Unplugging VIFs for instance {{(pid=61957) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 605.049661] env[61957]: DEBUG oslo_concurrency.lockutils [None req-423a0985-b17f-45fa-b4ad-54e09005ad46 tempest-TenantUsagesTestJSON-2011428287 tempest-TenantUsagesTestJSON-2011428287-project-member] Acquiring lock "refresh_cache-ed456ff1-249c-45d0-a007-3fef96ae8a2d" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 605.049661] env[61957]: DEBUG oslo_concurrency.lockutils [None req-423a0985-b17f-45fa-b4ad-54e09005ad46 tempest-TenantUsagesTestJSON-2011428287 tempest-TenantUsagesTestJSON-2011428287-project-member] Acquired lock "refresh_cache-ed456ff1-249c-45d0-a007-3fef96ae8a2d" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 605.050251] env[61957]: DEBUG nova.network.neutron [None req-423a0985-b17f-45fa-b4ad-54e09005ad46 tempest-TenantUsagesTestJSON-2011428287 tempest-TenantUsagesTestJSON-2011428287-project-member] [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 605.050251] env[61957]: DEBUG nova.compute.manager [None req-61bcf42d-e42b-40a8-98f8-af8cb211ebed tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] Start spawning the instance on the hypervisor. {{(pid=61957) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 605.052980] env[61957]: DEBUG nova.compute.manager [None req-adf5e3c1-3919-48f3-8d1a-68a60974255e tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] [instance: 3c099888-d5dc-4cee-9729-311808694625] Starting instance... {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 605.078821] env[61957]: DEBUG nova.virt.hardware [None req-61bcf42d-e42b-40a8-98f8-af8cb211ebed tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T17:22:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T17:21:50Z,direct_url=,disk_format='vmdk',id=11c76a2c-f705-470a-ba9d-4657858bab38,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='81d3096d0f094373913fc3dc8f5c3c6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T17:21:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 605.079467] env[61957]: DEBUG nova.virt.hardware [None req-61bcf42d-e42b-40a8-98f8-af8cb211ebed tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] Flavor limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 605.079467] env[61957]: DEBUG nova.virt.hardware [None req-61bcf42d-e42b-40a8-98f8-af8cb211ebed tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] Image limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 605.079467] env[61957]: DEBUG nova.virt.hardware [None req-61bcf42d-e42b-40a8-98f8-af8cb211ebed tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] Flavor pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 605.079594] env[61957]: DEBUG nova.virt.hardware [None req-61bcf42d-e42b-40a8-98f8-af8cb211ebed tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] Image pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 605.079808] env[61957]: DEBUG nova.virt.hardware [None req-61bcf42d-e42b-40a8-98f8-af8cb211ebed tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 605.079911] env[61957]: DEBUG nova.virt.hardware [None req-61bcf42d-e42b-40a8-98f8-af8cb211ebed tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 605.080078] env[61957]: DEBUG nova.virt.hardware [None req-61bcf42d-e42b-40a8-98f8-af8cb211ebed tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 605.080298] env[61957]: DEBUG nova.virt.hardware [None req-61bcf42d-e42b-40a8-98f8-af8cb211ebed tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] Got 1 possible topologies {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 605.080425] env[61957]: DEBUG nova.virt.hardware [None req-61bcf42d-e42b-40a8-98f8-af8cb211ebed tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 605.080549] env[61957]: DEBUG nova.virt.hardware [None req-61bcf42d-e42b-40a8-98f8-af8cb211ebed tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 605.081702] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27c1f08e-ad7c-4563-b98f-18a82fc8ae7b {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.090304] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e352df5d-0b93-48b4-95a3-61388be28552 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.349642] env[61957]: DEBUG nova.network.neutron [-] [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 605.586409] env[61957]: DEBUG oslo_concurrency.lockutils [None req-adf5e3c1-3919-48f3-8d1a-68a60974255e tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 605.626286] env[61957]: DEBUG nova.network.neutron [None req-423a0985-b17f-45fa-b4ad-54e09005ad46 tempest-TenantUsagesTestJSON-2011428287 tempest-TenantUsagesTestJSON-2011428287-project-member] [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 605.770580] env[61957]: DEBUG nova.network.neutron [None req-423a0985-b17f-45fa-b4ad-54e09005ad46 tempest-TenantUsagesTestJSON-2011428287 tempest-TenantUsagesTestJSON-2011428287-project-member] [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 605.851381] env[61957]: INFO nova.compute.manager [-] [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] Took 1.03 seconds to deallocate network for instance. [ 605.854046] env[61957]: DEBUG nova.compute.claims [None req-1b94927c-57d2-4b99-982c-53a5e69214ea tempest-VolumesAssistedSnapshotsTest-190365324 tempest-VolumesAssistedSnapshotsTest-190365324-project-member] [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] Aborting claim: {{(pid=61957) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 605.854196] env[61957]: DEBUG oslo_concurrency.lockutils [None req-1b94927c-57d2-4b99-982c-53a5e69214ea tempest-VolumesAssistedSnapshotsTest-190365324 tempest-VolumesAssistedSnapshotsTest-190365324-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 606.231356] env[61957]: DEBUG oslo_concurrency.lockutils [None req-af238887-1f40-4ff3-bfc6-fdb11b6cd8bf tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Acquiring lock "188b0fd3-5d71-4feb-aca5-75a2bd28895a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 606.231356] env[61957]: DEBUG oslo_concurrency.lockutils [None req-af238887-1f40-4ff3-bfc6-fdb11b6cd8bf tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Lock "188b0fd3-5d71-4feb-aca5-75a2bd28895a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 606.273066] env[61957]: DEBUG oslo_concurrency.lockutils [None req-423a0985-b17f-45fa-b4ad-54e09005ad46 tempest-TenantUsagesTestJSON-2011428287 tempest-TenantUsagesTestJSON-2011428287-project-member] Releasing lock "refresh_cache-ed456ff1-249c-45d0-a007-3fef96ae8a2d" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 606.273313] env[61957]: DEBUG nova.compute.manager [None req-423a0985-b17f-45fa-b4ad-54e09005ad46 tempest-TenantUsagesTestJSON-2011428287 tempest-TenantUsagesTestJSON-2011428287-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61957) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 606.273482] env[61957]: DEBUG nova.compute.manager [None req-423a0985-b17f-45fa-b4ad-54e09005ad46 tempest-TenantUsagesTestJSON-2011428287 tempest-TenantUsagesTestJSON-2011428287-project-member] [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 606.273647] env[61957]: DEBUG nova.network.neutron [None req-423a0985-b17f-45fa-b4ad-54e09005ad46 tempest-TenantUsagesTestJSON-2011428287 tempest-TenantUsagesTestJSON-2011428287-project-member] [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 606.305998] env[61957]: DEBUG nova.network.neutron [None req-423a0985-b17f-45fa-b4ad-54e09005ad46 tempest-TenantUsagesTestJSON-2011428287 tempest-TenantUsagesTestJSON-2011428287-project-member] [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 606.436960] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e22ecb64-d74e-4564-a099-bda0613dbc1e {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.447252] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51276513-b79e-4ab7-89ae-6501585d7bf6 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.492565] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5c2b97d-b233-451d-89fd-7e8f00245840 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.501313] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59268624-f507-4fbf-848d-e813e74b71c5 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.518056] env[61957]: DEBUG nova.compute.provider_tree [None req-7527ee8e-e60c-46dd-bb63-3e8733859a46 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 606.623320] env[61957]: DEBUG nova.compute.manager [req-3feb9e57-51b3-49dc-bfd0-1d56e7ee5761 req-4bddd2fa-418a-4f85-bf38-0287d5f870c3 service nova] [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] Received event network-changed-6af72034-3abb-47c4-94e2-1b47e7c69c0a {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 606.623625] env[61957]: DEBUG nova.compute.manager [req-3feb9e57-51b3-49dc-bfd0-1d56e7ee5761 req-4bddd2fa-418a-4f85-bf38-0287d5f870c3 service nova] [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] Refreshing instance network info cache due to event network-changed-6af72034-3abb-47c4-94e2-1b47e7c69c0a. {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 606.623705] env[61957]: DEBUG oslo_concurrency.lockutils [req-3feb9e57-51b3-49dc-bfd0-1d56e7ee5761 req-4bddd2fa-418a-4f85-bf38-0287d5f870c3 service nova] Acquiring lock "refresh_cache-ee479d3a-239f-426c-8bdb-484fdf8dcd76" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 606.624084] env[61957]: DEBUG oslo_concurrency.lockutils [req-3feb9e57-51b3-49dc-bfd0-1d56e7ee5761 req-4bddd2fa-418a-4f85-bf38-0287d5f870c3 service nova] Acquired lock "refresh_cache-ee479d3a-239f-426c-8bdb-484fdf8dcd76" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 606.624084] env[61957]: DEBUG nova.network.neutron [req-3feb9e57-51b3-49dc-bfd0-1d56e7ee5761 req-4bddd2fa-418a-4f85-bf38-0287d5f870c3 service nova] [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] Refreshing network info cache for port 6af72034-3abb-47c4-94e2-1b47e7c69c0a {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 606.811189] env[61957]: DEBUG nova.network.neutron [None req-423a0985-b17f-45fa-b4ad-54e09005ad46 tempest-TenantUsagesTestJSON-2011428287 tempest-TenantUsagesTestJSON-2011428287-project-member] [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 606.991036] env[61957]: ERROR nova.compute.manager [None req-61bcf42d-e42b-40a8-98f8-af8cb211ebed tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 6a3b4c95-304d-4644-ba91-67780e2d82ae, please check neutron logs for more information. [ 606.991036] env[61957]: ERROR nova.compute.manager Traceback (most recent call last): [ 606.991036] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 606.991036] env[61957]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 606.991036] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 606.991036] env[61957]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 606.991036] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 606.991036] env[61957]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 606.991036] env[61957]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 606.991036] env[61957]: ERROR nova.compute.manager self.force_reraise() [ 606.991036] env[61957]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 606.991036] env[61957]: ERROR nova.compute.manager raise self.value [ 606.991036] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 606.991036] env[61957]: ERROR nova.compute.manager updated_port = self._update_port( [ 606.991036] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 606.991036] env[61957]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 606.991593] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 606.991593] env[61957]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 606.991593] env[61957]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 6a3b4c95-304d-4644-ba91-67780e2d82ae, please check neutron logs for more information. [ 606.991593] env[61957]: ERROR nova.compute.manager [ 606.991593] env[61957]: Traceback (most recent call last): [ 606.991593] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 606.991593] env[61957]: listener.cb(fileno) [ 606.991593] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 606.991593] env[61957]: result = function(*args, **kwargs) [ 606.991593] env[61957]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 606.991593] env[61957]: return func(*args, **kwargs) [ 606.991593] env[61957]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 606.991593] env[61957]: raise e [ 606.991593] env[61957]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 606.991593] env[61957]: nwinfo = self.network_api.allocate_for_instance( [ 606.991593] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 606.991593] env[61957]: created_port_ids = self._update_ports_for_instance( [ 606.991593] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 606.991593] env[61957]: with excutils.save_and_reraise_exception(): [ 606.991593] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 606.991593] env[61957]: self.force_reraise() [ 606.991593] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 606.991593] env[61957]: raise self.value [ 606.991593] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 606.991593] env[61957]: updated_port = self._update_port( [ 606.991593] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 606.991593] env[61957]: _ensure_no_port_binding_failure(port) [ 606.991593] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 606.991593] env[61957]: raise exception.PortBindingFailed(port_id=port['id']) [ 606.997286] env[61957]: nova.exception.PortBindingFailed: Binding failed for port 6a3b4c95-304d-4644-ba91-67780e2d82ae, please check neutron logs for more information. [ 606.997286] env[61957]: Removing descriptor: 15 [ 606.997286] env[61957]: ERROR nova.compute.manager [None req-61bcf42d-e42b-40a8-98f8-af8cb211ebed tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 6a3b4c95-304d-4644-ba91-67780e2d82ae, please check neutron logs for more information. [ 606.997286] env[61957]: ERROR nova.compute.manager [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] Traceback (most recent call last): [ 606.997286] env[61957]: ERROR nova.compute.manager [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 606.997286] env[61957]: ERROR nova.compute.manager [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] yield resources [ 606.997286] env[61957]: ERROR nova.compute.manager [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 606.997286] env[61957]: ERROR nova.compute.manager [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] self.driver.spawn(context, instance, image_meta, [ 606.997286] env[61957]: ERROR nova.compute.manager [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 606.997286] env[61957]: ERROR nova.compute.manager [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] self._vmops.spawn(context, instance, image_meta, injected_files, [ 606.997286] env[61957]: ERROR nova.compute.manager [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 606.997286] env[61957]: ERROR nova.compute.manager [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] vm_ref = self.build_virtual_machine(instance, [ 607.000286] env[61957]: ERROR nova.compute.manager [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 607.000286] env[61957]: ERROR nova.compute.manager [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] vif_infos = vmwarevif.get_vif_info(self._session, [ 607.000286] env[61957]: ERROR nova.compute.manager [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 607.000286] env[61957]: ERROR nova.compute.manager [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] for vif in network_info: [ 607.000286] env[61957]: ERROR nova.compute.manager [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 607.000286] env[61957]: ERROR nova.compute.manager [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] return self._sync_wrapper(fn, *args, **kwargs) [ 607.000286] env[61957]: ERROR nova.compute.manager [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 607.000286] env[61957]: ERROR nova.compute.manager [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] self.wait() [ 607.000286] env[61957]: ERROR nova.compute.manager [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 607.000286] env[61957]: ERROR nova.compute.manager [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] self[:] = self._gt.wait() [ 607.000286] env[61957]: ERROR nova.compute.manager [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 607.000286] env[61957]: ERROR nova.compute.manager [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] return self._exit_event.wait() [ 607.000286] env[61957]: ERROR nova.compute.manager [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 607.000872] env[61957]: ERROR nova.compute.manager [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] result = hub.switch() [ 607.000872] env[61957]: ERROR nova.compute.manager [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 607.000872] env[61957]: ERROR nova.compute.manager [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] return self.greenlet.switch() [ 607.000872] env[61957]: ERROR nova.compute.manager [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 607.000872] env[61957]: ERROR nova.compute.manager [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] result = function(*args, **kwargs) [ 607.000872] env[61957]: ERROR nova.compute.manager [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 607.000872] env[61957]: ERROR nova.compute.manager [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] return func(*args, **kwargs) [ 607.000872] env[61957]: ERROR nova.compute.manager [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 607.000872] env[61957]: ERROR nova.compute.manager [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] raise e [ 607.000872] env[61957]: ERROR nova.compute.manager [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 607.000872] env[61957]: ERROR nova.compute.manager [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] nwinfo = self.network_api.allocate_for_instance( [ 607.000872] env[61957]: ERROR nova.compute.manager [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 607.000872] env[61957]: ERROR nova.compute.manager [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] created_port_ids = self._update_ports_for_instance( [ 607.001541] env[61957]: ERROR nova.compute.manager [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 607.001541] env[61957]: ERROR nova.compute.manager [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] with excutils.save_and_reraise_exception(): [ 607.001541] env[61957]: ERROR nova.compute.manager [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 607.001541] env[61957]: ERROR nova.compute.manager [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] self.force_reraise() [ 607.001541] env[61957]: ERROR nova.compute.manager [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 607.001541] env[61957]: ERROR nova.compute.manager [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] raise self.value [ 607.001541] env[61957]: ERROR nova.compute.manager [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 607.001541] env[61957]: ERROR nova.compute.manager [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] updated_port = self._update_port( [ 607.001541] env[61957]: ERROR nova.compute.manager [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 607.001541] env[61957]: ERROR nova.compute.manager [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] _ensure_no_port_binding_failure(port) [ 607.001541] env[61957]: ERROR nova.compute.manager [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 607.001541] env[61957]: ERROR nova.compute.manager [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] raise exception.PortBindingFailed(port_id=port['id']) [ 607.001948] env[61957]: ERROR nova.compute.manager [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] nova.exception.PortBindingFailed: Binding failed for port 6a3b4c95-304d-4644-ba91-67780e2d82ae, please check neutron logs for more information. [ 607.001948] env[61957]: ERROR nova.compute.manager [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] [ 607.001948] env[61957]: INFO nova.compute.manager [None req-61bcf42d-e42b-40a8-98f8-af8cb211ebed tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] Terminating instance [ 607.001948] env[61957]: DEBUG oslo_concurrency.lockutils [None req-61bcf42d-e42b-40a8-98f8-af8cb211ebed tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] Acquiring lock "refresh_cache-96c6b747-7293-4252-a1d0-b0d7684c5529" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 607.001948] env[61957]: DEBUG oslo_concurrency.lockutils [None req-61bcf42d-e42b-40a8-98f8-af8cb211ebed tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] Acquired lock "refresh_cache-96c6b747-7293-4252-a1d0-b0d7684c5529" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 607.001948] env[61957]: DEBUG nova.network.neutron [None req-61bcf42d-e42b-40a8-98f8-af8cb211ebed tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 607.021270] env[61957]: DEBUG nova.scheduler.client.report [None req-7527ee8e-e60c-46dd-bb63-3e8733859a46 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 607.139308] env[61957]: DEBUG oslo_concurrency.lockutils [None req-74fa7fa3-fb2c-4064-80f1-1405089fca10 tempest-ServerMetadataNegativeTestJSON-1373515003 tempest-ServerMetadataNegativeTestJSON-1373515003-project-member] Acquiring lock "3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 607.140414] env[61957]: DEBUG oslo_concurrency.lockutils [None req-74fa7fa3-fb2c-4064-80f1-1405089fca10 tempest-ServerMetadataNegativeTestJSON-1373515003 tempest-ServerMetadataNegativeTestJSON-1373515003-project-member] Lock "3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.003s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 607.159659] env[61957]: DEBUG nova.network.neutron [req-3feb9e57-51b3-49dc-bfd0-1d56e7ee5761 req-4bddd2fa-418a-4f85-bf38-0287d5f870c3 service nova] [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 607.299115] env[61957]: DEBUG nova.network.neutron [req-3feb9e57-51b3-49dc-bfd0-1d56e7ee5761 req-4bddd2fa-418a-4f85-bf38-0287d5f870c3 service nova] [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 607.316800] env[61957]: INFO nova.compute.manager [None req-423a0985-b17f-45fa-b4ad-54e09005ad46 tempest-TenantUsagesTestJSON-2011428287 tempest-TenantUsagesTestJSON-2011428287-project-member] [instance: ed456ff1-249c-45d0-a007-3fef96ae8a2d] Took 1.04 seconds to deallocate network for instance. [ 607.526144] env[61957]: DEBUG oslo_concurrency.lockutils [None req-7527ee8e-e60c-46dd-bb63-3e8733859a46 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.485s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 607.526731] env[61957]: DEBUG nova.compute.manager [None req-7527ee8e-e60c-46dd-bb63-3e8733859a46 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] [instance: 4ad746d0-4076-4c97-8ea9-20e49fda59ca] Start building networks asynchronously for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 607.531555] env[61957]: DEBUG nova.network.neutron [None req-61bcf42d-e42b-40a8-98f8-af8cb211ebed tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 607.534104] env[61957]: DEBUG oslo_concurrency.lockutils [None req-840b2d8a-f610-4c89-a367-da388d81d5aa tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 11.269s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 607.647647] env[61957]: DEBUG nova.network.neutron [None req-61bcf42d-e42b-40a8-98f8-af8cb211ebed tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 607.802494] env[61957]: DEBUG oslo_concurrency.lockutils [req-3feb9e57-51b3-49dc-bfd0-1d56e7ee5761 req-4bddd2fa-418a-4f85-bf38-0287d5f870c3 service nova] Releasing lock "refresh_cache-ee479d3a-239f-426c-8bdb-484fdf8dcd76" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 607.802494] env[61957]: DEBUG nova.compute.manager [req-3feb9e57-51b3-49dc-bfd0-1d56e7ee5761 req-4bddd2fa-418a-4f85-bf38-0287d5f870c3 service nova] [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] Received event network-vif-deleted-6af72034-3abb-47c4-94e2-1b47e7c69c0a {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 608.036892] env[61957]: DEBUG nova.compute.utils [None req-7527ee8e-e60c-46dd-bb63-3e8733859a46 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] Using /dev/sd instead of None {{(pid=61957) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 608.039779] env[61957]: DEBUG nova.compute.manager [None req-7527ee8e-e60c-46dd-bb63-3e8733859a46 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] [instance: 4ad746d0-4076-4c97-8ea9-20e49fda59ca] Not allocating networking since 'none' was specified. {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 608.150787] env[61957]: DEBUG oslo_concurrency.lockutils [None req-61bcf42d-e42b-40a8-98f8-af8cb211ebed tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] Releasing lock "refresh_cache-96c6b747-7293-4252-a1d0-b0d7684c5529" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 608.151373] env[61957]: DEBUG nova.compute.manager [None req-61bcf42d-e42b-40a8-98f8-af8cb211ebed tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] Start destroying the instance on the hypervisor. {{(pid=61957) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 608.151705] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-61bcf42d-e42b-40a8-98f8-af8cb211ebed tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] Destroying instance {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 608.152153] env[61957]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c9f4c138-3a35-4461-b9c4-7f683b9aa880 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.171310] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bbd4087-e179-42b8-b16b-62cc626a26a2 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.200737] env[61957]: WARNING nova.virt.vmwareapi.vmops [None req-61bcf42d-e42b-40a8-98f8-af8cb211ebed tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 96c6b747-7293-4252-a1d0-b0d7684c5529 could not be found. [ 608.200975] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-61bcf42d-e42b-40a8-98f8-af8cb211ebed tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] Instance destroyed {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 608.201174] env[61957]: INFO nova.compute.manager [None req-61bcf42d-e42b-40a8-98f8-af8cb211ebed tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] Took 0.05 seconds to destroy the instance on the hypervisor. [ 608.201425] env[61957]: DEBUG oslo.service.loopingcall [None req-61bcf42d-e42b-40a8-98f8-af8cb211ebed tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 608.201633] env[61957]: DEBUG nova.compute.manager [-] [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 608.201724] env[61957]: DEBUG nova.network.neutron [-] [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 608.266619] env[61957]: DEBUG nova.network.neutron [-] [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 608.376653] env[61957]: INFO nova.scheduler.client.report [None req-423a0985-b17f-45fa-b4ad-54e09005ad46 tempest-TenantUsagesTestJSON-2011428287 tempest-TenantUsagesTestJSON-2011428287-project-member] Deleted allocations for instance ed456ff1-249c-45d0-a007-3fef96ae8a2d [ 608.487139] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29cdeda2-c4c1-417a-a67d-a3d967f0e9a8 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.502711] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b3605cb-b59c-402e-adbf-1ad9e334d15b {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.543638] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a3020d6-c1d2-4043-aee9-06136626f663 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.550025] env[61957]: DEBUG nova.compute.manager [None req-7527ee8e-e60c-46dd-bb63-3e8733859a46 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] [instance: 4ad746d0-4076-4c97-8ea9-20e49fda59ca] Start building block device mappings for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 608.555013] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff839128-17e7-4b7b-bcb6-754e54018c40 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.570438] env[61957]: DEBUG nova.compute.provider_tree [None req-840b2d8a-f610-4c89-a367-da388d81d5aa tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 608.770025] env[61957]: DEBUG nova.network.neutron [-] [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 608.892978] env[61957]: DEBUG oslo_concurrency.lockutils [None req-423a0985-b17f-45fa-b4ad-54e09005ad46 tempest-TenantUsagesTestJSON-2011428287 tempest-TenantUsagesTestJSON-2011428287-project-member] Lock "ed456ff1-249c-45d0-a007-3fef96ae8a2d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 34.943s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 609.074040] env[61957]: DEBUG nova.scheduler.client.report [None req-840b2d8a-f610-4c89-a367-da388d81d5aa tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 609.279712] env[61957]: INFO nova.compute.manager [-] [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] Took 1.07 seconds to deallocate network for instance. [ 609.280650] env[61957]: DEBUG nova.compute.claims [None req-61bcf42d-e42b-40a8-98f8-af8cb211ebed tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] Aborting claim: {{(pid=61957) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 609.281126] env[61957]: DEBUG oslo_concurrency.lockutils [None req-61bcf42d-e42b-40a8-98f8-af8cb211ebed tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 609.397767] env[61957]: DEBUG nova.compute.manager [None req-4e7f7a28-6560-4b73-affe-356aa8d11b0b tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] Starting instance... {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 609.566941] env[61957]: DEBUG nova.compute.manager [None req-7527ee8e-e60c-46dd-bb63-3e8733859a46 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] [instance: 4ad746d0-4076-4c97-8ea9-20e49fda59ca] Start spawning the instance on the hypervisor. {{(pid=61957) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 609.581462] env[61957]: DEBUG oslo_concurrency.lockutils [None req-840b2d8a-f610-4c89-a367-da388d81d5aa tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.047s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 609.582281] env[61957]: ERROR nova.compute.manager [None req-840b2d8a-f610-4c89-a367-da388d81d5aa tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 25e905db-2ea8-4412-bfad-1406d998af37, please check neutron logs for more information. [ 609.582281] env[61957]: ERROR nova.compute.manager [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] Traceback (most recent call last): [ 609.582281] env[61957]: ERROR nova.compute.manager [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 609.582281] env[61957]: ERROR nova.compute.manager [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] self.driver.spawn(context, instance, image_meta, [ 609.582281] env[61957]: ERROR nova.compute.manager [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 609.582281] env[61957]: ERROR nova.compute.manager [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] self._vmops.spawn(context, instance, image_meta, injected_files, [ 609.582281] env[61957]: ERROR nova.compute.manager [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 609.582281] env[61957]: ERROR nova.compute.manager [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] vm_ref = self.build_virtual_machine(instance, [ 609.582281] env[61957]: ERROR nova.compute.manager [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 609.582281] env[61957]: ERROR nova.compute.manager [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] vif_infos = vmwarevif.get_vif_info(self._session, [ 609.582281] env[61957]: ERROR nova.compute.manager [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 609.582642] env[61957]: ERROR nova.compute.manager [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] for vif in network_info: [ 609.582642] env[61957]: ERROR nova.compute.manager [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 609.582642] env[61957]: ERROR nova.compute.manager [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] return self._sync_wrapper(fn, *args, **kwargs) [ 609.582642] env[61957]: ERROR nova.compute.manager [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 609.582642] env[61957]: ERROR nova.compute.manager [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] self.wait() [ 609.582642] env[61957]: ERROR nova.compute.manager [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 609.582642] env[61957]: ERROR nova.compute.manager [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] self[:] = self._gt.wait() [ 609.582642] env[61957]: ERROR nova.compute.manager [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 609.582642] env[61957]: ERROR nova.compute.manager [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] return self._exit_event.wait() [ 609.582642] env[61957]: ERROR nova.compute.manager [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 609.582642] env[61957]: ERROR nova.compute.manager [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] result = hub.switch() [ 609.582642] env[61957]: ERROR nova.compute.manager [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 609.582642] env[61957]: ERROR nova.compute.manager [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] return self.greenlet.switch() [ 609.583025] env[61957]: ERROR nova.compute.manager [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 609.583025] env[61957]: ERROR nova.compute.manager [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] result = function(*args, **kwargs) [ 609.583025] env[61957]: ERROR nova.compute.manager [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 609.583025] env[61957]: ERROR nova.compute.manager [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] return func(*args, **kwargs) [ 609.583025] env[61957]: ERROR nova.compute.manager [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 609.583025] env[61957]: ERROR nova.compute.manager [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] raise e [ 609.583025] env[61957]: ERROR nova.compute.manager [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 609.583025] env[61957]: ERROR nova.compute.manager [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] nwinfo = self.network_api.allocate_for_instance( [ 609.583025] env[61957]: ERROR nova.compute.manager [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 609.583025] env[61957]: ERROR nova.compute.manager [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] created_port_ids = self._update_ports_for_instance( [ 609.583025] env[61957]: ERROR nova.compute.manager [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 609.583025] env[61957]: ERROR nova.compute.manager [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] with excutils.save_and_reraise_exception(): [ 609.583025] env[61957]: ERROR nova.compute.manager [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 609.583433] env[61957]: ERROR nova.compute.manager [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] self.force_reraise() [ 609.583433] env[61957]: ERROR nova.compute.manager [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 609.583433] env[61957]: ERROR nova.compute.manager [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] raise self.value [ 609.583433] env[61957]: ERROR nova.compute.manager [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 609.583433] env[61957]: ERROR nova.compute.manager [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] updated_port = self._update_port( [ 609.583433] env[61957]: ERROR nova.compute.manager [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 609.583433] env[61957]: ERROR nova.compute.manager [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] _ensure_no_port_binding_failure(port) [ 609.583433] env[61957]: ERROR nova.compute.manager [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 609.583433] env[61957]: ERROR nova.compute.manager [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] raise exception.PortBindingFailed(port_id=port['id']) [ 609.583433] env[61957]: ERROR nova.compute.manager [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] nova.exception.PortBindingFailed: Binding failed for port 25e905db-2ea8-4412-bfad-1406d998af37, please check neutron logs for more information. [ 609.583433] env[61957]: ERROR nova.compute.manager [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] [ 609.583775] env[61957]: DEBUG nova.compute.utils [None req-840b2d8a-f610-4c89-a367-da388d81d5aa tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] Binding failed for port 25e905db-2ea8-4412-bfad-1406d998af37, please check neutron logs for more information. {{(pid=61957) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 609.586053] env[61957]: DEBUG nova.compute.manager [None req-840b2d8a-f610-4c89-a367-da388d81d5aa tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] Build of instance f9e04dbd-d5ae-4f03-9406-14b2caeaf010 was re-scheduled: Binding failed for port 25e905db-2ea8-4412-bfad-1406d998af37, please check neutron logs for more information. {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 609.586479] env[61957]: DEBUG nova.compute.manager [None req-840b2d8a-f610-4c89-a367-da388d81d5aa tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] Unplugging VIFs for instance {{(pid=61957) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 609.586699] env[61957]: DEBUG oslo_concurrency.lockutils [None req-840b2d8a-f610-4c89-a367-da388d81d5aa tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] Acquiring lock "refresh_cache-f9e04dbd-d5ae-4f03-9406-14b2caeaf010" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 609.586846] env[61957]: DEBUG oslo_concurrency.lockutils [None req-840b2d8a-f610-4c89-a367-da388d81d5aa tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] Acquired lock "refresh_cache-f9e04dbd-d5ae-4f03-9406-14b2caeaf010" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 609.587067] env[61957]: DEBUG nova.network.neutron [None req-840b2d8a-f610-4c89-a367-da388d81d5aa tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 609.596713] env[61957]: DEBUG oslo_concurrency.lockutils [None req-1b6e3bde-e95e-4682-9e04-c3f69f3be74f tempest-ServersWithSpecificFlavorTestJSON-568377525 tempest-ServersWithSpecificFlavorTestJSON-568377525-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.975s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 609.597439] env[61957]: INFO nova.compute.claims [None req-1b6e3bde-e95e-4682-9e04-c3f69f3be74f tempest-ServersWithSpecificFlavorTestJSON-568377525 tempest-ServersWithSpecificFlavorTestJSON-568377525-project-member] [instance: f660c725-fc06-42e4-8a34-5d80568b535e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 609.615363] env[61957]: DEBUG nova.virt.hardware [None req-7527ee8e-e60c-46dd-bb63-3e8733859a46 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T17:22:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T17:21:50Z,direct_url=,disk_format='vmdk',id=11c76a2c-f705-470a-ba9d-4657858bab38,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='81d3096d0f094373913fc3dc8f5c3c6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T17:21:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 609.615644] env[61957]: DEBUG nova.virt.hardware [None req-7527ee8e-e60c-46dd-bb63-3e8733859a46 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] Flavor limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 609.615868] env[61957]: DEBUG nova.virt.hardware [None req-7527ee8e-e60c-46dd-bb63-3e8733859a46 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] Image limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 609.616135] env[61957]: DEBUG nova.virt.hardware [None req-7527ee8e-e60c-46dd-bb63-3e8733859a46 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] Flavor pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 609.616991] env[61957]: DEBUG nova.virt.hardware [None req-7527ee8e-e60c-46dd-bb63-3e8733859a46 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] Image pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 609.616991] env[61957]: DEBUG nova.virt.hardware [None req-7527ee8e-e60c-46dd-bb63-3e8733859a46 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 609.616991] env[61957]: DEBUG nova.virt.hardware [None req-7527ee8e-e60c-46dd-bb63-3e8733859a46 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 609.616991] env[61957]: DEBUG nova.virt.hardware [None req-7527ee8e-e60c-46dd-bb63-3e8733859a46 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 609.616991] env[61957]: DEBUG nova.virt.hardware [None req-7527ee8e-e60c-46dd-bb63-3e8733859a46 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] Got 1 possible topologies {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 609.617248] env[61957]: DEBUG nova.virt.hardware [None req-7527ee8e-e60c-46dd-bb63-3e8733859a46 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 609.617294] env[61957]: DEBUG nova.virt.hardware [None req-7527ee8e-e60c-46dd-bb63-3e8733859a46 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 609.618798] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7162720-4108-4464-98af-2fdee72f77ba {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.630755] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d89ad338-e1dd-4c7a-9871-340aa0c846a8 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.650536] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-7527ee8e-e60c-46dd-bb63-3e8733859a46 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] [instance: 4ad746d0-4076-4c97-8ea9-20e49fda59ca] Instance VIF info [] {{(pid=61957) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 609.659122] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-7527ee8e-e60c-46dd-bb63-3e8733859a46 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] Creating folder: Project (0550e53c8edf42ad96e3be493c082b53). Parent ref: group-v274445. {{(pid=61957) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 609.659811] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-726f4cdc-4066-479e-9a69-ccef3e26de34 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.669988] env[61957]: INFO nova.virt.vmwareapi.vm_util [None req-7527ee8e-e60c-46dd-bb63-3e8733859a46 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] Created folder: Project (0550e53c8edf42ad96e3be493c082b53) in parent group-v274445. [ 609.670196] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-7527ee8e-e60c-46dd-bb63-3e8733859a46 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] Creating folder: Instances. Parent ref: group-v274453. {{(pid=61957) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 609.670425] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7aae5f2f-6e38-41d4-92fa-c126c41a54c1 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.680173] env[61957]: INFO nova.virt.vmwareapi.vm_util [None req-7527ee8e-e60c-46dd-bb63-3e8733859a46 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] Created folder: Instances in parent group-v274453. [ 609.680412] env[61957]: DEBUG oslo.service.loopingcall [None req-7527ee8e-e60c-46dd-bb63-3e8733859a46 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 609.680598] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4ad746d0-4076-4c97-8ea9-20e49fda59ca] Creating VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 609.680795] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-62ef3492-134f-4de5-844b-27b68fe889b4 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.699500] env[61957]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 609.699500] env[61957]: value = "task-1277311" [ 609.699500] env[61957]: _type = "Task" [ 609.699500] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 609.711459] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277311, 'name': CreateVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 609.762790] env[61957]: DEBUG nova.compute.manager [req-7f85499f-2f98-40da-8424-e6046ab26acb req-49638530-ea4f-4463-834c-9564d0e8aa41 service nova] [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] Received event network-changed-6a3b4c95-304d-4644-ba91-67780e2d82ae {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 609.762992] env[61957]: DEBUG nova.compute.manager [req-7f85499f-2f98-40da-8424-e6046ab26acb req-49638530-ea4f-4463-834c-9564d0e8aa41 service nova] [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] Refreshing instance network info cache due to event network-changed-6a3b4c95-304d-4644-ba91-67780e2d82ae. {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 609.763230] env[61957]: DEBUG oslo_concurrency.lockutils [req-7f85499f-2f98-40da-8424-e6046ab26acb req-49638530-ea4f-4463-834c-9564d0e8aa41 service nova] Acquiring lock "refresh_cache-96c6b747-7293-4252-a1d0-b0d7684c5529" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 609.763364] env[61957]: DEBUG oslo_concurrency.lockutils [req-7f85499f-2f98-40da-8424-e6046ab26acb req-49638530-ea4f-4463-834c-9564d0e8aa41 service nova] Acquired lock "refresh_cache-96c6b747-7293-4252-a1d0-b0d7684c5529" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 609.763523] env[61957]: DEBUG nova.network.neutron [req-7f85499f-2f98-40da-8424-e6046ab26acb req-49638530-ea4f-4463-834c-9564d0e8aa41 service nova] [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] Refreshing network info cache for port 6a3b4c95-304d-4644-ba91-67780e2d82ae {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 609.921483] env[61957]: DEBUG oslo_concurrency.lockutils [None req-4e7f7a28-6560-4b73-affe-356aa8d11b0b tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 610.139236] env[61957]: DEBUG nova.network.neutron [None req-840b2d8a-f610-4c89-a367-da388d81d5aa tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 610.211122] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277311, 'name': CreateVM_Task} progress is 99%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 610.308609] env[61957]: DEBUG oslo_concurrency.lockutils [None req-dbfaab8a-0784-4441-ad61-79e56d9c3c60 tempest-FloatingIPsAssociationTestJSON-1786153562 tempest-FloatingIPsAssociationTestJSON-1786153562-project-member] Acquiring lock "8257a521-2c25-45a2-a2e5-c735ece03da2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 610.308839] env[61957]: DEBUG oslo_concurrency.lockutils [None req-dbfaab8a-0784-4441-ad61-79e56d9c3c60 tempest-FloatingIPsAssociationTestJSON-1786153562 tempest-FloatingIPsAssociationTestJSON-1786153562-project-member] Lock "8257a521-2c25-45a2-a2e5-c735ece03da2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 610.316487] env[61957]: DEBUG nova.network.neutron [req-7f85499f-2f98-40da-8424-e6046ab26acb req-49638530-ea4f-4463-834c-9564d0e8aa41 service nova] [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 610.462442] env[61957]: DEBUG nova.network.neutron [None req-840b2d8a-f610-4c89-a367-da388d81d5aa tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 610.472141] env[61957]: DEBUG nova.network.neutron [req-7f85499f-2f98-40da-8424-e6046ab26acb req-49638530-ea4f-4463-834c-9564d0e8aa41 service nova] [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 610.714456] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277311, 'name': CreateVM_Task} progress is 99%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 610.965390] env[61957]: DEBUG oslo_concurrency.lockutils [None req-840b2d8a-f610-4c89-a367-da388d81d5aa tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] Releasing lock "refresh_cache-f9e04dbd-d5ae-4f03-9406-14b2caeaf010" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 610.965723] env[61957]: DEBUG nova.compute.manager [None req-840b2d8a-f610-4c89-a367-da388d81d5aa tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61957) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 610.965845] env[61957]: DEBUG nova.compute.manager [None req-840b2d8a-f610-4c89-a367-da388d81d5aa tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 610.966071] env[61957]: DEBUG nova.network.neutron [None req-840b2d8a-f610-4c89-a367-da388d81d5aa tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 610.976023] env[61957]: DEBUG oslo_concurrency.lockutils [req-7f85499f-2f98-40da-8424-e6046ab26acb req-49638530-ea4f-4463-834c-9564d0e8aa41 service nova] Releasing lock "refresh_cache-96c6b747-7293-4252-a1d0-b0d7684c5529" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 610.976023] env[61957]: DEBUG nova.compute.manager [req-7f85499f-2f98-40da-8424-e6046ab26acb req-49638530-ea4f-4463-834c-9564d0e8aa41 service nova] [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] Received event network-vif-deleted-6a3b4c95-304d-4644-ba91-67780e2d82ae {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 610.993965] env[61957]: DEBUG nova.network.neutron [None req-840b2d8a-f610-4c89-a367-da388d81d5aa tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 611.035728] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97183ff1-700c-4277-9e13-d809d11e2dc3 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.045299] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11903a31-789e-4370-9d9d-91c0ab72b533 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.084285] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2897cc67-3bf1-4d98-ba0d-f1599c2f5da7 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.092837] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e01d8cb0-d9c1-4c84-b73f-e11edca34003 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.107928] env[61957]: DEBUG nova.compute.provider_tree [None req-1b6e3bde-e95e-4682-9e04-c3f69f3be74f tempest-ServersWithSpecificFlavorTestJSON-568377525 tempest-ServersWithSpecificFlavorTestJSON-568377525-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 611.214290] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277311, 'name': CreateVM_Task, 'duration_secs': 1.383136} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 611.214429] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4ad746d0-4076-4c97-8ea9-20e49fda59ca] Created VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 611.215351] env[61957]: DEBUG oslo_vmware.service [None req-7527ee8e-e60c-46dd-bb63-3e8733859a46 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6be6f99-2b67-41c7-a689-bdd6a33c06ca {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.222618] env[61957]: DEBUG oslo_concurrency.lockutils [None req-7527ee8e-e60c-46dd-bb63-3e8733859a46 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 611.222671] env[61957]: DEBUG oslo_concurrency.lockutils [None req-7527ee8e-e60c-46dd-bb63-3e8733859a46 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] Acquired lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 611.223238] env[61957]: DEBUG oslo_concurrency.lockutils [None req-7527ee8e-e60c-46dd-bb63-3e8733859a46 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 611.223526] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-90171bdc-fe06-4ca3-ba81-fbf9db974be8 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.231221] env[61957]: DEBUG oslo_vmware.api [None req-7527ee8e-e60c-46dd-bb63-3e8733859a46 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] Waiting for the task: (returnval){ [ 611.231221] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]5250d9a7-63b3-36f1-535e-9d011b3b5d2b" [ 611.231221] env[61957]: _type = "Task" [ 611.231221] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 611.244793] env[61957]: DEBUG oslo_vmware.api [None req-7527ee8e-e60c-46dd-bb63-3e8733859a46 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]5250d9a7-63b3-36f1-535e-9d011b3b5d2b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 611.499799] env[61957]: DEBUG nova.network.neutron [None req-840b2d8a-f610-4c89-a367-da388d81d5aa tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 611.509257] env[61957]: DEBUG oslo_concurrency.lockutils [None req-c2f4fc75-b8bd-417b-8703-3b9048472a92 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] Acquiring lock "d0bde8ec-da55-42b5-8d2c-9df2b90b88a7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 611.509550] env[61957]: DEBUG oslo_concurrency.lockutils [None req-c2f4fc75-b8bd-417b-8703-3b9048472a92 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] Lock "d0bde8ec-da55-42b5-8d2c-9df2b90b88a7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 611.611969] env[61957]: DEBUG nova.scheduler.client.report [None req-1b6e3bde-e95e-4682-9e04-c3f69f3be74f tempest-ServersWithSpecificFlavorTestJSON-568377525 tempest-ServersWithSpecificFlavorTestJSON-568377525-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 611.744994] env[61957]: DEBUG oslo_concurrency.lockutils [None req-7527ee8e-e60c-46dd-bb63-3e8733859a46 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] Releasing lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 611.746050] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-7527ee8e-e60c-46dd-bb63-3e8733859a46 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] [instance: 4ad746d0-4076-4c97-8ea9-20e49fda59ca] Processing image 11c76a2c-f705-470a-ba9d-4657858bab38 {{(pid=61957) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 611.746050] env[61957]: DEBUG oslo_concurrency.lockutils [None req-7527ee8e-e60c-46dd-bb63-3e8733859a46 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 611.746050] env[61957]: DEBUG oslo_concurrency.lockutils [None req-7527ee8e-e60c-46dd-bb63-3e8733859a46 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] Acquired lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 611.746050] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-7527ee8e-e60c-46dd-bb63-3e8733859a46 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 611.746259] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0f665eee-a9a3-46d4-9059-e76286ac5568 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.754300] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-7527ee8e-e60c-46dd-bb63-3e8733859a46 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 611.754607] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-7527ee8e-e60c-46dd-bb63-3e8733859a46 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61957) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 611.755420] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea29c718-dac7-48ae-a68f-05b47d2a3ed8 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.769787] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7ce8bbeb-9229-4688-a0f1-5d1ebd9e63c1 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.777446] env[61957]: DEBUG oslo_vmware.api [None req-7527ee8e-e60c-46dd-bb63-3e8733859a46 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] Waiting for the task: (returnval){ [ 611.777446] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]5219e9c1-a179-3397-328c-f20d2e8390a0" [ 611.777446] env[61957]: _type = "Task" [ 611.777446] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 611.784246] env[61957]: DEBUG oslo_vmware.api [None req-7527ee8e-e60c-46dd-bb63-3e8733859a46 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]5219e9c1-a179-3397-328c-f20d2e8390a0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 612.003765] env[61957]: INFO nova.compute.manager [None req-840b2d8a-f610-4c89-a367-da388d81d5aa tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] [instance: f9e04dbd-d5ae-4f03-9406-14b2caeaf010] Took 1.04 seconds to deallocate network for instance. [ 612.118279] env[61957]: DEBUG oslo_concurrency.lockutils [None req-1b6e3bde-e95e-4682-9e04-c3f69f3be74f tempest-ServersWithSpecificFlavorTestJSON-568377525 tempest-ServersWithSpecificFlavorTestJSON-568377525-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.523s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 612.118824] env[61957]: DEBUG nova.compute.manager [None req-1b6e3bde-e95e-4682-9e04-c3f69f3be74f tempest-ServersWithSpecificFlavorTestJSON-568377525 tempest-ServersWithSpecificFlavorTestJSON-568377525-project-member] [instance: f660c725-fc06-42e4-8a34-5d80568b535e] Start building networks asynchronously for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 612.122153] env[61957]: DEBUG oslo_concurrency.lockutils [None req-42e67e16-6377-41f9-9501-426c4335c678 tempest-ServersTestBootFromVolume-482289819 tempest-ServersTestBootFromVolume-482289819-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.827s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 612.125086] env[61957]: INFO nova.compute.claims [None req-42e67e16-6377-41f9-9501-426c4335c678 tempest-ServersTestBootFromVolume-482289819 tempest-ServersTestBootFromVolume-482289819-project-member] [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 612.290849] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-7527ee8e-e60c-46dd-bb63-3e8733859a46 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] [instance: 4ad746d0-4076-4c97-8ea9-20e49fda59ca] Preparing fetch location {{(pid=61957) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 612.291150] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-7527ee8e-e60c-46dd-bb63-3e8733859a46 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] Creating directory with path [datastore1] vmware_temp/fccdd68b-c15a-4edf-860e-918b76bb9680/11c76a2c-f705-470a-ba9d-4657858bab38 {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 612.291393] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b58531cf-c9d2-4e52-8019-e8c78098c09b {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.321509] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-7527ee8e-e60c-46dd-bb63-3e8733859a46 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] Created directory with path [datastore1] vmware_temp/fccdd68b-c15a-4edf-860e-918b76bb9680/11c76a2c-f705-470a-ba9d-4657858bab38 {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 612.321719] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-7527ee8e-e60c-46dd-bb63-3e8733859a46 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] [instance: 4ad746d0-4076-4c97-8ea9-20e49fda59ca] Fetch image to [datastore1] vmware_temp/fccdd68b-c15a-4edf-860e-918b76bb9680/11c76a2c-f705-470a-ba9d-4657858bab38/tmp-sparse.vmdk {{(pid=61957) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 612.321869] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-7527ee8e-e60c-46dd-bb63-3e8733859a46 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] [instance: 4ad746d0-4076-4c97-8ea9-20e49fda59ca] Downloading image file data 11c76a2c-f705-470a-ba9d-4657858bab38 to [datastore1] vmware_temp/fccdd68b-c15a-4edf-860e-918b76bb9680/11c76a2c-f705-470a-ba9d-4657858bab38/tmp-sparse.vmdk on the data store datastore1 {{(pid=61957) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 612.322696] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc339e58-a8b5-4b8d-bd10-0f7ae092d9e4 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.331067] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d56e8e1c-a391-45d9-981d-1b5f962f0fdb {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.342280] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b6cb5c3-e606-4ab6-867c-061278ebb307 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.373537] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43f011ae-b5b3-4056-bda3-957744e95dd3 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.380170] env[61957]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-1537b9aa-c804-4e71-8968-6012e6965287 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.410241] env[61957]: DEBUG nova.virt.vmwareapi.images [None req-7527ee8e-e60c-46dd-bb63-3e8733859a46 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] [instance: 4ad746d0-4076-4c97-8ea9-20e49fda59ca] Downloading image file data 11c76a2c-f705-470a-ba9d-4657858bab38 to the data store datastore1 {{(pid=61957) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 612.474238] env[61957]: DEBUG oslo_vmware.rw_handles [None req-7527ee8e-e60c-46dd-bb63-3e8733859a46 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c2n1.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/fccdd68b-c15a-4edf-860e-918b76bb9680/11c76a2c-f705-470a-ba9d-4657858bab38/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=61957) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 612.636584] env[61957]: DEBUG nova.compute.utils [None req-1b6e3bde-e95e-4682-9e04-c3f69f3be74f tempest-ServersWithSpecificFlavorTestJSON-568377525 tempest-ServersWithSpecificFlavorTestJSON-568377525-project-member] Using /dev/sd instead of None {{(pid=61957) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 612.642607] env[61957]: DEBUG nova.compute.manager [None req-1b6e3bde-e95e-4682-9e04-c3f69f3be74f tempest-ServersWithSpecificFlavorTestJSON-568377525 tempest-ServersWithSpecificFlavorTestJSON-568377525-project-member] [instance: f660c725-fc06-42e4-8a34-5d80568b535e] Allocating IP information in the background. {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 612.642781] env[61957]: DEBUG nova.network.neutron [None req-1b6e3bde-e95e-4682-9e04-c3f69f3be74f tempest-ServersWithSpecificFlavorTestJSON-568377525 tempest-ServersWithSpecificFlavorTestJSON-568377525-project-member] [instance: f660c725-fc06-42e4-8a34-5d80568b535e] allocate_for_instance() {{(pid=61957) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 612.692798] env[61957]: DEBUG nova.compute.manager [None req-5e81c56f-9fb5-4d2d-9d97-efcb009e05a1 tempest-ServerDiagnosticsV248Test-248432757 tempest-ServerDiagnosticsV248Test-248432757-project-admin] [instance: c5af147e-3526-4014-98bc-7ad163ae89d9] Checking state {{(pid=61957) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 612.694294] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0d90f63-6442-44c4-8528-8df60baa0991 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.702605] env[61957]: INFO nova.compute.manager [None req-5e81c56f-9fb5-4d2d-9d97-efcb009e05a1 tempest-ServerDiagnosticsV248Test-248432757 tempest-ServerDiagnosticsV248Test-248432757-project-admin] [instance: c5af147e-3526-4014-98bc-7ad163ae89d9] Retrieving diagnostics [ 612.703800] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb8487c1-9e36-4087-8624-2c63b1f5b3fc {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.711735] env[61957]: DEBUG nova.policy [None req-1b6e3bde-e95e-4682-9e04-c3f69f3be74f tempest-ServersWithSpecificFlavorTestJSON-568377525 tempest-ServersWithSpecificFlavorTestJSON-568377525-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f0743fcb86844264bc749af409bb4c28', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '345b8654817443389eb6dbbca0d7f44a', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61957) authorize /opt/stack/nova/nova/policy.py:203}} [ 613.071734] env[61957]: INFO nova.scheduler.client.report [None req-840b2d8a-f610-4c89-a367-da388d81d5aa tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] Deleted allocations for instance f9e04dbd-d5ae-4f03-9406-14b2caeaf010 [ 613.151772] env[61957]: DEBUG nova.compute.manager [None req-1b6e3bde-e95e-4682-9e04-c3f69f3be74f tempest-ServersWithSpecificFlavorTestJSON-568377525 tempest-ServersWithSpecificFlavorTestJSON-568377525-project-member] [instance: f660c725-fc06-42e4-8a34-5d80568b535e] Start building block device mappings for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 613.244332] env[61957]: DEBUG oslo_vmware.rw_handles [None req-7527ee8e-e60c-46dd-bb63-3e8733859a46 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] Completed reading data from the image iterator. {{(pid=61957) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 613.244693] env[61957]: DEBUG oslo_vmware.rw_handles [None req-7527ee8e-e60c-46dd-bb63-3e8733859a46 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] Closing write handle for https://esx7c2n1.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/fccdd68b-c15a-4edf-860e-918b76bb9680/11c76a2c-f705-470a-ba9d-4657858bab38/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=61957) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 613.293476] env[61957]: DEBUG nova.virt.vmwareapi.images [None req-7527ee8e-e60c-46dd-bb63-3e8733859a46 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] [instance: 4ad746d0-4076-4c97-8ea9-20e49fda59ca] Downloaded image file data 11c76a2c-f705-470a-ba9d-4657858bab38 to vmware_temp/fccdd68b-c15a-4edf-860e-918b76bb9680/11c76a2c-f705-470a-ba9d-4657858bab38/tmp-sparse.vmdk on the data store datastore1 {{(pid=61957) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 613.295882] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-7527ee8e-e60c-46dd-bb63-3e8733859a46 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] [instance: 4ad746d0-4076-4c97-8ea9-20e49fda59ca] Caching image {{(pid=61957) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 613.296176] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-7527ee8e-e60c-46dd-bb63-3e8733859a46 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] Copying Virtual Disk [datastore1] vmware_temp/fccdd68b-c15a-4edf-860e-918b76bb9680/11c76a2c-f705-470a-ba9d-4657858bab38/tmp-sparse.vmdk to [datastore1] vmware_temp/fccdd68b-c15a-4edf-860e-918b76bb9680/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 613.297381] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-225e959a-5c0b-4d85-90bf-9b2bef3918de {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.306415] env[61957]: DEBUG oslo_vmware.api [None req-7527ee8e-e60c-46dd-bb63-3e8733859a46 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] Waiting for the task: (returnval){ [ 613.306415] env[61957]: value = "task-1277312" [ 613.306415] env[61957]: _type = "Task" [ 613.306415] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 613.314788] env[61957]: DEBUG oslo_vmware.api [None req-7527ee8e-e60c-46dd-bb63-3e8733859a46 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] Task: {'id': task-1277312, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 613.374639] env[61957]: DEBUG oslo_concurrency.lockutils [None req-691c932a-6913-4bb6-98f2-4142c7d131a8 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Acquiring lock "c5024e22-6499-4f2e-8d1b-99ca34b33104" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 613.374873] env[61957]: DEBUG oslo_concurrency.lockutils [None req-691c932a-6913-4bb6-98f2-4142c7d131a8 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Lock "c5024e22-6499-4f2e-8d1b-99ca34b33104" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 613.407330] env[61957]: DEBUG nova.network.neutron [None req-1b6e3bde-e95e-4682-9e04-c3f69f3be74f tempest-ServersWithSpecificFlavorTestJSON-568377525 tempest-ServersWithSpecificFlavorTestJSON-568377525-project-member] [instance: f660c725-fc06-42e4-8a34-5d80568b535e] Successfully created port: 511dfb63-c95d-4f13-9a94-1bbc532cbb38 {{(pid=61957) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 613.530619] env[61957]: DEBUG oslo_concurrency.lockutils [None req-456a04ab-263c-4c55-898a-70a189651712 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] Acquiring lock "6241fb44-ce1c-4881-bc32-7d19ddc83a61" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 613.531041] env[61957]: DEBUG oslo_concurrency.lockutils [None req-456a04ab-263c-4c55-898a-70a189651712 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] Lock "6241fb44-ce1c-4881-bc32-7d19ddc83a61" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 613.582683] env[61957]: DEBUG oslo_concurrency.lockutils [None req-840b2d8a-f610-4c89-a367-da388d81d5aa tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] Lock "f9e04dbd-d5ae-4f03-9406-14b2caeaf010" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 37.199s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 613.715927] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4dfb52f0-3af5-42e9-98e5-78efce957f6c {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.728956] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d9a2cfd-0cfd-4246-af80-8d27a388f524 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.767875] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2732b6a4-090f-48cd-aa5a-3fe825af80b8 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.776147] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-756c39df-21ad-461a-8828-3f32253a99fe {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.792323] env[61957]: DEBUG nova.compute.provider_tree [None req-42e67e16-6377-41f9-9501-426c4335c678 tempest-ServersTestBootFromVolume-482289819 tempest-ServersTestBootFromVolume-482289819-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 613.821863] env[61957]: DEBUG oslo_vmware.api [None req-7527ee8e-e60c-46dd-bb63-3e8733859a46 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] Task: {'id': task-1277312, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 614.089056] env[61957]: DEBUG nova.compute.manager [None req-9c3c6101-e50c-439a-91f6-f293e059d0d2 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] Starting instance... {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 614.169989] env[61957]: DEBUG nova.compute.manager [None req-1b6e3bde-e95e-4682-9e04-c3f69f3be74f tempest-ServersWithSpecificFlavorTestJSON-568377525 tempest-ServersWithSpecificFlavorTestJSON-568377525-project-member] [instance: f660c725-fc06-42e4-8a34-5d80568b535e] Start spawning the instance on the hypervisor. {{(pid=61957) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 614.200778] env[61957]: DEBUG nova.virt.hardware [None req-1b6e3bde-e95e-4682-9e04-c3f69f3be74f tempest-ServersWithSpecificFlavorTestJSON-568377525 tempest-ServersWithSpecificFlavorTestJSON-568377525-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T17:23:59Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='1459910260',id=31,is_public=True,memory_mb=192,name='tempest-flavor_with_ephemeral_0-728154207',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T17:21:50Z,direct_url=,disk_format='vmdk',id=11c76a2c-f705-470a-ba9d-4657858bab38,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='81d3096d0f094373913fc3dc8f5c3c6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T17:21:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 614.203039] env[61957]: DEBUG nova.virt.hardware [None req-1b6e3bde-e95e-4682-9e04-c3f69f3be74f tempest-ServersWithSpecificFlavorTestJSON-568377525 tempest-ServersWithSpecificFlavorTestJSON-568377525-project-member] Flavor limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 614.203039] env[61957]: DEBUG nova.virt.hardware [None req-1b6e3bde-e95e-4682-9e04-c3f69f3be74f tempest-ServersWithSpecificFlavorTestJSON-568377525 tempest-ServersWithSpecificFlavorTestJSON-568377525-project-member] Image limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 614.203039] env[61957]: DEBUG nova.virt.hardware [None req-1b6e3bde-e95e-4682-9e04-c3f69f3be74f tempest-ServersWithSpecificFlavorTestJSON-568377525 tempest-ServersWithSpecificFlavorTestJSON-568377525-project-member] Flavor pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 614.203039] env[61957]: DEBUG nova.virt.hardware [None req-1b6e3bde-e95e-4682-9e04-c3f69f3be74f tempest-ServersWithSpecificFlavorTestJSON-568377525 tempest-ServersWithSpecificFlavorTestJSON-568377525-project-member] Image pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 614.203039] env[61957]: DEBUG nova.virt.hardware [None req-1b6e3bde-e95e-4682-9e04-c3f69f3be74f tempest-ServersWithSpecificFlavorTestJSON-568377525 tempest-ServersWithSpecificFlavorTestJSON-568377525-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 614.203595] env[61957]: DEBUG nova.virt.hardware [None req-1b6e3bde-e95e-4682-9e04-c3f69f3be74f tempest-ServersWithSpecificFlavorTestJSON-568377525 tempest-ServersWithSpecificFlavorTestJSON-568377525-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 614.203595] env[61957]: DEBUG nova.virt.hardware [None req-1b6e3bde-e95e-4682-9e04-c3f69f3be74f tempest-ServersWithSpecificFlavorTestJSON-568377525 tempest-ServersWithSpecificFlavorTestJSON-568377525-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 614.203595] env[61957]: DEBUG nova.virt.hardware [None req-1b6e3bde-e95e-4682-9e04-c3f69f3be74f tempest-ServersWithSpecificFlavorTestJSON-568377525 tempest-ServersWithSpecificFlavorTestJSON-568377525-project-member] Got 1 possible topologies {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 614.203595] env[61957]: DEBUG nova.virt.hardware [None req-1b6e3bde-e95e-4682-9e04-c3f69f3be74f tempest-ServersWithSpecificFlavorTestJSON-568377525 tempest-ServersWithSpecificFlavorTestJSON-568377525-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 614.203595] env[61957]: DEBUG nova.virt.hardware [None req-1b6e3bde-e95e-4682-9e04-c3f69f3be74f tempest-ServersWithSpecificFlavorTestJSON-568377525 tempest-ServersWithSpecificFlavorTestJSON-568377525-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 614.203952] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b25a676d-50c4-40dd-aa33-9eb36aee6fe4 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.216063] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a95ed1c9-d13d-4717-a7f3-82097aa3865a {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.294853] env[61957]: DEBUG nova.scheduler.client.report [None req-42e67e16-6377-41f9-9501-426c4335c678 tempest-ServersTestBootFromVolume-482289819 tempest-ServersTestBootFromVolume-482289819-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 614.318906] env[61957]: DEBUG oslo_vmware.api [None req-7527ee8e-e60c-46dd-bb63-3e8733859a46 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] Task: {'id': task-1277312, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.660966} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 614.320998] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-7527ee8e-e60c-46dd-bb63-3e8733859a46 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] Copied Virtual Disk [datastore1] vmware_temp/fccdd68b-c15a-4edf-860e-918b76bb9680/11c76a2c-f705-470a-ba9d-4657858bab38/tmp-sparse.vmdk to [datastore1] vmware_temp/fccdd68b-c15a-4edf-860e-918b76bb9680/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 614.321275] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-7527ee8e-e60c-46dd-bb63-3e8733859a46 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] Deleting the datastore file [datastore1] vmware_temp/fccdd68b-c15a-4edf-860e-918b76bb9680/11c76a2c-f705-470a-ba9d-4657858bab38/tmp-sparse.vmdk {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 614.321957] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b147ded5-ff7c-4ee1-a246-29f9acaba137 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.333680] env[61957]: DEBUG oslo_vmware.api [None req-7527ee8e-e60c-46dd-bb63-3e8733859a46 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] Waiting for the task: (returnval){ [ 614.333680] env[61957]: value = "task-1277313" [ 614.333680] env[61957]: _type = "Task" [ 614.333680] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 614.341966] env[61957]: DEBUG oslo_vmware.api [None req-7527ee8e-e60c-46dd-bb63-3e8733859a46 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] Task: {'id': task-1277313, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 614.627643] env[61957]: DEBUG oslo_concurrency.lockutils [None req-9c3c6101-e50c-439a-91f6-f293e059d0d2 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 614.803612] env[61957]: DEBUG oslo_concurrency.lockutils [None req-42e67e16-6377-41f9-9501-426c4335c678 tempest-ServersTestBootFromVolume-482289819 tempest-ServersTestBootFromVolume-482289819-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.681s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 614.804093] env[61957]: DEBUG nova.compute.manager [None req-42e67e16-6377-41f9-9501-426c4335c678 tempest-ServersTestBootFromVolume-482289819 tempest-ServersTestBootFromVolume-482289819-project-member] [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] Start building networks asynchronously for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 614.812679] env[61957]: DEBUG oslo_concurrency.lockutils [None req-260de5ae-9540-404d-a23e-9facd81096ee tempest-AttachInterfacesUnderV243Test-1229304814 tempest-AttachInterfacesUnderV243Test-1229304814-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.425s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 614.814714] env[61957]: INFO nova.compute.claims [None req-260de5ae-9540-404d-a23e-9facd81096ee tempest-AttachInterfacesUnderV243Test-1229304814 tempest-AttachInterfacesUnderV243Test-1229304814-project-member] [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 614.844863] env[61957]: DEBUG oslo_vmware.api [None req-7527ee8e-e60c-46dd-bb63-3e8733859a46 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] Task: {'id': task-1277313, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.025274} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 614.847171] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-7527ee8e-e60c-46dd-bb63-3e8733859a46 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] Deleted the datastore file {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 614.847504] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-7527ee8e-e60c-46dd-bb63-3e8733859a46 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] Moving file from [datastore1] vmware_temp/fccdd68b-c15a-4edf-860e-918b76bb9680/11c76a2c-f705-470a-ba9d-4657858bab38 to [datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38. {{(pid=61957) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 614.847845] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-221e808e-058c-4a1a-b393-cb77a4e1889f {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.855193] env[61957]: DEBUG oslo_vmware.api [None req-7527ee8e-e60c-46dd-bb63-3e8733859a46 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] Waiting for the task: (returnval){ [ 614.855193] env[61957]: value = "task-1277314" [ 614.855193] env[61957]: _type = "Task" [ 614.855193] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 614.865780] env[61957]: DEBUG oslo_vmware.api [None req-7527ee8e-e60c-46dd-bb63-3e8733859a46 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] Task: {'id': task-1277314, 'name': MoveDatastoreFile_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 615.290590] env[61957]: DEBUG oslo_concurrency.lockutils [None req-3e8259a8-69fb-4692-917a-5e7f03f3483c tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] Acquiring lock "c5af147e-3526-4014-98bc-7ad163ae89d9" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 615.293661] env[61957]: DEBUG oslo_concurrency.lockutils [None req-3e8259a8-69fb-4692-917a-5e7f03f3483c tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] Lock "c5af147e-3526-4014-98bc-7ad163ae89d9" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.003s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 615.296357] env[61957]: DEBUG oslo_concurrency.lockutils [None req-3e8259a8-69fb-4692-917a-5e7f03f3483c tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] Acquiring lock "c5af147e-3526-4014-98bc-7ad163ae89d9-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 615.296357] env[61957]: DEBUG oslo_concurrency.lockutils [None req-3e8259a8-69fb-4692-917a-5e7f03f3483c tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] Lock "c5af147e-3526-4014-98bc-7ad163ae89d9-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 615.296357] env[61957]: DEBUG oslo_concurrency.lockutils [None req-3e8259a8-69fb-4692-917a-5e7f03f3483c tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] Lock "c5af147e-3526-4014-98bc-7ad163ae89d9-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 615.297428] env[61957]: INFO nova.compute.manager [None req-3e8259a8-69fb-4692-917a-5e7f03f3483c tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] [instance: c5af147e-3526-4014-98bc-7ad163ae89d9] Terminating instance [ 615.300678] env[61957]: DEBUG oslo_concurrency.lockutils [None req-3e8259a8-69fb-4692-917a-5e7f03f3483c tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] Acquiring lock "refresh_cache-c5af147e-3526-4014-98bc-7ad163ae89d9" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 615.300979] env[61957]: DEBUG oslo_concurrency.lockutils [None req-3e8259a8-69fb-4692-917a-5e7f03f3483c tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] Acquired lock "refresh_cache-c5af147e-3526-4014-98bc-7ad163ae89d9" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 615.301572] env[61957]: DEBUG nova.network.neutron [None req-3e8259a8-69fb-4692-917a-5e7f03f3483c tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] [instance: c5af147e-3526-4014-98bc-7ad163ae89d9] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 615.322982] env[61957]: ERROR nova.compute.manager [None req-1b6e3bde-e95e-4682-9e04-c3f69f3be74f tempest-ServersWithSpecificFlavorTestJSON-568377525 tempest-ServersWithSpecificFlavorTestJSON-568377525-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 511dfb63-c95d-4f13-9a94-1bbc532cbb38, please check neutron logs for more information. [ 615.322982] env[61957]: ERROR nova.compute.manager Traceback (most recent call last): [ 615.322982] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 615.322982] env[61957]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 615.322982] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 615.322982] env[61957]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 615.322982] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 615.322982] env[61957]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 615.322982] env[61957]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 615.322982] env[61957]: ERROR nova.compute.manager self.force_reraise() [ 615.322982] env[61957]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 615.322982] env[61957]: ERROR nova.compute.manager raise self.value [ 615.322982] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 615.322982] env[61957]: ERROR nova.compute.manager updated_port = self._update_port( [ 615.322982] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 615.322982] env[61957]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 615.323605] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 615.323605] env[61957]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 615.323605] env[61957]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 511dfb63-c95d-4f13-9a94-1bbc532cbb38, please check neutron logs for more information. [ 615.323605] env[61957]: ERROR nova.compute.manager [ 615.323605] env[61957]: Traceback (most recent call last): [ 615.323605] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 615.323605] env[61957]: listener.cb(fileno) [ 615.323605] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 615.323605] env[61957]: result = function(*args, **kwargs) [ 615.323605] env[61957]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 615.323605] env[61957]: return func(*args, **kwargs) [ 615.323605] env[61957]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 615.323605] env[61957]: raise e [ 615.323605] env[61957]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 615.323605] env[61957]: nwinfo = self.network_api.allocate_for_instance( [ 615.323605] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 615.323605] env[61957]: created_port_ids = self._update_ports_for_instance( [ 615.323605] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 615.323605] env[61957]: with excutils.save_and_reraise_exception(): [ 615.323605] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 615.323605] env[61957]: self.force_reraise() [ 615.323605] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 615.323605] env[61957]: raise self.value [ 615.323605] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 615.323605] env[61957]: updated_port = self._update_port( [ 615.323605] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 615.323605] env[61957]: _ensure_no_port_binding_failure(port) [ 615.323605] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 615.323605] env[61957]: raise exception.PortBindingFailed(port_id=port['id']) [ 615.324501] env[61957]: nova.exception.PortBindingFailed: Binding failed for port 511dfb63-c95d-4f13-9a94-1bbc532cbb38, please check neutron logs for more information. [ 615.324501] env[61957]: Removing descriptor: 17 [ 615.324716] env[61957]: DEBUG nova.compute.utils [None req-42e67e16-6377-41f9-9501-426c4335c678 tempest-ServersTestBootFromVolume-482289819 tempest-ServersTestBootFromVolume-482289819-project-member] Using /dev/sd instead of None {{(pid=61957) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 615.330667] env[61957]: ERROR nova.compute.manager [None req-1b6e3bde-e95e-4682-9e04-c3f69f3be74f tempest-ServersWithSpecificFlavorTestJSON-568377525 tempest-ServersWithSpecificFlavorTestJSON-568377525-project-member] [instance: f660c725-fc06-42e4-8a34-5d80568b535e] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 511dfb63-c95d-4f13-9a94-1bbc532cbb38, please check neutron logs for more information. [ 615.330667] env[61957]: ERROR nova.compute.manager [instance: f660c725-fc06-42e4-8a34-5d80568b535e] Traceback (most recent call last): [ 615.330667] env[61957]: ERROR nova.compute.manager [instance: f660c725-fc06-42e4-8a34-5d80568b535e] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 615.330667] env[61957]: ERROR nova.compute.manager [instance: f660c725-fc06-42e4-8a34-5d80568b535e] yield resources [ 615.330667] env[61957]: ERROR nova.compute.manager [instance: f660c725-fc06-42e4-8a34-5d80568b535e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 615.330667] env[61957]: ERROR nova.compute.manager [instance: f660c725-fc06-42e4-8a34-5d80568b535e] self.driver.spawn(context, instance, image_meta, [ 615.330667] env[61957]: ERROR nova.compute.manager [instance: f660c725-fc06-42e4-8a34-5d80568b535e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 615.330667] env[61957]: ERROR nova.compute.manager [instance: f660c725-fc06-42e4-8a34-5d80568b535e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 615.330667] env[61957]: ERROR nova.compute.manager [instance: f660c725-fc06-42e4-8a34-5d80568b535e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 615.330667] env[61957]: ERROR nova.compute.manager [instance: f660c725-fc06-42e4-8a34-5d80568b535e] vm_ref = self.build_virtual_machine(instance, [ 615.330667] env[61957]: ERROR nova.compute.manager [instance: f660c725-fc06-42e4-8a34-5d80568b535e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 615.331116] env[61957]: ERROR nova.compute.manager [instance: f660c725-fc06-42e4-8a34-5d80568b535e] vif_infos = vmwarevif.get_vif_info(self._session, [ 615.331116] env[61957]: ERROR nova.compute.manager [instance: f660c725-fc06-42e4-8a34-5d80568b535e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 615.331116] env[61957]: ERROR nova.compute.manager [instance: f660c725-fc06-42e4-8a34-5d80568b535e] for vif in network_info: [ 615.331116] env[61957]: ERROR nova.compute.manager [instance: f660c725-fc06-42e4-8a34-5d80568b535e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 615.331116] env[61957]: ERROR nova.compute.manager [instance: f660c725-fc06-42e4-8a34-5d80568b535e] return self._sync_wrapper(fn, *args, **kwargs) [ 615.331116] env[61957]: ERROR nova.compute.manager [instance: f660c725-fc06-42e4-8a34-5d80568b535e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 615.331116] env[61957]: ERROR nova.compute.manager [instance: f660c725-fc06-42e4-8a34-5d80568b535e] self.wait() [ 615.331116] env[61957]: ERROR nova.compute.manager [instance: f660c725-fc06-42e4-8a34-5d80568b535e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 615.331116] env[61957]: ERROR nova.compute.manager [instance: f660c725-fc06-42e4-8a34-5d80568b535e] self[:] = self._gt.wait() [ 615.331116] env[61957]: ERROR nova.compute.manager [instance: f660c725-fc06-42e4-8a34-5d80568b535e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 615.331116] env[61957]: ERROR nova.compute.manager [instance: f660c725-fc06-42e4-8a34-5d80568b535e] return self._exit_event.wait() [ 615.331116] env[61957]: ERROR nova.compute.manager [instance: f660c725-fc06-42e4-8a34-5d80568b535e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 615.331116] env[61957]: ERROR nova.compute.manager [instance: f660c725-fc06-42e4-8a34-5d80568b535e] result = hub.switch() [ 615.331507] env[61957]: ERROR nova.compute.manager [instance: f660c725-fc06-42e4-8a34-5d80568b535e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 615.331507] env[61957]: ERROR nova.compute.manager [instance: f660c725-fc06-42e4-8a34-5d80568b535e] return self.greenlet.switch() [ 615.331507] env[61957]: ERROR nova.compute.manager [instance: f660c725-fc06-42e4-8a34-5d80568b535e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 615.331507] env[61957]: ERROR nova.compute.manager [instance: f660c725-fc06-42e4-8a34-5d80568b535e] result = function(*args, **kwargs) [ 615.331507] env[61957]: ERROR nova.compute.manager [instance: f660c725-fc06-42e4-8a34-5d80568b535e] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 615.331507] env[61957]: ERROR nova.compute.manager [instance: f660c725-fc06-42e4-8a34-5d80568b535e] return func(*args, **kwargs) [ 615.331507] env[61957]: ERROR nova.compute.manager [instance: f660c725-fc06-42e4-8a34-5d80568b535e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 615.331507] env[61957]: ERROR nova.compute.manager [instance: f660c725-fc06-42e4-8a34-5d80568b535e] raise e [ 615.331507] env[61957]: ERROR nova.compute.manager [instance: f660c725-fc06-42e4-8a34-5d80568b535e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 615.331507] env[61957]: ERROR nova.compute.manager [instance: f660c725-fc06-42e4-8a34-5d80568b535e] nwinfo = self.network_api.allocate_for_instance( [ 615.331507] env[61957]: ERROR nova.compute.manager [instance: f660c725-fc06-42e4-8a34-5d80568b535e] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 615.331507] env[61957]: ERROR nova.compute.manager [instance: f660c725-fc06-42e4-8a34-5d80568b535e] created_port_ids = self._update_ports_for_instance( [ 615.331507] env[61957]: ERROR nova.compute.manager [instance: f660c725-fc06-42e4-8a34-5d80568b535e] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 615.331927] env[61957]: ERROR nova.compute.manager [instance: f660c725-fc06-42e4-8a34-5d80568b535e] with excutils.save_and_reraise_exception(): [ 615.331927] env[61957]: ERROR nova.compute.manager [instance: f660c725-fc06-42e4-8a34-5d80568b535e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 615.331927] env[61957]: ERROR nova.compute.manager [instance: f660c725-fc06-42e4-8a34-5d80568b535e] self.force_reraise() [ 615.331927] env[61957]: ERROR nova.compute.manager [instance: f660c725-fc06-42e4-8a34-5d80568b535e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 615.331927] env[61957]: ERROR nova.compute.manager [instance: f660c725-fc06-42e4-8a34-5d80568b535e] raise self.value [ 615.331927] env[61957]: ERROR nova.compute.manager [instance: f660c725-fc06-42e4-8a34-5d80568b535e] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 615.331927] env[61957]: ERROR nova.compute.manager [instance: f660c725-fc06-42e4-8a34-5d80568b535e] updated_port = self._update_port( [ 615.331927] env[61957]: ERROR nova.compute.manager [instance: f660c725-fc06-42e4-8a34-5d80568b535e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 615.331927] env[61957]: ERROR nova.compute.manager [instance: f660c725-fc06-42e4-8a34-5d80568b535e] _ensure_no_port_binding_failure(port) [ 615.331927] env[61957]: ERROR nova.compute.manager [instance: f660c725-fc06-42e4-8a34-5d80568b535e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 615.331927] env[61957]: ERROR nova.compute.manager [instance: f660c725-fc06-42e4-8a34-5d80568b535e] raise exception.PortBindingFailed(port_id=port['id']) [ 615.331927] env[61957]: ERROR nova.compute.manager [instance: f660c725-fc06-42e4-8a34-5d80568b535e] nova.exception.PortBindingFailed: Binding failed for port 511dfb63-c95d-4f13-9a94-1bbc532cbb38, please check neutron logs for more information. [ 615.331927] env[61957]: ERROR nova.compute.manager [instance: f660c725-fc06-42e4-8a34-5d80568b535e] [ 615.332353] env[61957]: INFO nova.compute.manager [None req-1b6e3bde-e95e-4682-9e04-c3f69f3be74f tempest-ServersWithSpecificFlavorTestJSON-568377525 tempest-ServersWithSpecificFlavorTestJSON-568377525-project-member] [instance: f660c725-fc06-42e4-8a34-5d80568b535e] Terminating instance [ 615.333511] env[61957]: DEBUG nova.compute.manager [None req-42e67e16-6377-41f9-9501-426c4335c678 tempest-ServersTestBootFromVolume-482289819 tempest-ServersTestBootFromVolume-482289819-project-member] [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] Allocating IP information in the background. {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 615.333717] env[61957]: DEBUG nova.network.neutron [None req-42e67e16-6377-41f9-9501-426c4335c678 tempest-ServersTestBootFromVolume-482289819 tempest-ServersTestBootFromVolume-482289819-project-member] [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] allocate_for_instance() {{(pid=61957) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 615.336507] env[61957]: DEBUG oslo_concurrency.lockutils [None req-1b6e3bde-e95e-4682-9e04-c3f69f3be74f tempest-ServersWithSpecificFlavorTestJSON-568377525 tempest-ServersWithSpecificFlavorTestJSON-568377525-project-member] Acquiring lock "refresh_cache-f660c725-fc06-42e4-8a34-5d80568b535e" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 615.337190] env[61957]: DEBUG oslo_concurrency.lockutils [None req-1b6e3bde-e95e-4682-9e04-c3f69f3be74f tempest-ServersWithSpecificFlavorTestJSON-568377525 tempest-ServersWithSpecificFlavorTestJSON-568377525-project-member] Acquired lock "refresh_cache-f660c725-fc06-42e4-8a34-5d80568b535e" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 615.337469] env[61957]: DEBUG nova.network.neutron [None req-1b6e3bde-e95e-4682-9e04-c3f69f3be74f tempest-ServersWithSpecificFlavorTestJSON-568377525 tempest-ServersWithSpecificFlavorTestJSON-568377525-project-member] [instance: f660c725-fc06-42e4-8a34-5d80568b535e] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 615.367454] env[61957]: DEBUG oslo_vmware.api [None req-7527ee8e-e60c-46dd-bb63-3e8733859a46 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] Task: {'id': task-1277314, 'name': MoveDatastoreFile_Task, 'duration_secs': 0.025304} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 615.368521] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-7527ee8e-e60c-46dd-bb63-3e8733859a46 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] File moved {{(pid=61957) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 615.368884] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-7527ee8e-e60c-46dd-bb63-3e8733859a46 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] [instance: 4ad746d0-4076-4c97-8ea9-20e49fda59ca] Cleaning up location [datastore1] vmware_temp/fccdd68b-c15a-4edf-860e-918b76bb9680 {{(pid=61957) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 615.369143] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-7527ee8e-e60c-46dd-bb63-3e8733859a46 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] Deleting the datastore file [datastore1] vmware_temp/fccdd68b-c15a-4edf-860e-918b76bb9680 {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 615.369403] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7546ea3c-2a16-49dc-b971-45d52f6533b2 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.375902] env[61957]: DEBUG oslo_vmware.api [None req-7527ee8e-e60c-46dd-bb63-3e8733859a46 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] Waiting for the task: (returnval){ [ 615.375902] env[61957]: value = "task-1277315" [ 615.375902] env[61957]: _type = "Task" [ 615.375902] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 615.387969] env[61957]: DEBUG oslo_vmware.api [None req-7527ee8e-e60c-46dd-bb63-3e8733859a46 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] Task: {'id': task-1277315, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 615.393046] env[61957]: DEBUG nova.policy [None req-42e67e16-6377-41f9-9501-426c4335c678 tempest-ServersTestBootFromVolume-482289819 tempest-ServersTestBootFromVolume-482289819-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '013dc5591507456895e53fee519f1403', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '61e9db9bf2f743e4afc4d7af70c9607c', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61957) authorize /opt/stack/nova/nova/policy.py:203}} [ 615.705969] env[61957]: DEBUG nova.network.neutron [None req-42e67e16-6377-41f9-9501-426c4335c678 tempest-ServersTestBootFromVolume-482289819 tempest-ServersTestBootFromVolume-482289819-project-member] [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] Successfully created port: bb5c53cf-6a2a-4efe-9be1-07da3dd400d6 {{(pid=61957) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 615.783936] env[61957]: DEBUG nova.compute.manager [req-dc58e1b1-ce88-4ef6-8580-5ae1b680a1a6 req-e5750388-ea73-44e2-900d-00b98640485a service nova] [instance: f660c725-fc06-42e4-8a34-5d80568b535e] Received event network-changed-511dfb63-c95d-4f13-9a94-1bbc532cbb38 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 615.783936] env[61957]: DEBUG nova.compute.manager [req-dc58e1b1-ce88-4ef6-8580-5ae1b680a1a6 req-e5750388-ea73-44e2-900d-00b98640485a service nova] [instance: f660c725-fc06-42e4-8a34-5d80568b535e] Refreshing instance network info cache due to event network-changed-511dfb63-c95d-4f13-9a94-1bbc532cbb38. {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 615.783936] env[61957]: DEBUG oslo_concurrency.lockutils [req-dc58e1b1-ce88-4ef6-8580-5ae1b680a1a6 req-e5750388-ea73-44e2-900d-00b98640485a service nova] Acquiring lock "refresh_cache-f660c725-fc06-42e4-8a34-5d80568b535e" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 615.829388] env[61957]: DEBUG nova.network.neutron [None req-3e8259a8-69fb-4692-917a-5e7f03f3483c tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] [instance: c5af147e-3526-4014-98bc-7ad163ae89d9] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 615.834694] env[61957]: DEBUG nova.compute.manager [None req-42e67e16-6377-41f9-9501-426c4335c678 tempest-ServersTestBootFromVolume-482289819 tempest-ServersTestBootFromVolume-482289819-project-member] [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] Start building block device mappings for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 615.877820] env[61957]: DEBUG nova.network.neutron [None req-1b6e3bde-e95e-4682-9e04-c3f69f3be74f tempest-ServersWithSpecificFlavorTestJSON-568377525 tempest-ServersWithSpecificFlavorTestJSON-568377525-project-member] [instance: f660c725-fc06-42e4-8a34-5d80568b535e] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 615.888226] env[61957]: DEBUG oslo_vmware.api [None req-7527ee8e-e60c-46dd-bb63-3e8733859a46 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] Task: {'id': task-1277315, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.044492} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 615.891294] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-7527ee8e-e60c-46dd-bb63-3e8733859a46 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] Deleted the datastore file {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 615.892204] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-071e590f-aa34-47b2-ba57-4fb08391ddaa {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.898191] env[61957]: DEBUG oslo_vmware.api [None req-7527ee8e-e60c-46dd-bb63-3e8733859a46 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] Waiting for the task: (returnval){ [ 615.898191] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52aaac2c-a8f0-7810-2326-46c37e335203" [ 615.898191] env[61957]: _type = "Task" [ 615.898191] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 615.910174] env[61957]: DEBUG oslo_vmware.api [None req-7527ee8e-e60c-46dd-bb63-3e8733859a46 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52aaac2c-a8f0-7810-2326-46c37e335203, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 615.943199] env[61957]: DEBUG nova.network.neutron [None req-3e8259a8-69fb-4692-917a-5e7f03f3483c tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] [instance: c5af147e-3526-4014-98bc-7ad163ae89d9] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 616.166121] env[61957]: DEBUG nova.network.neutron [None req-1b6e3bde-e95e-4682-9e04-c3f69f3be74f tempest-ServersWithSpecificFlavorTestJSON-568377525 tempest-ServersWithSpecificFlavorTestJSON-568377525-project-member] [instance: f660c725-fc06-42e4-8a34-5d80568b535e] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 616.238781] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ecdaa73-1648-4a0c-991c-5a59e18277c2 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.247130] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-453b04a9-ee5f-4aa5-9745-838788267ad0 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.286587] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0978736-ebd5-4735-aff2-deefd1952e30 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.294969] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a61e6054-9a2b-48c1-a948-f20ab50f44f9 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.311455] env[61957]: DEBUG nova.compute.provider_tree [None req-260de5ae-9540-404d-a23e-9facd81096ee tempest-AttachInterfacesUnderV243Test-1229304814 tempest-AttachInterfacesUnderV243Test-1229304814-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 616.344314] env[61957]: INFO nova.virt.block_device [None req-42e67e16-6377-41f9-9501-426c4335c678 tempest-ServersTestBootFromVolume-482289819 tempest-ServersTestBootFromVolume-482289819-project-member] [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] Booting with volume bb0475df-5d35-4e9e-b7ed-adf5fb3f1018 at /dev/sda [ 616.409546] env[61957]: DEBUG oslo_vmware.api [None req-7527ee8e-e60c-46dd-bb63-3e8733859a46 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52aaac2c-a8f0-7810-2326-46c37e335203, 'name': SearchDatastore_Task, 'duration_secs': 0.007953} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 616.411130] env[61957]: DEBUG oslo_concurrency.lockutils [None req-7527ee8e-e60c-46dd-bb63-3e8733859a46 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] Releasing lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 616.411504] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-7527ee8e-e60c-46dd-bb63-3e8733859a46 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore1] 4ad746d0-4076-4c97-8ea9-20e49fda59ca/4ad746d0-4076-4c97-8ea9-20e49fda59ca.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 616.411959] env[61957]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-219a4c8f-9bf4-468a-8e35-199a47d69e9b {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.416104] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2b1f0735-0be4-41b5-a53a-e37c6e2b1125 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.424865] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b26d64c0-73d4-4fb1-bb59-5f4fa7b38a08 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.443036] env[61957]: DEBUG oslo_vmware.api [None req-7527ee8e-e60c-46dd-bb63-3e8733859a46 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] Waiting for the task: (returnval){ [ 616.443036] env[61957]: value = "task-1277316" [ 616.443036] env[61957]: _type = "Task" [ 616.443036] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 616.446650] env[61957]: DEBUG oslo_concurrency.lockutils [None req-3e8259a8-69fb-4692-917a-5e7f03f3483c tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] Releasing lock "refresh_cache-c5af147e-3526-4014-98bc-7ad163ae89d9" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 616.447690] env[61957]: DEBUG nova.compute.manager [None req-3e8259a8-69fb-4692-917a-5e7f03f3483c tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] [instance: c5af147e-3526-4014-98bc-7ad163ae89d9] Start destroying the instance on the hypervisor. {{(pid=61957) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 616.447690] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-3e8259a8-69fb-4692-917a-5e7f03f3483c tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] [instance: c5af147e-3526-4014-98bc-7ad163ae89d9] Destroying instance {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 616.448044] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d4e69b1-bf27-4288-93d5-432bc2956a5e {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.460356] env[61957]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e435339e-de0c-4c47-b3bf-fb3499d429ba {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.462372] env[61957]: DEBUG oslo_vmware.api [None req-7527ee8e-e60c-46dd-bb63-3e8733859a46 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] Task: {'id': task-1277316, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 616.471968] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b38ee98-672b-44a8-9255-0ae5ccbdeb7b {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.483456] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-3e8259a8-69fb-4692-917a-5e7f03f3483c tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] [instance: c5af147e-3526-4014-98bc-7ad163ae89d9] Powering off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 616.483741] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ac1c026a-a0b9-4dd7-bbf3-2a8a270ad038 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.490850] env[61957]: DEBUG oslo_vmware.api [None req-3e8259a8-69fb-4692-917a-5e7f03f3483c tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] Waiting for the task: (returnval){ [ 616.490850] env[61957]: value = "task-1277317" [ 616.490850] env[61957]: _type = "Task" [ 616.490850] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 616.502447] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4102e0ed-5dd6-47e9-b8ef-7679fc5c078c {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.505090] env[61957]: DEBUG oslo_vmware.api [None req-3e8259a8-69fb-4692-917a-5e7f03f3483c tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] Task: {'id': task-1277317, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 616.509417] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3c34dbb-f94a-46ee-b1de-9f94c59cd71b {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.525664] env[61957]: DEBUG nova.virt.block_device [None req-42e67e16-6377-41f9-9501-426c4335c678 tempest-ServersTestBootFromVolume-482289819 tempest-ServersTestBootFromVolume-482289819-project-member] [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] Updating existing volume attachment record: 83144ec0-7edc-4f38-bd13-fa96cff2dccd {{(pid=61957) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 616.672223] env[61957]: DEBUG oslo_concurrency.lockutils [None req-1b6e3bde-e95e-4682-9e04-c3f69f3be74f tempest-ServersWithSpecificFlavorTestJSON-568377525 tempest-ServersWithSpecificFlavorTestJSON-568377525-project-member] Releasing lock "refresh_cache-f660c725-fc06-42e4-8a34-5d80568b535e" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 616.673328] env[61957]: DEBUG nova.compute.manager [None req-1b6e3bde-e95e-4682-9e04-c3f69f3be74f tempest-ServersWithSpecificFlavorTestJSON-568377525 tempest-ServersWithSpecificFlavorTestJSON-568377525-project-member] [instance: f660c725-fc06-42e4-8a34-5d80568b535e] Start destroying the instance on the hypervisor. {{(pid=61957) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 616.673467] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-1b6e3bde-e95e-4682-9e04-c3f69f3be74f tempest-ServersWithSpecificFlavorTestJSON-568377525 tempest-ServersWithSpecificFlavorTestJSON-568377525-project-member] [instance: f660c725-fc06-42e4-8a34-5d80568b535e] Destroying instance {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 616.673792] env[61957]: DEBUG oslo_concurrency.lockutils [req-dc58e1b1-ce88-4ef6-8580-5ae1b680a1a6 req-e5750388-ea73-44e2-900d-00b98640485a service nova] Acquired lock "refresh_cache-f660c725-fc06-42e4-8a34-5d80568b535e" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 616.673962] env[61957]: DEBUG nova.network.neutron [req-dc58e1b1-ce88-4ef6-8580-5ae1b680a1a6 req-e5750388-ea73-44e2-900d-00b98640485a service nova] [instance: f660c725-fc06-42e4-8a34-5d80568b535e] Refreshing network info cache for port 511dfb63-c95d-4f13-9a94-1bbc532cbb38 {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 616.675103] env[61957]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1f3ed15d-55b3-4bd3-98a8-533c1ddf2b64 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.689820] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd6ea918-1af3-4024-93df-242bbd079596 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.717595] env[61957]: WARNING nova.virt.vmwareapi.vmops [None req-1b6e3bde-e95e-4682-9e04-c3f69f3be74f tempest-ServersWithSpecificFlavorTestJSON-568377525 tempest-ServersWithSpecificFlavorTestJSON-568377525-project-member] [instance: f660c725-fc06-42e4-8a34-5d80568b535e] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance f660c725-fc06-42e4-8a34-5d80568b535e could not be found. [ 616.717953] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-1b6e3bde-e95e-4682-9e04-c3f69f3be74f tempest-ServersWithSpecificFlavorTestJSON-568377525 tempest-ServersWithSpecificFlavorTestJSON-568377525-project-member] [instance: f660c725-fc06-42e4-8a34-5d80568b535e] Instance destroyed {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 616.718303] env[61957]: INFO nova.compute.manager [None req-1b6e3bde-e95e-4682-9e04-c3f69f3be74f tempest-ServersWithSpecificFlavorTestJSON-568377525 tempest-ServersWithSpecificFlavorTestJSON-568377525-project-member] [instance: f660c725-fc06-42e4-8a34-5d80568b535e] Took 0.04 seconds to destroy the instance on the hypervisor. [ 616.718829] env[61957]: DEBUG oslo.service.loopingcall [None req-1b6e3bde-e95e-4682-9e04-c3f69f3be74f tempest-ServersWithSpecificFlavorTestJSON-568377525 tempest-ServersWithSpecificFlavorTestJSON-568377525-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 616.719125] env[61957]: DEBUG nova.compute.manager [-] [instance: f660c725-fc06-42e4-8a34-5d80568b535e] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 616.719244] env[61957]: DEBUG nova.network.neutron [-] [instance: f660c725-fc06-42e4-8a34-5d80568b535e] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 616.737932] env[61957]: DEBUG nova.network.neutron [-] [instance: f660c725-fc06-42e4-8a34-5d80568b535e] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 616.814844] env[61957]: DEBUG nova.scheduler.client.report [None req-260de5ae-9540-404d-a23e-9facd81096ee tempest-AttachInterfacesUnderV243Test-1229304814 tempest-AttachInterfacesUnderV243Test-1229304814-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 616.953506] env[61957]: DEBUG oslo_vmware.api [None req-7527ee8e-e60c-46dd-bb63-3e8733859a46 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] Task: {'id': task-1277316, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.463557} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 616.953652] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-7527ee8e-e60c-46dd-bb63-3e8733859a46 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore1] 4ad746d0-4076-4c97-8ea9-20e49fda59ca/4ad746d0-4076-4c97-8ea9-20e49fda59ca.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 616.953929] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-7527ee8e-e60c-46dd-bb63-3e8733859a46 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] [instance: 4ad746d0-4076-4c97-8ea9-20e49fda59ca] Extending root virtual disk to 1048576 {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 616.954117] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0d9554cf-55b8-40ad-a5ea-9ad0b0fe60a1 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.960781] env[61957]: DEBUG oslo_vmware.api [None req-7527ee8e-e60c-46dd-bb63-3e8733859a46 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] Waiting for the task: (returnval){ [ 616.960781] env[61957]: value = "task-1277318" [ 616.960781] env[61957]: _type = "Task" [ 616.960781] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 616.969399] env[61957]: DEBUG oslo_vmware.api [None req-7527ee8e-e60c-46dd-bb63-3e8733859a46 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] Task: {'id': task-1277318, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 617.004770] env[61957]: DEBUG oslo_vmware.api [None req-3e8259a8-69fb-4692-917a-5e7f03f3483c tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] Task: {'id': task-1277317, 'name': PowerOffVM_Task, 'duration_secs': 0.12813} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 617.005042] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-3e8259a8-69fb-4692-917a-5e7f03f3483c tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] [instance: c5af147e-3526-4014-98bc-7ad163ae89d9] Powered off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 617.005237] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-3e8259a8-69fb-4692-917a-5e7f03f3483c tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] [instance: c5af147e-3526-4014-98bc-7ad163ae89d9] Unregistering the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 617.005917] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f976e636-76f5-4599-96b9-20fc3bf70ef3 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.041418] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-3e8259a8-69fb-4692-917a-5e7f03f3483c tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] [instance: c5af147e-3526-4014-98bc-7ad163ae89d9] Unregistered the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 617.041418] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-3e8259a8-69fb-4692-917a-5e7f03f3483c tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] [instance: c5af147e-3526-4014-98bc-7ad163ae89d9] Deleting contents of the VM from datastore datastore2 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 617.041418] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-3e8259a8-69fb-4692-917a-5e7f03f3483c tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] Deleting the datastore file [datastore2] c5af147e-3526-4014-98bc-7ad163ae89d9 {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 617.041418] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-dc972c93-f6ed-44b4-8f96-b7f9e015d38f {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.047508] env[61957]: DEBUG oslo_vmware.api [None req-3e8259a8-69fb-4692-917a-5e7f03f3483c tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] Waiting for the task: (returnval){ [ 617.047508] env[61957]: value = "task-1277320" [ 617.047508] env[61957]: _type = "Task" [ 617.047508] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 617.055274] env[61957]: DEBUG oslo_vmware.api [None req-3e8259a8-69fb-4692-917a-5e7f03f3483c tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] Task: {'id': task-1277320, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 617.136972] env[61957]: ERROR nova.compute.manager [None req-42e67e16-6377-41f9-9501-426c4335c678 tempest-ServersTestBootFromVolume-482289819 tempest-ServersTestBootFromVolume-482289819-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port bb5c53cf-6a2a-4efe-9be1-07da3dd400d6, please check neutron logs for more information. [ 617.136972] env[61957]: ERROR nova.compute.manager Traceback (most recent call last): [ 617.136972] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 617.136972] env[61957]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 617.136972] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 617.136972] env[61957]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 617.136972] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 617.136972] env[61957]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 617.136972] env[61957]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 617.136972] env[61957]: ERROR nova.compute.manager self.force_reraise() [ 617.136972] env[61957]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 617.136972] env[61957]: ERROR nova.compute.manager raise self.value [ 617.136972] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 617.136972] env[61957]: ERROR nova.compute.manager updated_port = self._update_port( [ 617.136972] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 617.136972] env[61957]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 617.137934] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 617.137934] env[61957]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 617.137934] env[61957]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port bb5c53cf-6a2a-4efe-9be1-07da3dd400d6, please check neutron logs for more information. [ 617.137934] env[61957]: ERROR nova.compute.manager [ 617.137934] env[61957]: Traceback (most recent call last): [ 617.137934] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 617.137934] env[61957]: listener.cb(fileno) [ 617.137934] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 617.137934] env[61957]: result = function(*args, **kwargs) [ 617.137934] env[61957]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 617.137934] env[61957]: return func(*args, **kwargs) [ 617.137934] env[61957]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 617.137934] env[61957]: raise e [ 617.137934] env[61957]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 617.137934] env[61957]: nwinfo = self.network_api.allocate_for_instance( [ 617.137934] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 617.137934] env[61957]: created_port_ids = self._update_ports_for_instance( [ 617.137934] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 617.137934] env[61957]: with excutils.save_and_reraise_exception(): [ 617.137934] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 617.137934] env[61957]: self.force_reraise() [ 617.137934] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 617.137934] env[61957]: raise self.value [ 617.137934] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 617.137934] env[61957]: updated_port = self._update_port( [ 617.137934] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 617.137934] env[61957]: _ensure_no_port_binding_failure(port) [ 617.137934] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 617.137934] env[61957]: raise exception.PortBindingFailed(port_id=port['id']) [ 617.138800] env[61957]: nova.exception.PortBindingFailed: Binding failed for port bb5c53cf-6a2a-4efe-9be1-07da3dd400d6, please check neutron logs for more information. [ 617.138800] env[61957]: Removing descriptor: 17 [ 617.212681] env[61957]: DEBUG nova.network.neutron [req-dc58e1b1-ce88-4ef6-8580-5ae1b680a1a6 req-e5750388-ea73-44e2-900d-00b98640485a service nova] [instance: f660c725-fc06-42e4-8a34-5d80568b535e] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 617.240440] env[61957]: DEBUG nova.network.neutron [-] [instance: f660c725-fc06-42e4-8a34-5d80568b535e] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 617.329129] env[61957]: DEBUG oslo_concurrency.lockutils [None req-260de5ae-9540-404d-a23e-9facd81096ee tempest-AttachInterfacesUnderV243Test-1229304814 tempest-AttachInterfacesUnderV243Test-1229304814-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.513s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 617.329129] env[61957]: DEBUG nova.compute.manager [None req-260de5ae-9540-404d-a23e-9facd81096ee tempest-AttachInterfacesUnderV243Test-1229304814 tempest-AttachInterfacesUnderV243Test-1229304814-project-member] [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] Start building networks asynchronously for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 617.333404] env[61957]: DEBUG oslo_concurrency.lockutils [None req-e4765209-cb68-4173-9918-09c43244c91d tempest-ServerMetadataTestJSON-1199197451 tempest-ServerMetadataTestJSON-1199197451-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.369s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 617.385177] env[61957]: DEBUG nova.network.neutron [req-dc58e1b1-ce88-4ef6-8580-5ae1b680a1a6 req-e5750388-ea73-44e2-900d-00b98640485a service nova] [instance: f660c725-fc06-42e4-8a34-5d80568b535e] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 617.477529] env[61957]: DEBUG oslo_vmware.api [None req-7527ee8e-e60c-46dd-bb63-3e8733859a46 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] Task: {'id': task-1277318, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068546} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 617.478222] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-7527ee8e-e60c-46dd-bb63-3e8733859a46 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] [instance: 4ad746d0-4076-4c97-8ea9-20e49fda59ca] Extended root virtual disk {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 617.478801] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a9d2ca2-1ae0-4669-96d6-925e6ce86978 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.509553] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-7527ee8e-e60c-46dd-bb63-3e8733859a46 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] [instance: 4ad746d0-4076-4c97-8ea9-20e49fda59ca] Reconfiguring VM instance instance-0000000c to attach disk [datastore1] 4ad746d0-4076-4c97-8ea9-20e49fda59ca/4ad746d0-4076-4c97-8ea9-20e49fda59ca.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 617.509988] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-809a5088-82c6-4e8d-be99-31dc81452d17 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.532890] env[61957]: DEBUG oslo_vmware.api [None req-7527ee8e-e60c-46dd-bb63-3e8733859a46 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] Waiting for the task: (returnval){ [ 617.532890] env[61957]: value = "task-1277321" [ 617.532890] env[61957]: _type = "Task" [ 617.532890] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 617.544667] env[61957]: DEBUG oslo_vmware.api [None req-7527ee8e-e60c-46dd-bb63-3e8733859a46 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] Task: {'id': task-1277321, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 617.557964] env[61957]: DEBUG oslo_vmware.api [None req-3e8259a8-69fb-4692-917a-5e7f03f3483c tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] Task: {'id': task-1277320, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.090526} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 617.558303] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-3e8259a8-69fb-4692-917a-5e7f03f3483c tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] Deleted the datastore file {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 617.558508] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-3e8259a8-69fb-4692-917a-5e7f03f3483c tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] [instance: c5af147e-3526-4014-98bc-7ad163ae89d9] Deleted contents of the VM from datastore datastore2 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 617.558696] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-3e8259a8-69fb-4692-917a-5e7f03f3483c tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] [instance: c5af147e-3526-4014-98bc-7ad163ae89d9] Instance destroyed {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 617.558965] env[61957]: INFO nova.compute.manager [None req-3e8259a8-69fb-4692-917a-5e7f03f3483c tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] [instance: c5af147e-3526-4014-98bc-7ad163ae89d9] Took 1.11 seconds to destroy the instance on the hypervisor. [ 617.559201] env[61957]: DEBUG oslo.service.loopingcall [None req-3e8259a8-69fb-4692-917a-5e7f03f3483c tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 617.559346] env[61957]: DEBUG nova.compute.manager [-] [instance: c5af147e-3526-4014-98bc-7ad163ae89d9] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 617.559573] env[61957]: DEBUG nova.network.neutron [-] [instance: c5af147e-3526-4014-98bc-7ad163ae89d9] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 617.591529] env[61957]: DEBUG nova.network.neutron [-] [instance: c5af147e-3526-4014-98bc-7ad163ae89d9] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 617.743253] env[61957]: INFO nova.compute.manager [-] [instance: f660c725-fc06-42e4-8a34-5d80568b535e] Took 1.02 seconds to deallocate network for instance. [ 617.746439] env[61957]: DEBUG nova.compute.claims [None req-1b6e3bde-e95e-4682-9e04-c3f69f3be74f tempest-ServersWithSpecificFlavorTestJSON-568377525 tempest-ServersWithSpecificFlavorTestJSON-568377525-project-member] [instance: f660c725-fc06-42e4-8a34-5d80568b535e] Aborting claim: {{(pid=61957) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 617.746867] env[61957]: DEBUG oslo_concurrency.lockutils [None req-1b6e3bde-e95e-4682-9e04-c3f69f3be74f tempest-ServersWithSpecificFlavorTestJSON-568377525 tempest-ServersWithSpecificFlavorTestJSON-568377525-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 617.837993] env[61957]: DEBUG nova.compute.utils [None req-260de5ae-9540-404d-a23e-9facd81096ee tempest-AttachInterfacesUnderV243Test-1229304814 tempest-AttachInterfacesUnderV243Test-1229304814-project-member] Using /dev/sd instead of None {{(pid=61957) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 617.837993] env[61957]: DEBUG nova.compute.manager [None req-260de5ae-9540-404d-a23e-9facd81096ee tempest-AttachInterfacesUnderV243Test-1229304814 tempest-AttachInterfacesUnderV243Test-1229304814-project-member] [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] Allocating IP information in the background. {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 617.837993] env[61957]: DEBUG nova.network.neutron [None req-260de5ae-9540-404d-a23e-9facd81096ee tempest-AttachInterfacesUnderV243Test-1229304814 tempest-AttachInterfacesUnderV243Test-1229304814-project-member] [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] allocate_for_instance() {{(pid=61957) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 617.890701] env[61957]: DEBUG oslo_concurrency.lockutils [req-dc58e1b1-ce88-4ef6-8580-5ae1b680a1a6 req-e5750388-ea73-44e2-900d-00b98640485a service nova] Releasing lock "refresh_cache-f660c725-fc06-42e4-8a34-5d80568b535e" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 617.913817] env[61957]: DEBUG nova.policy [None req-260de5ae-9540-404d-a23e-9facd81096ee tempest-AttachInterfacesUnderV243Test-1229304814 tempest-AttachInterfacesUnderV243Test-1229304814-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0986ad62085246c0a3610b0d5a08681c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b5bfeac9795548c4bf2a1c2eed3d8146', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61957) authorize /opt/stack/nova/nova/policy.py:203}} [ 618.043745] env[61957]: DEBUG oslo_vmware.api [None req-7527ee8e-e60c-46dd-bb63-3e8733859a46 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] Task: {'id': task-1277321, 'name': ReconfigVM_Task, 'duration_secs': 0.286183} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 618.047696] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-7527ee8e-e60c-46dd-bb63-3e8733859a46 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] [instance: 4ad746d0-4076-4c97-8ea9-20e49fda59ca] Reconfigured VM instance instance-0000000c to attach disk [datastore1] 4ad746d0-4076-4c97-8ea9-20e49fda59ca/4ad746d0-4076-4c97-8ea9-20e49fda59ca.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 618.053364] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1755be59-9b39-4262-a9a8-3c0194fbe695 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.061842] env[61957]: DEBUG oslo_vmware.api [None req-7527ee8e-e60c-46dd-bb63-3e8733859a46 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] Waiting for the task: (returnval){ [ 618.061842] env[61957]: value = "task-1277322" [ 618.061842] env[61957]: _type = "Task" [ 618.061842] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 618.075550] env[61957]: DEBUG oslo_vmware.api [None req-7527ee8e-e60c-46dd-bb63-3e8733859a46 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] Task: {'id': task-1277322, 'name': Rename_Task} progress is 10%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 618.094713] env[61957]: DEBUG nova.network.neutron [-] [instance: c5af147e-3526-4014-98bc-7ad163ae89d9] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 618.249874] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-221136cd-63ba-4a11-955b-8e0fa88e02ea {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.256972] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4c93ee2-2d23-428e-9f97-ac7c59115ea5 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.294660] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc2785ec-37d1-4778-93ed-2719c3085f21 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.304123] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-411ee9f5-b5cf-4b97-8599-d4118ab5182d {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.321852] env[61957]: DEBUG nova.compute.provider_tree [None req-e4765209-cb68-4173-9918-09c43244c91d tempest-ServerMetadataTestJSON-1199197451 tempest-ServerMetadataTestJSON-1199197451-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 618.345702] env[61957]: DEBUG nova.compute.manager [None req-260de5ae-9540-404d-a23e-9facd81096ee tempest-AttachInterfacesUnderV243Test-1229304814 tempest-AttachInterfacesUnderV243Test-1229304814-project-member] [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] Start building block device mappings for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 618.485320] env[61957]: DEBUG nova.network.neutron [None req-260de5ae-9540-404d-a23e-9facd81096ee tempest-AttachInterfacesUnderV243Test-1229304814 tempest-AttachInterfacesUnderV243Test-1229304814-project-member] [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] Successfully created port: 3e020214-1678-4e89-930e-bb37e8f3e31b {{(pid=61957) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 618.578160] env[61957]: DEBUG oslo_vmware.api [None req-7527ee8e-e60c-46dd-bb63-3e8733859a46 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] Task: {'id': task-1277322, 'name': Rename_Task, 'duration_secs': 0.126595} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 618.579079] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-7527ee8e-e60c-46dd-bb63-3e8733859a46 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] [instance: 4ad746d0-4076-4c97-8ea9-20e49fda59ca] Powering on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 618.579079] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-be953f5e-c986-4ba7-9f6b-1c5878fb77d2 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.587123] env[61957]: DEBUG oslo_vmware.api [None req-7527ee8e-e60c-46dd-bb63-3e8733859a46 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] Waiting for the task: (returnval){ [ 618.587123] env[61957]: value = "task-1277323" [ 618.587123] env[61957]: _type = "Task" [ 618.587123] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 618.598897] env[61957]: INFO nova.compute.manager [-] [instance: c5af147e-3526-4014-98bc-7ad163ae89d9] Took 1.04 seconds to deallocate network for instance. [ 618.599248] env[61957]: DEBUG oslo_vmware.api [None req-7527ee8e-e60c-46dd-bb63-3e8733859a46 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] Task: {'id': task-1277323, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 618.685207] env[61957]: DEBUG nova.compute.manager [req-c91acb3c-ce80-473f-b7e9-8898d7323707 req-a8751f8b-8922-439e-b828-d745c4f3567a service nova] [instance: f660c725-fc06-42e4-8a34-5d80568b535e] Received event network-vif-deleted-511dfb63-c95d-4f13-9a94-1bbc532cbb38 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 618.685639] env[61957]: DEBUG nova.compute.manager [req-c91acb3c-ce80-473f-b7e9-8898d7323707 req-a8751f8b-8922-439e-b828-d745c4f3567a service nova] [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] Received event network-changed-bb5c53cf-6a2a-4efe-9be1-07da3dd400d6 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 618.685803] env[61957]: DEBUG nova.compute.manager [req-c91acb3c-ce80-473f-b7e9-8898d7323707 req-a8751f8b-8922-439e-b828-d745c4f3567a service nova] [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] Refreshing instance network info cache due to event network-changed-bb5c53cf-6a2a-4efe-9be1-07da3dd400d6. {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 618.689233] env[61957]: DEBUG oslo_concurrency.lockutils [req-c91acb3c-ce80-473f-b7e9-8898d7323707 req-a8751f8b-8922-439e-b828-d745c4f3567a service nova] Acquiring lock "refresh_cache-32e56693-5100-4ed8-b26a-7d377f9ab7b6" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 618.689420] env[61957]: DEBUG oslo_concurrency.lockutils [req-c91acb3c-ce80-473f-b7e9-8898d7323707 req-a8751f8b-8922-439e-b828-d745c4f3567a service nova] Acquired lock "refresh_cache-32e56693-5100-4ed8-b26a-7d377f9ab7b6" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 618.689602] env[61957]: DEBUG nova.network.neutron [req-c91acb3c-ce80-473f-b7e9-8898d7323707 req-a8751f8b-8922-439e-b828-d745c4f3567a service nova] [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] Refreshing network info cache for port bb5c53cf-6a2a-4efe-9be1-07da3dd400d6 {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 618.711415] env[61957]: DEBUG nova.compute.manager [None req-42e67e16-6377-41f9-9501-426c4335c678 tempest-ServersTestBootFromVolume-482289819 tempest-ServersTestBootFromVolume-482289819-project-member] [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] Start spawning the instance on the hypervisor. {{(pid=61957) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 618.712413] env[61957]: DEBUG nova.virt.hardware [None req-42e67e16-6377-41f9-9501-426c4335c678 tempest-ServersTestBootFromVolume-482289819 tempest-ServersTestBootFromVolume-482289819-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T17:22:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 618.712413] env[61957]: DEBUG nova.virt.hardware [None req-42e67e16-6377-41f9-9501-426c4335c678 tempest-ServersTestBootFromVolume-482289819 tempest-ServersTestBootFromVolume-482289819-project-member] Flavor limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 618.712413] env[61957]: DEBUG nova.virt.hardware [None req-42e67e16-6377-41f9-9501-426c4335c678 tempest-ServersTestBootFromVolume-482289819 tempest-ServersTestBootFromVolume-482289819-project-member] Image limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 618.713056] env[61957]: DEBUG nova.virt.hardware [None req-42e67e16-6377-41f9-9501-426c4335c678 tempest-ServersTestBootFromVolume-482289819 tempest-ServersTestBootFromVolume-482289819-project-member] Flavor pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 618.713056] env[61957]: DEBUG nova.virt.hardware [None req-42e67e16-6377-41f9-9501-426c4335c678 tempest-ServersTestBootFromVolume-482289819 tempest-ServersTestBootFromVolume-482289819-project-member] Image pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 618.713056] env[61957]: DEBUG nova.virt.hardware [None req-42e67e16-6377-41f9-9501-426c4335c678 tempest-ServersTestBootFromVolume-482289819 tempest-ServersTestBootFromVolume-482289819-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 618.716033] env[61957]: DEBUG nova.virt.hardware [None req-42e67e16-6377-41f9-9501-426c4335c678 tempest-ServersTestBootFromVolume-482289819 tempest-ServersTestBootFromVolume-482289819-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 618.716033] env[61957]: DEBUG nova.virt.hardware [None req-42e67e16-6377-41f9-9501-426c4335c678 tempest-ServersTestBootFromVolume-482289819 tempest-ServersTestBootFromVolume-482289819-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 618.716033] env[61957]: DEBUG nova.virt.hardware [None req-42e67e16-6377-41f9-9501-426c4335c678 tempest-ServersTestBootFromVolume-482289819 tempest-ServersTestBootFromVolume-482289819-project-member] Got 1 possible topologies {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 618.716033] env[61957]: DEBUG nova.virt.hardware [None req-42e67e16-6377-41f9-9501-426c4335c678 tempest-ServersTestBootFromVolume-482289819 tempest-ServersTestBootFromVolume-482289819-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 618.716033] env[61957]: DEBUG nova.virt.hardware [None req-42e67e16-6377-41f9-9501-426c4335c678 tempest-ServersTestBootFromVolume-482289819 tempest-ServersTestBootFromVolume-482289819-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 618.716502] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7175d533-84c5-45f5-aec0-d908c8ef7b6d {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.728541] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83cd707d-32c7-4b76-8cc9-6e014cde40d6 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.750948] env[61957]: ERROR nova.compute.manager [None req-42e67e16-6377-41f9-9501-426c4335c678 tempest-ServersTestBootFromVolume-482289819 tempest-ServersTestBootFromVolume-482289819-project-member] [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port bb5c53cf-6a2a-4efe-9be1-07da3dd400d6, please check neutron logs for more information. [ 618.750948] env[61957]: ERROR nova.compute.manager [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] Traceback (most recent call last): [ 618.750948] env[61957]: ERROR nova.compute.manager [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 618.750948] env[61957]: ERROR nova.compute.manager [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] yield resources [ 618.750948] env[61957]: ERROR nova.compute.manager [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 618.750948] env[61957]: ERROR nova.compute.manager [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] self.driver.spawn(context, instance, image_meta, [ 618.750948] env[61957]: ERROR nova.compute.manager [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 618.750948] env[61957]: ERROR nova.compute.manager [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 618.750948] env[61957]: ERROR nova.compute.manager [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 618.750948] env[61957]: ERROR nova.compute.manager [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] vm_ref = self.build_virtual_machine(instance, [ 618.750948] env[61957]: ERROR nova.compute.manager [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 618.751386] env[61957]: ERROR nova.compute.manager [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] vif_infos = vmwarevif.get_vif_info(self._session, [ 618.751386] env[61957]: ERROR nova.compute.manager [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 618.751386] env[61957]: ERROR nova.compute.manager [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] for vif in network_info: [ 618.751386] env[61957]: ERROR nova.compute.manager [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 618.751386] env[61957]: ERROR nova.compute.manager [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] return self._sync_wrapper(fn, *args, **kwargs) [ 618.751386] env[61957]: ERROR nova.compute.manager [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 618.751386] env[61957]: ERROR nova.compute.manager [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] self.wait() [ 618.751386] env[61957]: ERROR nova.compute.manager [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 618.751386] env[61957]: ERROR nova.compute.manager [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] self[:] = self._gt.wait() [ 618.751386] env[61957]: ERROR nova.compute.manager [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 618.751386] env[61957]: ERROR nova.compute.manager [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] return self._exit_event.wait() [ 618.751386] env[61957]: ERROR nova.compute.manager [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 618.751386] env[61957]: ERROR nova.compute.manager [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] current.throw(*self._exc) [ 618.751751] env[61957]: ERROR nova.compute.manager [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 618.751751] env[61957]: ERROR nova.compute.manager [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] result = function(*args, **kwargs) [ 618.751751] env[61957]: ERROR nova.compute.manager [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 618.751751] env[61957]: ERROR nova.compute.manager [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] return func(*args, **kwargs) [ 618.751751] env[61957]: ERROR nova.compute.manager [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 618.751751] env[61957]: ERROR nova.compute.manager [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] raise e [ 618.751751] env[61957]: ERROR nova.compute.manager [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 618.751751] env[61957]: ERROR nova.compute.manager [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] nwinfo = self.network_api.allocate_for_instance( [ 618.751751] env[61957]: ERROR nova.compute.manager [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 618.751751] env[61957]: ERROR nova.compute.manager [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] created_port_ids = self._update_ports_for_instance( [ 618.751751] env[61957]: ERROR nova.compute.manager [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 618.751751] env[61957]: ERROR nova.compute.manager [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] with excutils.save_and_reraise_exception(): [ 618.751751] env[61957]: ERROR nova.compute.manager [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 618.752122] env[61957]: ERROR nova.compute.manager [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] self.force_reraise() [ 618.752122] env[61957]: ERROR nova.compute.manager [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 618.752122] env[61957]: ERROR nova.compute.manager [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] raise self.value [ 618.752122] env[61957]: ERROR nova.compute.manager [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 618.752122] env[61957]: ERROR nova.compute.manager [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] updated_port = self._update_port( [ 618.752122] env[61957]: ERROR nova.compute.manager [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 618.752122] env[61957]: ERROR nova.compute.manager [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] _ensure_no_port_binding_failure(port) [ 618.752122] env[61957]: ERROR nova.compute.manager [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 618.752122] env[61957]: ERROR nova.compute.manager [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] raise exception.PortBindingFailed(port_id=port['id']) [ 618.752122] env[61957]: ERROR nova.compute.manager [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] nova.exception.PortBindingFailed: Binding failed for port bb5c53cf-6a2a-4efe-9be1-07da3dd400d6, please check neutron logs for more information. [ 618.752122] env[61957]: ERROR nova.compute.manager [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] [ 618.752122] env[61957]: INFO nova.compute.manager [None req-42e67e16-6377-41f9-9501-426c4335c678 tempest-ServersTestBootFromVolume-482289819 tempest-ServersTestBootFromVolume-482289819-project-member] [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] Terminating instance [ 618.761647] env[61957]: DEBUG oslo_concurrency.lockutils [None req-42e67e16-6377-41f9-9501-426c4335c678 tempest-ServersTestBootFromVolume-482289819 tempest-ServersTestBootFromVolume-482289819-project-member] Acquiring lock "refresh_cache-32e56693-5100-4ed8-b26a-7d377f9ab7b6" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 618.824841] env[61957]: DEBUG nova.scheduler.client.report [None req-e4765209-cb68-4173-9918-09c43244c91d tempest-ServerMetadataTestJSON-1199197451 tempest-ServerMetadataTestJSON-1199197451-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 619.098035] env[61957]: DEBUG oslo_vmware.api [None req-7527ee8e-e60c-46dd-bb63-3e8733859a46 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] Task: {'id': task-1277323, 'name': PowerOnVM_Task, 'duration_secs': 0.435893} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 619.098035] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-7527ee8e-e60c-46dd-bb63-3e8733859a46 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] [instance: 4ad746d0-4076-4c97-8ea9-20e49fda59ca] Powered on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 619.098035] env[61957]: INFO nova.compute.manager [None req-7527ee8e-e60c-46dd-bb63-3e8733859a46 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] [instance: 4ad746d0-4076-4c97-8ea9-20e49fda59ca] Took 9.53 seconds to spawn the instance on the hypervisor. [ 619.098035] env[61957]: DEBUG nova.compute.manager [None req-7527ee8e-e60c-46dd-bb63-3e8733859a46 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] [instance: 4ad746d0-4076-4c97-8ea9-20e49fda59ca] Checking state {{(pid=61957) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 619.098574] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8302666-781c-4550-b74b-9f3cb2617c8f {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.110120] env[61957]: DEBUG oslo_concurrency.lockutils [None req-3e8259a8-69fb-4692-917a-5e7f03f3483c tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 619.221537] env[61957]: DEBUG nova.network.neutron [req-c91acb3c-ce80-473f-b7e9-8898d7323707 req-a8751f8b-8922-439e-b828-d745c4f3567a service nova] [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 619.334778] env[61957]: DEBUG oslo_concurrency.lockutils [None req-e4765209-cb68-4173-9918-09c43244c91d tempest-ServerMetadataTestJSON-1199197451 tempest-ServerMetadataTestJSON-1199197451-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.999s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 619.334778] env[61957]: ERROR nova.compute.manager [None req-e4765209-cb68-4173-9918-09c43244c91d tempest-ServerMetadataTestJSON-1199197451 tempest-ServerMetadataTestJSON-1199197451-project-member] [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 8834628f-9270-4eee-9918-dfc555410c7c, please check neutron logs for more information. [ 619.334778] env[61957]: ERROR nova.compute.manager [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] Traceback (most recent call last): [ 619.334778] env[61957]: ERROR nova.compute.manager [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 619.334778] env[61957]: ERROR nova.compute.manager [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] self.driver.spawn(context, instance, image_meta, [ 619.334778] env[61957]: ERROR nova.compute.manager [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 619.334778] env[61957]: ERROR nova.compute.manager [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 619.334778] env[61957]: ERROR nova.compute.manager [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 619.334778] env[61957]: ERROR nova.compute.manager [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] vm_ref = self.build_virtual_machine(instance, [ 619.335331] env[61957]: ERROR nova.compute.manager [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 619.335331] env[61957]: ERROR nova.compute.manager [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] vif_infos = vmwarevif.get_vif_info(self._session, [ 619.335331] env[61957]: ERROR nova.compute.manager [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 619.335331] env[61957]: ERROR nova.compute.manager [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] for vif in network_info: [ 619.335331] env[61957]: ERROR nova.compute.manager [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 619.335331] env[61957]: ERROR nova.compute.manager [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] return self._sync_wrapper(fn, *args, **kwargs) [ 619.335331] env[61957]: ERROR nova.compute.manager [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 619.335331] env[61957]: ERROR nova.compute.manager [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] self.wait() [ 619.335331] env[61957]: ERROR nova.compute.manager [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 619.335331] env[61957]: ERROR nova.compute.manager [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] self[:] = self._gt.wait() [ 619.335331] env[61957]: ERROR nova.compute.manager [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 619.335331] env[61957]: ERROR nova.compute.manager [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] return self._exit_event.wait() [ 619.335331] env[61957]: ERROR nova.compute.manager [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 619.335886] env[61957]: ERROR nova.compute.manager [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] result = hub.switch() [ 619.335886] env[61957]: ERROR nova.compute.manager [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 619.335886] env[61957]: ERROR nova.compute.manager [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] return self.greenlet.switch() [ 619.335886] env[61957]: ERROR nova.compute.manager [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 619.335886] env[61957]: ERROR nova.compute.manager [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] result = function(*args, **kwargs) [ 619.335886] env[61957]: ERROR nova.compute.manager [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 619.335886] env[61957]: ERROR nova.compute.manager [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] return func(*args, **kwargs) [ 619.335886] env[61957]: ERROR nova.compute.manager [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 619.335886] env[61957]: ERROR nova.compute.manager [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] raise e [ 619.335886] env[61957]: ERROR nova.compute.manager [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 619.335886] env[61957]: ERROR nova.compute.manager [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] nwinfo = self.network_api.allocate_for_instance( [ 619.335886] env[61957]: ERROR nova.compute.manager [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 619.335886] env[61957]: ERROR nova.compute.manager [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] created_port_ids = self._update_ports_for_instance( [ 619.336467] env[61957]: ERROR nova.compute.manager [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 619.336467] env[61957]: ERROR nova.compute.manager [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] with excutils.save_and_reraise_exception(): [ 619.336467] env[61957]: ERROR nova.compute.manager [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 619.336467] env[61957]: ERROR nova.compute.manager [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] self.force_reraise() [ 619.336467] env[61957]: ERROR nova.compute.manager [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 619.336467] env[61957]: ERROR nova.compute.manager [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] raise self.value [ 619.336467] env[61957]: ERROR nova.compute.manager [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 619.336467] env[61957]: ERROR nova.compute.manager [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] updated_port = self._update_port( [ 619.336467] env[61957]: ERROR nova.compute.manager [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 619.336467] env[61957]: ERROR nova.compute.manager [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] _ensure_no_port_binding_failure(port) [ 619.336467] env[61957]: ERROR nova.compute.manager [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 619.336467] env[61957]: ERROR nova.compute.manager [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] raise exception.PortBindingFailed(port_id=port['id']) [ 619.337039] env[61957]: ERROR nova.compute.manager [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] nova.exception.PortBindingFailed: Binding failed for port 8834628f-9270-4eee-9918-dfc555410c7c, please check neutron logs for more information. [ 619.337039] env[61957]: ERROR nova.compute.manager [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] [ 619.337039] env[61957]: DEBUG nova.compute.utils [None req-e4765209-cb68-4173-9918-09c43244c91d tempest-ServerMetadataTestJSON-1199197451 tempest-ServerMetadataTestJSON-1199197451-project-member] [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] Binding failed for port 8834628f-9270-4eee-9918-dfc555410c7c, please check neutron logs for more information. {{(pid=61957) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 619.337039] env[61957]: DEBUG oslo_concurrency.lockutils [None req-adf5e3c1-3919-48f3-8d1a-68a60974255e tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.749s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 619.341928] env[61957]: INFO nova.compute.claims [None req-adf5e3c1-3919-48f3-8d1a-68a60974255e tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] [instance: 3c099888-d5dc-4cee-9729-311808694625] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 619.342542] env[61957]: DEBUG nova.compute.manager [None req-e4765209-cb68-4173-9918-09c43244c91d tempest-ServerMetadataTestJSON-1199197451 tempest-ServerMetadataTestJSON-1199197451-project-member] [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] Build of instance d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f was re-scheduled: Binding failed for port 8834628f-9270-4eee-9918-dfc555410c7c, please check neutron logs for more information. {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 619.342974] env[61957]: DEBUG nova.compute.manager [None req-e4765209-cb68-4173-9918-09c43244c91d tempest-ServerMetadataTestJSON-1199197451 tempest-ServerMetadataTestJSON-1199197451-project-member] [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] Unplugging VIFs for instance {{(pid=61957) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 619.343365] env[61957]: DEBUG oslo_concurrency.lockutils [None req-e4765209-cb68-4173-9918-09c43244c91d tempest-ServerMetadataTestJSON-1199197451 tempest-ServerMetadataTestJSON-1199197451-project-member] Acquiring lock "refresh_cache-d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 619.343627] env[61957]: DEBUG oslo_concurrency.lockutils [None req-e4765209-cb68-4173-9918-09c43244c91d tempest-ServerMetadataTestJSON-1199197451 tempest-ServerMetadataTestJSON-1199197451-project-member] Acquired lock "refresh_cache-d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 619.343668] env[61957]: DEBUG nova.network.neutron [None req-e4765209-cb68-4173-9918-09c43244c91d tempest-ServerMetadataTestJSON-1199197451 tempest-ServerMetadataTestJSON-1199197451-project-member] [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 619.357787] env[61957]: DEBUG nova.compute.manager [None req-260de5ae-9540-404d-a23e-9facd81096ee tempest-AttachInterfacesUnderV243Test-1229304814 tempest-AttachInterfacesUnderV243Test-1229304814-project-member] [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] Start spawning the instance on the hypervisor. {{(pid=61957) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 619.392479] env[61957]: DEBUG nova.virt.hardware [None req-260de5ae-9540-404d-a23e-9facd81096ee tempest-AttachInterfacesUnderV243Test-1229304814 tempest-AttachInterfacesUnderV243Test-1229304814-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T17:22:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T17:21:50Z,direct_url=,disk_format='vmdk',id=11c76a2c-f705-470a-ba9d-4657858bab38,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='81d3096d0f094373913fc3dc8f5c3c6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T17:21:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 619.392743] env[61957]: DEBUG nova.virt.hardware [None req-260de5ae-9540-404d-a23e-9facd81096ee tempest-AttachInterfacesUnderV243Test-1229304814 tempest-AttachInterfacesUnderV243Test-1229304814-project-member] Flavor limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 619.392869] env[61957]: DEBUG nova.virt.hardware [None req-260de5ae-9540-404d-a23e-9facd81096ee tempest-AttachInterfacesUnderV243Test-1229304814 tempest-AttachInterfacesUnderV243Test-1229304814-project-member] Image limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 619.393059] env[61957]: DEBUG nova.virt.hardware [None req-260de5ae-9540-404d-a23e-9facd81096ee tempest-AttachInterfacesUnderV243Test-1229304814 tempest-AttachInterfacesUnderV243Test-1229304814-project-member] Flavor pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 619.393208] env[61957]: DEBUG nova.virt.hardware [None req-260de5ae-9540-404d-a23e-9facd81096ee tempest-AttachInterfacesUnderV243Test-1229304814 tempest-AttachInterfacesUnderV243Test-1229304814-project-member] Image pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 619.393351] env[61957]: DEBUG nova.virt.hardware [None req-260de5ae-9540-404d-a23e-9facd81096ee tempest-AttachInterfacesUnderV243Test-1229304814 tempest-AttachInterfacesUnderV243Test-1229304814-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 619.393552] env[61957]: DEBUG nova.virt.hardware [None req-260de5ae-9540-404d-a23e-9facd81096ee tempest-AttachInterfacesUnderV243Test-1229304814 tempest-AttachInterfacesUnderV243Test-1229304814-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 619.393703] env[61957]: DEBUG nova.virt.hardware [None req-260de5ae-9540-404d-a23e-9facd81096ee tempest-AttachInterfacesUnderV243Test-1229304814 tempest-AttachInterfacesUnderV243Test-1229304814-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 619.393858] env[61957]: DEBUG nova.virt.hardware [None req-260de5ae-9540-404d-a23e-9facd81096ee tempest-AttachInterfacesUnderV243Test-1229304814 tempest-AttachInterfacesUnderV243Test-1229304814-project-member] Got 1 possible topologies {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 619.394495] env[61957]: DEBUG nova.virt.hardware [None req-260de5ae-9540-404d-a23e-9facd81096ee tempest-AttachInterfacesUnderV243Test-1229304814 tempest-AttachInterfacesUnderV243Test-1229304814-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 619.394495] env[61957]: DEBUG nova.virt.hardware [None req-260de5ae-9540-404d-a23e-9facd81096ee tempest-AttachInterfacesUnderV243Test-1229304814 tempest-AttachInterfacesUnderV243Test-1229304814-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 619.396146] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0cf40056-0400-448c-bb6c-7f9cf0253baf {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.406806] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80d19d5f-3a12-4ad8-901e-f08b92c2bc18 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.482802] env[61957]: DEBUG nova.network.neutron [req-c91acb3c-ce80-473f-b7e9-8898d7323707 req-a8751f8b-8922-439e-b828-d745c4f3567a service nova] [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 619.628084] env[61957]: INFO nova.compute.manager [None req-7527ee8e-e60c-46dd-bb63-3e8733859a46 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] [instance: 4ad746d0-4076-4c97-8ea9-20e49fda59ca] Took 27.40 seconds to build instance. [ 619.987622] env[61957]: DEBUG oslo_concurrency.lockutils [req-c91acb3c-ce80-473f-b7e9-8898d7323707 req-a8751f8b-8922-439e-b828-d745c4f3567a service nova] Releasing lock "refresh_cache-32e56693-5100-4ed8-b26a-7d377f9ab7b6" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 619.987622] env[61957]: DEBUG nova.compute.manager [req-c91acb3c-ce80-473f-b7e9-8898d7323707 req-a8751f8b-8922-439e-b828-d745c4f3567a service nova] [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] Received event network-vif-deleted-bb5c53cf-6a2a-4efe-9be1-07da3dd400d6 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 619.987622] env[61957]: DEBUG oslo_concurrency.lockutils [None req-42e67e16-6377-41f9-9501-426c4335c678 tempest-ServersTestBootFromVolume-482289819 tempest-ServersTestBootFromVolume-482289819-project-member] Acquired lock "refresh_cache-32e56693-5100-4ed8-b26a-7d377f9ab7b6" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 619.987622] env[61957]: DEBUG nova.network.neutron [None req-42e67e16-6377-41f9-9501-426c4335c678 tempest-ServersTestBootFromVolume-482289819 tempest-ServersTestBootFromVolume-482289819-project-member] [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 620.012130] env[61957]: DEBUG nova.network.neutron [None req-e4765209-cb68-4173-9918-09c43244c91d tempest-ServerMetadataTestJSON-1199197451 tempest-ServerMetadataTestJSON-1199197451-project-member] [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 620.130782] env[61957]: DEBUG oslo_concurrency.lockutils [None req-7527ee8e-e60c-46dd-bb63-3e8733859a46 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] Lock "4ad746d0-4076-4c97-8ea9-20e49fda59ca" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 38.673s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 620.525450] env[61957]: DEBUG nova.network.neutron [None req-e4765209-cb68-4173-9918-09c43244c91d tempest-ServerMetadataTestJSON-1199197451 tempest-ServerMetadataTestJSON-1199197451-project-member] [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 620.560933] env[61957]: DEBUG nova.network.neutron [None req-42e67e16-6377-41f9-9501-426c4335c678 tempest-ServersTestBootFromVolume-482289819 tempest-ServersTestBootFromVolume-482289819-project-member] [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 620.634065] env[61957]: DEBUG nova.compute.manager [None req-af971b81-6348-4338-b0a9-b06675d34db5 tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] Starting instance... {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 620.758965] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c46acdc1-3741-418d-bda5-c79980e19998 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.766200] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cb7bf85-4bf5-4ae3-9a8e-c5f40ba72d88 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.801639] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3181b97d-bae9-4f29-b7fb-c6291f64e969 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.810696] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be0b30c0-d66b-466f-9016-cdf83be5664f {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.828139] env[61957]: DEBUG nova.compute.provider_tree [None req-adf5e3c1-3919-48f3-8d1a-68a60974255e tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 620.883652] env[61957]: DEBUG nova.network.neutron [None req-42e67e16-6377-41f9-9501-426c4335c678 tempest-ServersTestBootFromVolume-482289819 tempest-ServersTestBootFromVolume-482289819-project-member] [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 621.028638] env[61957]: DEBUG oslo_concurrency.lockutils [None req-e4765209-cb68-4173-9918-09c43244c91d tempest-ServerMetadataTestJSON-1199197451 tempest-ServerMetadataTestJSON-1199197451-project-member] Releasing lock "refresh_cache-d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 621.029326] env[61957]: DEBUG nova.compute.manager [None req-e4765209-cb68-4173-9918-09c43244c91d tempest-ServerMetadataTestJSON-1199197451 tempest-ServerMetadataTestJSON-1199197451-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61957) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 621.029549] env[61957]: DEBUG nova.compute.manager [None req-e4765209-cb68-4173-9918-09c43244c91d tempest-ServerMetadataTestJSON-1199197451 tempest-ServerMetadataTestJSON-1199197451-project-member] [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 621.029846] env[61957]: DEBUG nova.network.neutron [None req-e4765209-cb68-4173-9918-09c43244c91d tempest-ServerMetadataTestJSON-1199197451 tempest-ServerMetadataTestJSON-1199197451-project-member] [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 621.067183] env[61957]: DEBUG nova.network.neutron [None req-e4765209-cb68-4173-9918-09c43244c91d tempest-ServerMetadataTestJSON-1199197451 tempest-ServerMetadataTestJSON-1199197451-project-member] [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 621.169366] env[61957]: DEBUG oslo_concurrency.lockutils [None req-af971b81-6348-4338-b0a9-b06675d34db5 tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 621.330929] env[61957]: DEBUG nova.scheduler.client.report [None req-adf5e3c1-3919-48f3-8d1a-68a60974255e tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 621.386878] env[61957]: DEBUG oslo_concurrency.lockutils [None req-42e67e16-6377-41f9-9501-426c4335c678 tempest-ServersTestBootFromVolume-482289819 tempest-ServersTestBootFromVolume-482289819-project-member] Releasing lock "refresh_cache-32e56693-5100-4ed8-b26a-7d377f9ab7b6" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 621.388136] env[61957]: DEBUG nova.compute.manager [None req-42e67e16-6377-41f9-9501-426c4335c678 tempest-ServersTestBootFromVolume-482289819 tempest-ServersTestBootFromVolume-482289819-project-member] [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] Start destroying the instance on the hypervisor. {{(pid=61957) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 621.389933] env[61957]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d60ce696-a361-4150-840c-bb7ef443274d {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.400080] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d3fe697-c238-4f11-9c63-fafe5f447fa9 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.429098] env[61957]: WARNING nova.virt.vmwareapi.driver [None req-42e67e16-6377-41f9-9501-426c4335c678 tempest-ServersTestBootFromVolume-482289819 tempest-ServersTestBootFromVolume-482289819-project-member] [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] Instance does not exists. Proceeding to delete instance properties on datastore: nova.exception.InstanceNotFound: Instance 32e56693-5100-4ed8-b26a-7d377f9ab7b6 could not be found. [ 621.429525] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-42e67e16-6377-41f9-9501-426c4335c678 tempest-ServersTestBootFromVolume-482289819 tempest-ServersTestBootFromVolume-482289819-project-member] [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] Destroying instance {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 621.430016] env[61957]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9c7f0f20-d815-4cae-b1a9-3312424e87ea {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.440872] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de26d811-dc3a-4fd0-9822-5646731b0d21 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.469482] env[61957]: WARNING nova.virt.vmwareapi.vmops [None req-42e67e16-6377-41f9-9501-426c4335c678 tempest-ServersTestBootFromVolume-482289819 tempest-ServersTestBootFromVolume-482289819-project-member] [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 32e56693-5100-4ed8-b26a-7d377f9ab7b6 could not be found. [ 621.469784] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-42e67e16-6377-41f9-9501-426c4335c678 tempest-ServersTestBootFromVolume-482289819 tempest-ServersTestBootFromVolume-482289819-project-member] [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] Instance destroyed {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 621.470099] env[61957]: INFO nova.compute.manager [None req-42e67e16-6377-41f9-9501-426c4335c678 tempest-ServersTestBootFromVolume-482289819 tempest-ServersTestBootFromVolume-482289819-project-member] [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] Took 0.08 seconds to destroy the instance on the hypervisor. [ 621.470269] env[61957]: DEBUG oslo.service.loopingcall [None req-42e67e16-6377-41f9-9501-426c4335c678 tempest-ServersTestBootFromVolume-482289819 tempest-ServersTestBootFromVolume-482289819-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 621.470503] env[61957]: DEBUG nova.compute.manager [-] [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 621.470610] env[61957]: DEBUG nova.network.neutron [-] [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 621.515708] env[61957]: DEBUG nova.network.neutron [-] [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 621.569921] env[61957]: DEBUG nova.network.neutron [None req-e4765209-cb68-4173-9918-09c43244c91d tempest-ServerMetadataTestJSON-1199197451 tempest-ServerMetadataTestJSON-1199197451-project-member] [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 621.616387] env[61957]: ERROR nova.compute.manager [None req-260de5ae-9540-404d-a23e-9facd81096ee tempest-AttachInterfacesUnderV243Test-1229304814 tempest-AttachInterfacesUnderV243Test-1229304814-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 3e020214-1678-4e89-930e-bb37e8f3e31b, please check neutron logs for more information. [ 621.616387] env[61957]: ERROR nova.compute.manager Traceback (most recent call last): [ 621.616387] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 621.616387] env[61957]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 621.616387] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 621.616387] env[61957]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 621.616387] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 621.616387] env[61957]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 621.616387] env[61957]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 621.616387] env[61957]: ERROR nova.compute.manager self.force_reraise() [ 621.616387] env[61957]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 621.616387] env[61957]: ERROR nova.compute.manager raise self.value [ 621.616387] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 621.616387] env[61957]: ERROR nova.compute.manager updated_port = self._update_port( [ 621.616387] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 621.616387] env[61957]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 621.616918] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 621.616918] env[61957]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 621.616918] env[61957]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 3e020214-1678-4e89-930e-bb37e8f3e31b, please check neutron logs for more information. [ 621.616918] env[61957]: ERROR nova.compute.manager [ 621.616918] env[61957]: Traceback (most recent call last): [ 621.616918] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 621.616918] env[61957]: listener.cb(fileno) [ 621.616918] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 621.616918] env[61957]: result = function(*args, **kwargs) [ 621.616918] env[61957]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 621.616918] env[61957]: return func(*args, **kwargs) [ 621.616918] env[61957]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 621.616918] env[61957]: raise e [ 621.616918] env[61957]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 621.616918] env[61957]: nwinfo = self.network_api.allocate_for_instance( [ 621.616918] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 621.616918] env[61957]: created_port_ids = self._update_ports_for_instance( [ 621.616918] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 621.616918] env[61957]: with excutils.save_and_reraise_exception(): [ 621.616918] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 621.616918] env[61957]: self.force_reraise() [ 621.616918] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 621.616918] env[61957]: raise self.value [ 621.616918] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 621.616918] env[61957]: updated_port = self._update_port( [ 621.616918] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 621.616918] env[61957]: _ensure_no_port_binding_failure(port) [ 621.616918] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 621.616918] env[61957]: raise exception.PortBindingFailed(port_id=port['id']) [ 621.617792] env[61957]: nova.exception.PortBindingFailed: Binding failed for port 3e020214-1678-4e89-930e-bb37e8f3e31b, please check neutron logs for more information. [ 621.617792] env[61957]: Removing descriptor: 17 [ 621.619407] env[61957]: ERROR nova.compute.manager [None req-260de5ae-9540-404d-a23e-9facd81096ee tempest-AttachInterfacesUnderV243Test-1229304814 tempest-AttachInterfacesUnderV243Test-1229304814-project-member] [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 3e020214-1678-4e89-930e-bb37e8f3e31b, please check neutron logs for more information. [ 621.619407] env[61957]: ERROR nova.compute.manager [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] Traceback (most recent call last): [ 621.619407] env[61957]: ERROR nova.compute.manager [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 621.619407] env[61957]: ERROR nova.compute.manager [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] yield resources [ 621.619407] env[61957]: ERROR nova.compute.manager [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 621.619407] env[61957]: ERROR nova.compute.manager [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] self.driver.spawn(context, instance, image_meta, [ 621.619407] env[61957]: ERROR nova.compute.manager [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 621.619407] env[61957]: ERROR nova.compute.manager [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] self._vmops.spawn(context, instance, image_meta, injected_files, [ 621.619407] env[61957]: ERROR nova.compute.manager [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 621.619407] env[61957]: ERROR nova.compute.manager [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] vm_ref = self.build_virtual_machine(instance, [ 621.619407] env[61957]: ERROR nova.compute.manager [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 621.620358] env[61957]: ERROR nova.compute.manager [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] vif_infos = vmwarevif.get_vif_info(self._session, [ 621.620358] env[61957]: ERROR nova.compute.manager [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 621.620358] env[61957]: ERROR nova.compute.manager [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] for vif in network_info: [ 621.620358] env[61957]: ERROR nova.compute.manager [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 621.620358] env[61957]: ERROR nova.compute.manager [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] return self._sync_wrapper(fn, *args, **kwargs) [ 621.620358] env[61957]: ERROR nova.compute.manager [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 621.620358] env[61957]: ERROR nova.compute.manager [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] self.wait() [ 621.620358] env[61957]: ERROR nova.compute.manager [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 621.620358] env[61957]: ERROR nova.compute.manager [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] self[:] = self._gt.wait() [ 621.620358] env[61957]: ERROR nova.compute.manager [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 621.620358] env[61957]: ERROR nova.compute.manager [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] return self._exit_event.wait() [ 621.620358] env[61957]: ERROR nova.compute.manager [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 621.620358] env[61957]: ERROR nova.compute.manager [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] result = hub.switch() [ 621.620777] env[61957]: ERROR nova.compute.manager [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 621.620777] env[61957]: ERROR nova.compute.manager [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] return self.greenlet.switch() [ 621.620777] env[61957]: ERROR nova.compute.manager [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 621.620777] env[61957]: ERROR nova.compute.manager [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] result = function(*args, **kwargs) [ 621.620777] env[61957]: ERROR nova.compute.manager [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 621.620777] env[61957]: ERROR nova.compute.manager [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] return func(*args, **kwargs) [ 621.620777] env[61957]: ERROR nova.compute.manager [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 621.620777] env[61957]: ERROR nova.compute.manager [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] raise e [ 621.620777] env[61957]: ERROR nova.compute.manager [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 621.620777] env[61957]: ERROR nova.compute.manager [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] nwinfo = self.network_api.allocate_for_instance( [ 621.620777] env[61957]: ERROR nova.compute.manager [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 621.620777] env[61957]: ERROR nova.compute.manager [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] created_port_ids = self._update_ports_for_instance( [ 621.620777] env[61957]: ERROR nova.compute.manager [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 621.621219] env[61957]: ERROR nova.compute.manager [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] with excutils.save_and_reraise_exception(): [ 621.621219] env[61957]: ERROR nova.compute.manager [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 621.621219] env[61957]: ERROR nova.compute.manager [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] self.force_reraise() [ 621.621219] env[61957]: ERROR nova.compute.manager [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 621.621219] env[61957]: ERROR nova.compute.manager [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] raise self.value [ 621.621219] env[61957]: ERROR nova.compute.manager [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 621.621219] env[61957]: ERROR nova.compute.manager [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] updated_port = self._update_port( [ 621.621219] env[61957]: ERROR nova.compute.manager [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 621.621219] env[61957]: ERROR nova.compute.manager [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] _ensure_no_port_binding_failure(port) [ 621.621219] env[61957]: ERROR nova.compute.manager [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 621.621219] env[61957]: ERROR nova.compute.manager [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] raise exception.PortBindingFailed(port_id=port['id']) [ 621.621219] env[61957]: ERROR nova.compute.manager [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] nova.exception.PortBindingFailed: Binding failed for port 3e020214-1678-4e89-930e-bb37e8f3e31b, please check neutron logs for more information. [ 621.621219] env[61957]: ERROR nova.compute.manager [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] [ 621.621678] env[61957]: INFO nova.compute.manager [None req-260de5ae-9540-404d-a23e-9facd81096ee tempest-AttachInterfacesUnderV243Test-1229304814 tempest-AttachInterfacesUnderV243Test-1229304814-project-member] [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] Terminating instance [ 621.623978] env[61957]: DEBUG oslo_concurrency.lockutils [None req-260de5ae-9540-404d-a23e-9facd81096ee tempest-AttachInterfacesUnderV243Test-1229304814 tempest-AttachInterfacesUnderV243Test-1229304814-project-member] Acquiring lock "refresh_cache-02d215d8-dd42-49cc-9509-08335c7bafe3" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 621.624341] env[61957]: DEBUG oslo_concurrency.lockutils [None req-260de5ae-9540-404d-a23e-9facd81096ee tempest-AttachInterfacesUnderV243Test-1229304814 tempest-AttachInterfacesUnderV243Test-1229304814-project-member] Acquired lock "refresh_cache-02d215d8-dd42-49cc-9509-08335c7bafe3" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 621.624786] env[61957]: DEBUG nova.network.neutron [None req-260de5ae-9540-404d-a23e-9facd81096ee tempest-AttachInterfacesUnderV243Test-1229304814 tempest-AttachInterfacesUnderV243Test-1229304814-project-member] [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 621.798923] env[61957]: DEBUG nova.compute.manager [req-801c2770-1f8e-460f-b013-99b0545f1596 req-ff5053a6-5bae-4133-8053-5cc6aab8f837 service nova] [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] Received event network-changed-3e020214-1678-4e89-930e-bb37e8f3e31b {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 621.799241] env[61957]: DEBUG nova.compute.manager [req-801c2770-1f8e-460f-b013-99b0545f1596 req-ff5053a6-5bae-4133-8053-5cc6aab8f837 service nova] [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] Refreshing instance network info cache due to event network-changed-3e020214-1678-4e89-930e-bb37e8f3e31b. {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 621.799470] env[61957]: DEBUG oslo_concurrency.lockutils [req-801c2770-1f8e-460f-b013-99b0545f1596 req-ff5053a6-5bae-4133-8053-5cc6aab8f837 service nova] Acquiring lock "refresh_cache-02d215d8-dd42-49cc-9509-08335c7bafe3" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 621.838597] env[61957]: DEBUG oslo_concurrency.lockutils [None req-adf5e3c1-3919-48f3-8d1a-68a60974255e tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.503s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 621.839162] env[61957]: DEBUG nova.compute.manager [None req-adf5e3c1-3919-48f3-8d1a-68a60974255e tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] [instance: 3c099888-d5dc-4cee-9729-311808694625] Start building networks asynchronously for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 621.842263] env[61957]: DEBUG oslo_concurrency.lockutils [None req-1b94927c-57d2-4b99-982c-53a5e69214ea tempest-VolumesAssistedSnapshotsTest-190365324 tempest-VolumesAssistedSnapshotsTest-190365324-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.988s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 622.019565] env[61957]: DEBUG nova.network.neutron [-] [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 622.076904] env[61957]: INFO nova.compute.manager [None req-e4765209-cb68-4173-9918-09c43244c91d tempest-ServerMetadataTestJSON-1199197451 tempest-ServerMetadataTestJSON-1199197451-project-member] [instance: d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f] Took 1.05 seconds to deallocate network for instance. [ 622.101536] env[61957]: DEBUG nova.compute.manager [None req-8680a796-a008-4b3e-b09f-6cdebb64d987 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] [instance: 4ad746d0-4076-4c97-8ea9-20e49fda59ca] Checking state {{(pid=61957) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 622.104188] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-661bcb6b-c21a-4625-8ef1-e24a9682e491 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.178430] env[61957]: DEBUG nova.network.neutron [None req-260de5ae-9540-404d-a23e-9facd81096ee tempest-AttachInterfacesUnderV243Test-1229304814 tempest-AttachInterfacesUnderV243Test-1229304814-project-member] [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 622.348720] env[61957]: DEBUG nova.compute.utils [None req-adf5e3c1-3919-48f3-8d1a-68a60974255e tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] Using /dev/sd instead of None {{(pid=61957) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 622.355053] env[61957]: DEBUG nova.compute.manager [None req-adf5e3c1-3919-48f3-8d1a-68a60974255e tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] [instance: 3c099888-d5dc-4cee-9729-311808694625] Allocating IP information in the background. {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 622.355053] env[61957]: DEBUG nova.network.neutron [None req-adf5e3c1-3919-48f3-8d1a-68a60974255e tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] [instance: 3c099888-d5dc-4cee-9729-311808694625] allocate_for_instance() {{(pid=61957) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 622.524529] env[61957]: INFO nova.compute.manager [-] [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] Took 1.05 seconds to deallocate network for instance. [ 622.536624] env[61957]: DEBUG nova.network.neutron [None req-260de5ae-9540-404d-a23e-9facd81096ee tempest-AttachInterfacesUnderV243Test-1229304814 tempest-AttachInterfacesUnderV243Test-1229304814-project-member] [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 622.578154] env[61957]: DEBUG nova.policy [None req-adf5e3c1-3919-48f3-8d1a-68a60974255e tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '73ccf8befeac4f6f8eac11ebdb700684', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3d2369599ed547b9964e02d3549958dd', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61957) authorize /opt/stack/nova/nova/policy.py:203}} [ 622.627054] env[61957]: INFO nova.compute.manager [None req-8680a796-a008-4b3e-b09f-6cdebb64d987 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] [instance: 4ad746d0-4076-4c97-8ea9-20e49fda59ca] instance snapshotting [ 622.628123] env[61957]: DEBUG nova.objects.instance [None req-8680a796-a008-4b3e-b09f-6cdebb64d987 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] Lazy-loading 'flavor' on Instance uuid 4ad746d0-4076-4c97-8ea9-20e49fda59ca {{(pid=61957) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 622.772345] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c82916c-e2bd-405e-aece-767b2f1df77a {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.789705] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0bbf3d1-679c-41fc-ae99-e130be25183c {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.823035] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1216718e-a766-4cbb-8bbd-e3fba69504bd {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.831777] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-927aedd7-2ba3-435c-b877-3efd638d5bf3 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.846973] env[61957]: DEBUG nova.compute.provider_tree [None req-1b94927c-57d2-4b99-982c-53a5e69214ea tempest-VolumesAssistedSnapshotsTest-190365324 tempest-VolumesAssistedSnapshotsTest-190365324-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 622.855249] env[61957]: DEBUG nova.compute.manager [None req-adf5e3c1-3919-48f3-8d1a-68a60974255e tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] [instance: 3c099888-d5dc-4cee-9729-311808694625] Start building block device mappings for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 623.040668] env[61957]: DEBUG oslo_concurrency.lockutils [None req-260de5ae-9540-404d-a23e-9facd81096ee tempest-AttachInterfacesUnderV243Test-1229304814 tempest-AttachInterfacesUnderV243Test-1229304814-project-member] Releasing lock "refresh_cache-02d215d8-dd42-49cc-9509-08335c7bafe3" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 623.041116] env[61957]: DEBUG nova.compute.manager [None req-260de5ae-9540-404d-a23e-9facd81096ee tempest-AttachInterfacesUnderV243Test-1229304814 tempest-AttachInterfacesUnderV243Test-1229304814-project-member] [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] Start destroying the instance on the hypervisor. {{(pid=61957) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 623.041543] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-260de5ae-9540-404d-a23e-9facd81096ee tempest-AttachInterfacesUnderV243Test-1229304814 tempest-AttachInterfacesUnderV243Test-1229304814-project-member] [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] Destroying instance {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 623.041665] env[61957]: DEBUG oslo_concurrency.lockutils [req-801c2770-1f8e-460f-b013-99b0545f1596 req-ff5053a6-5bae-4133-8053-5cc6aab8f837 service nova] Acquired lock "refresh_cache-02d215d8-dd42-49cc-9509-08335c7bafe3" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 623.041932] env[61957]: DEBUG nova.network.neutron [req-801c2770-1f8e-460f-b013-99b0545f1596 req-ff5053a6-5bae-4133-8053-5cc6aab8f837 service nova] [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] Refreshing network info cache for port 3e020214-1678-4e89-930e-bb37e8f3e31b {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 623.043428] env[61957]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0f93b564-3df8-4e4a-ad78-48088c52c336 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.054904] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b08ae3b0-023e-4f3b-9926-2d251c77f579 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.088052] env[61957]: WARNING nova.virt.vmwareapi.vmops [None req-260de5ae-9540-404d-a23e-9facd81096ee tempest-AttachInterfacesUnderV243Test-1229304814 tempest-AttachInterfacesUnderV243Test-1229304814-project-member] [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 02d215d8-dd42-49cc-9509-08335c7bafe3 could not be found. [ 623.088322] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-260de5ae-9540-404d-a23e-9facd81096ee tempest-AttachInterfacesUnderV243Test-1229304814 tempest-AttachInterfacesUnderV243Test-1229304814-project-member] [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] Instance destroyed {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 623.088556] env[61957]: INFO nova.compute.manager [None req-260de5ae-9540-404d-a23e-9facd81096ee tempest-AttachInterfacesUnderV243Test-1229304814 tempest-AttachInterfacesUnderV243Test-1229304814-project-member] [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] Took 0.05 seconds to destroy the instance on the hypervisor. [ 623.088873] env[61957]: DEBUG oslo.service.loopingcall [None req-260de5ae-9540-404d-a23e-9facd81096ee tempest-AttachInterfacesUnderV243Test-1229304814 tempest-AttachInterfacesUnderV243Test-1229304814-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 623.092068] env[61957]: DEBUG nova.compute.manager [-] [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 623.092134] env[61957]: DEBUG nova.network.neutron [-] [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 623.126295] env[61957]: INFO nova.scheduler.client.report [None req-e4765209-cb68-4173-9918-09c43244c91d tempest-ServerMetadataTestJSON-1199197451 tempest-ServerMetadataTestJSON-1199197451-project-member] Deleted allocations for instance d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f [ 623.139197] env[61957]: INFO nova.compute.manager [None req-42e67e16-6377-41f9-9501-426c4335c678 tempest-ServersTestBootFromVolume-482289819 tempest-ServersTestBootFromVolume-482289819-project-member] [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] Took 0.61 seconds to detach 1 volumes for instance. [ 623.141878] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc60d25c-4653-47fc-b629-b5db62069ca9 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.163218] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-923bad54-c6a3-420e-8f20-9b0d1703fcd4 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.165749] env[61957]: DEBUG nova.network.neutron [-] [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 623.178202] env[61957]: DEBUG nova.compute.claims [None req-42e67e16-6377-41f9-9501-426c4335c678 tempest-ServersTestBootFromVolume-482289819 tempest-ServersTestBootFromVolume-482289819-project-member] [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] Aborting claim: {{(pid=61957) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 623.178422] env[61957]: DEBUG oslo_concurrency.lockutils [None req-42e67e16-6377-41f9-9501-426c4335c678 tempest-ServersTestBootFromVolume-482289819 tempest-ServersTestBootFromVolume-482289819-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 623.351868] env[61957]: DEBUG nova.scheduler.client.report [None req-1b94927c-57d2-4b99-982c-53a5e69214ea tempest-VolumesAssistedSnapshotsTest-190365324 tempest-VolumesAssistedSnapshotsTest-190365324-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 623.530586] env[61957]: DEBUG oslo_concurrency.lockutils [None req-a12d1497-5e6b-4b01-8612-e7e1d2fcba68 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] Acquiring lock "4ad746d0-4076-4c97-8ea9-20e49fda59ca" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 623.530724] env[61957]: DEBUG oslo_concurrency.lockutils [None req-a12d1497-5e6b-4b01-8612-e7e1d2fcba68 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] Lock "4ad746d0-4076-4c97-8ea9-20e49fda59ca" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 623.530953] env[61957]: DEBUG oslo_concurrency.lockutils [None req-a12d1497-5e6b-4b01-8612-e7e1d2fcba68 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] Acquiring lock "4ad746d0-4076-4c97-8ea9-20e49fda59ca-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 623.531163] env[61957]: DEBUG oslo_concurrency.lockutils [None req-a12d1497-5e6b-4b01-8612-e7e1d2fcba68 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] Lock "4ad746d0-4076-4c97-8ea9-20e49fda59ca-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 623.531336] env[61957]: DEBUG oslo_concurrency.lockutils [None req-a12d1497-5e6b-4b01-8612-e7e1d2fcba68 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] Lock "4ad746d0-4076-4c97-8ea9-20e49fda59ca-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 623.540031] env[61957]: INFO nova.compute.manager [None req-a12d1497-5e6b-4b01-8612-e7e1d2fcba68 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] [instance: 4ad746d0-4076-4c97-8ea9-20e49fda59ca] Terminating instance [ 623.542360] env[61957]: DEBUG oslo_concurrency.lockutils [None req-a12d1497-5e6b-4b01-8612-e7e1d2fcba68 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] Acquiring lock "refresh_cache-4ad746d0-4076-4c97-8ea9-20e49fda59ca" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 623.542577] env[61957]: DEBUG oslo_concurrency.lockutils [None req-a12d1497-5e6b-4b01-8612-e7e1d2fcba68 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] Acquired lock "refresh_cache-4ad746d0-4076-4c97-8ea9-20e49fda59ca" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 623.542806] env[61957]: DEBUG nova.network.neutron [None req-a12d1497-5e6b-4b01-8612-e7e1d2fcba68 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] [instance: 4ad746d0-4076-4c97-8ea9-20e49fda59ca] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 623.644663] env[61957]: DEBUG oslo_concurrency.lockutils [None req-e4765209-cb68-4173-9918-09c43244c91d tempest-ServerMetadataTestJSON-1199197451 tempest-ServerMetadataTestJSON-1199197451-project-member] Lock "d7a929e2-c3fd-4fcb-b6e1-ae1caae6da1f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 44.897s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 623.668383] env[61957]: DEBUG nova.network.neutron [req-801c2770-1f8e-460f-b013-99b0545f1596 req-ff5053a6-5bae-4133-8053-5cc6aab8f837 service nova] [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 623.671901] env[61957]: DEBUG nova.network.neutron [-] [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 623.677817] env[61957]: DEBUG nova.compute.manager [None req-8680a796-a008-4b3e-b09f-6cdebb64d987 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] [instance: 4ad746d0-4076-4c97-8ea9-20e49fda59ca] Instance disappeared during snapshot {{(pid=61957) _snapshot_instance /opt/stack/nova/nova/compute/manager.py:4494}} [ 623.833740] env[61957]: DEBUG nova.compute.manager [None req-8680a796-a008-4b3e-b09f-6cdebb64d987 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] [instance: 4ad746d0-4076-4c97-8ea9-20e49fda59ca] Found 0 images (rotation: 2) {{(pid=61957) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4554}} [ 623.857560] env[61957]: DEBUG oslo_concurrency.lockutils [None req-1b94927c-57d2-4b99-982c-53a5e69214ea tempest-VolumesAssistedSnapshotsTest-190365324 tempest-VolumesAssistedSnapshotsTest-190365324-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.015s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 623.858855] env[61957]: ERROR nova.compute.manager [None req-1b94927c-57d2-4b99-982c-53a5e69214ea tempest-VolumesAssistedSnapshotsTest-190365324 tempest-VolumesAssistedSnapshotsTest-190365324-project-member] [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 6af72034-3abb-47c4-94e2-1b47e7c69c0a, please check neutron logs for more information. [ 623.858855] env[61957]: ERROR nova.compute.manager [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] Traceback (most recent call last): [ 623.858855] env[61957]: ERROR nova.compute.manager [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 623.858855] env[61957]: ERROR nova.compute.manager [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] self.driver.spawn(context, instance, image_meta, [ 623.858855] env[61957]: ERROR nova.compute.manager [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 623.858855] env[61957]: ERROR nova.compute.manager [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] self._vmops.spawn(context, instance, image_meta, injected_files, [ 623.858855] env[61957]: ERROR nova.compute.manager [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 623.858855] env[61957]: ERROR nova.compute.manager [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] vm_ref = self.build_virtual_machine(instance, [ 623.858855] env[61957]: ERROR nova.compute.manager [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 623.858855] env[61957]: ERROR nova.compute.manager [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] vif_infos = vmwarevif.get_vif_info(self._session, [ 623.858855] env[61957]: ERROR nova.compute.manager [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 623.859410] env[61957]: ERROR nova.compute.manager [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] for vif in network_info: [ 623.859410] env[61957]: ERROR nova.compute.manager [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 623.859410] env[61957]: ERROR nova.compute.manager [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] return self._sync_wrapper(fn, *args, **kwargs) [ 623.859410] env[61957]: ERROR nova.compute.manager [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 623.859410] env[61957]: ERROR nova.compute.manager [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] self.wait() [ 623.859410] env[61957]: ERROR nova.compute.manager [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 623.859410] env[61957]: ERROR nova.compute.manager [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] self[:] = self._gt.wait() [ 623.859410] env[61957]: ERROR nova.compute.manager [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 623.859410] env[61957]: ERROR nova.compute.manager [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] return self._exit_event.wait() [ 623.859410] env[61957]: ERROR nova.compute.manager [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 623.859410] env[61957]: ERROR nova.compute.manager [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] result = hub.switch() [ 623.859410] env[61957]: ERROR nova.compute.manager [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 623.859410] env[61957]: ERROR nova.compute.manager [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] return self.greenlet.switch() [ 623.860290] env[61957]: ERROR nova.compute.manager [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 623.860290] env[61957]: ERROR nova.compute.manager [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] result = function(*args, **kwargs) [ 623.860290] env[61957]: ERROR nova.compute.manager [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 623.860290] env[61957]: ERROR nova.compute.manager [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] return func(*args, **kwargs) [ 623.860290] env[61957]: ERROR nova.compute.manager [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 623.860290] env[61957]: ERROR nova.compute.manager [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] raise e [ 623.860290] env[61957]: ERROR nova.compute.manager [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 623.860290] env[61957]: ERROR nova.compute.manager [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] nwinfo = self.network_api.allocate_for_instance( [ 623.860290] env[61957]: ERROR nova.compute.manager [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 623.860290] env[61957]: ERROR nova.compute.manager [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] created_port_ids = self._update_ports_for_instance( [ 623.860290] env[61957]: ERROR nova.compute.manager [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 623.860290] env[61957]: ERROR nova.compute.manager [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] with excutils.save_and_reraise_exception(): [ 623.860290] env[61957]: ERROR nova.compute.manager [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 623.860889] env[61957]: ERROR nova.compute.manager [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] self.force_reraise() [ 623.860889] env[61957]: ERROR nova.compute.manager [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 623.860889] env[61957]: ERROR nova.compute.manager [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] raise self.value [ 623.860889] env[61957]: ERROR nova.compute.manager [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 623.860889] env[61957]: ERROR nova.compute.manager [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] updated_port = self._update_port( [ 623.860889] env[61957]: ERROR nova.compute.manager [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 623.860889] env[61957]: ERROR nova.compute.manager [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] _ensure_no_port_binding_failure(port) [ 623.860889] env[61957]: ERROR nova.compute.manager [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 623.860889] env[61957]: ERROR nova.compute.manager [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] raise exception.PortBindingFailed(port_id=port['id']) [ 623.860889] env[61957]: ERROR nova.compute.manager [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] nova.exception.PortBindingFailed: Binding failed for port 6af72034-3abb-47c4-94e2-1b47e7c69c0a, please check neutron logs for more information. [ 623.860889] env[61957]: ERROR nova.compute.manager [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] [ 623.861556] env[61957]: DEBUG nova.compute.utils [None req-1b94927c-57d2-4b99-982c-53a5e69214ea tempest-VolumesAssistedSnapshotsTest-190365324 tempest-VolumesAssistedSnapshotsTest-190365324-project-member] [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] Binding failed for port 6af72034-3abb-47c4-94e2-1b47e7c69c0a, please check neutron logs for more information. {{(pid=61957) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 623.861556] env[61957]: DEBUG oslo_concurrency.lockutils [None req-61bcf42d-e42b-40a8-98f8-af8cb211ebed tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.579s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 623.865158] env[61957]: DEBUG nova.compute.manager [None req-adf5e3c1-3919-48f3-8d1a-68a60974255e tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] [instance: 3c099888-d5dc-4cee-9729-311808694625] Start spawning the instance on the hypervisor. {{(pid=61957) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 623.871073] env[61957]: DEBUG nova.compute.manager [None req-1b94927c-57d2-4b99-982c-53a5e69214ea tempest-VolumesAssistedSnapshotsTest-190365324 tempest-VolumesAssistedSnapshotsTest-190365324-project-member] [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] Build of instance ee479d3a-239f-426c-8bdb-484fdf8dcd76 was re-scheduled: Binding failed for port 6af72034-3abb-47c4-94e2-1b47e7c69c0a, please check neutron logs for more information. {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 623.872203] env[61957]: DEBUG nova.compute.manager [None req-1b94927c-57d2-4b99-982c-53a5e69214ea tempest-VolumesAssistedSnapshotsTest-190365324 tempest-VolumesAssistedSnapshotsTest-190365324-project-member] [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] Unplugging VIFs for instance {{(pid=61957) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 623.872203] env[61957]: DEBUG oslo_concurrency.lockutils [None req-1b94927c-57d2-4b99-982c-53a5e69214ea tempest-VolumesAssistedSnapshotsTest-190365324 tempest-VolumesAssistedSnapshotsTest-190365324-project-member] Acquiring lock "refresh_cache-ee479d3a-239f-426c-8bdb-484fdf8dcd76" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 623.872203] env[61957]: DEBUG oslo_concurrency.lockutils [None req-1b94927c-57d2-4b99-982c-53a5e69214ea tempest-VolumesAssistedSnapshotsTest-190365324 tempest-VolumesAssistedSnapshotsTest-190365324-project-member] Acquired lock "refresh_cache-ee479d3a-239f-426c-8bdb-484fdf8dcd76" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 623.872203] env[61957]: DEBUG nova.network.neutron [None req-1b94927c-57d2-4b99-982c-53a5e69214ea tempest-VolumesAssistedSnapshotsTest-190365324 tempest-VolumesAssistedSnapshotsTest-190365324-project-member] [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 623.903164] env[61957]: DEBUG nova.virt.hardware [None req-adf5e3c1-3919-48f3-8d1a-68a60974255e tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T17:22:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T17:21:50Z,direct_url=,disk_format='vmdk',id=11c76a2c-f705-470a-ba9d-4657858bab38,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='81d3096d0f094373913fc3dc8f5c3c6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T17:21:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 623.903353] env[61957]: DEBUG nova.virt.hardware [None req-adf5e3c1-3919-48f3-8d1a-68a60974255e tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] Flavor limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 623.903508] env[61957]: DEBUG nova.virt.hardware [None req-adf5e3c1-3919-48f3-8d1a-68a60974255e tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] Image limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 623.904084] env[61957]: DEBUG nova.virt.hardware [None req-adf5e3c1-3919-48f3-8d1a-68a60974255e tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] Flavor pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 623.904084] env[61957]: DEBUG nova.virt.hardware [None req-adf5e3c1-3919-48f3-8d1a-68a60974255e tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] Image pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 623.904084] env[61957]: DEBUG nova.virt.hardware [None req-adf5e3c1-3919-48f3-8d1a-68a60974255e tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 623.904297] env[61957]: DEBUG nova.virt.hardware [None req-adf5e3c1-3919-48f3-8d1a-68a60974255e tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 623.905624] env[61957]: DEBUG nova.virt.hardware [None req-adf5e3c1-3919-48f3-8d1a-68a60974255e tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 623.905624] env[61957]: DEBUG nova.virt.hardware [None req-adf5e3c1-3919-48f3-8d1a-68a60974255e tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] Got 1 possible topologies {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 623.905624] env[61957]: DEBUG nova.virt.hardware [None req-adf5e3c1-3919-48f3-8d1a-68a60974255e tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 623.905624] env[61957]: DEBUG nova.virt.hardware [None req-adf5e3c1-3919-48f3-8d1a-68a60974255e tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 623.905995] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4224f1bc-f3a4-405d-9750-ef200565a3ed {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.919488] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74afb9c8-4eb8-48be-a789-94573ba1808a {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.017130] env[61957]: DEBUG nova.network.neutron [req-801c2770-1f8e-460f-b013-99b0545f1596 req-ff5053a6-5bae-4133-8053-5cc6aab8f837 service nova] [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 624.079581] env[61957]: DEBUG nova.network.neutron [None req-a12d1497-5e6b-4b01-8612-e7e1d2fcba68 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] [instance: 4ad746d0-4076-4c97-8ea9-20e49fda59ca] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 624.148475] env[61957]: DEBUG nova.compute.manager [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] Starting instance... {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 624.174404] env[61957]: INFO nova.compute.manager [-] [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] Took 1.08 seconds to deallocate network for instance. [ 624.177248] env[61957]: DEBUG nova.compute.claims [None req-260de5ae-9540-404d-a23e-9facd81096ee tempest-AttachInterfacesUnderV243Test-1229304814 tempest-AttachInterfacesUnderV243Test-1229304814-project-member] [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] Aborting claim: {{(pid=61957) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 624.186505] env[61957]: DEBUG oslo_concurrency.lockutils [None req-260de5ae-9540-404d-a23e-9facd81096ee tempest-AttachInterfacesUnderV243Test-1229304814 tempest-AttachInterfacesUnderV243Test-1229304814-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 624.249877] env[61957]: DEBUG nova.network.neutron [None req-a12d1497-5e6b-4b01-8612-e7e1d2fcba68 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] [instance: 4ad746d0-4076-4c97-8ea9-20e49fda59ca] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 624.257112] env[61957]: DEBUG nova.network.neutron [None req-adf5e3c1-3919-48f3-8d1a-68a60974255e tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] [instance: 3c099888-d5dc-4cee-9729-311808694625] Successfully created port: cd92211b-fb72-4034-a229-b18203e54c1e {{(pid=61957) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 624.421484] env[61957]: DEBUG nova.network.neutron [None req-1b94927c-57d2-4b99-982c-53a5e69214ea tempest-VolumesAssistedSnapshotsTest-190365324 tempest-VolumesAssistedSnapshotsTest-190365324-project-member] [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 624.521983] env[61957]: DEBUG oslo_concurrency.lockutils [req-801c2770-1f8e-460f-b013-99b0545f1596 req-ff5053a6-5bae-4133-8053-5cc6aab8f837 service nova] Releasing lock "refresh_cache-02d215d8-dd42-49cc-9509-08335c7bafe3" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 624.677971] env[61957]: DEBUG oslo_concurrency.lockutils [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 624.731686] env[61957]: DEBUG nova.network.neutron [None req-1b94927c-57d2-4b99-982c-53a5e69214ea tempest-VolumesAssistedSnapshotsTest-190365324 tempest-VolumesAssistedSnapshotsTest-190365324-project-member] [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 624.752275] env[61957]: DEBUG oslo_concurrency.lockutils [None req-a12d1497-5e6b-4b01-8612-e7e1d2fcba68 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] Releasing lock "refresh_cache-4ad746d0-4076-4c97-8ea9-20e49fda59ca" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 624.753043] env[61957]: DEBUG nova.compute.manager [None req-a12d1497-5e6b-4b01-8612-e7e1d2fcba68 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] [instance: 4ad746d0-4076-4c97-8ea9-20e49fda59ca] Start destroying the instance on the hypervisor. {{(pid=61957) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 624.753043] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-a12d1497-5e6b-4b01-8612-e7e1d2fcba68 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] [instance: 4ad746d0-4076-4c97-8ea9-20e49fda59ca] Destroying instance {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 624.754174] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-200ca89b-4535-42b8-b07c-ef096b3401c4 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.763835] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-a12d1497-5e6b-4b01-8612-e7e1d2fcba68 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] [instance: 4ad746d0-4076-4c97-8ea9-20e49fda59ca] Powering off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 624.764698] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b1891d49-05eb-4535-8217-ed89bdbcdd99 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.768257] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29033ee7-956a-4dd3-b79e-58a2feb5458d {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.777369] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5213b6f8-3a3b-4390-b0f2-713471e2f9b0 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.781260] env[61957]: DEBUG oslo_vmware.api [None req-a12d1497-5e6b-4b01-8612-e7e1d2fcba68 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] Waiting for the task: (returnval){ [ 624.781260] env[61957]: value = "task-1277327" [ 624.781260] env[61957]: _type = "Task" [ 624.781260] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 624.815059] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1fdd548-d686-4f19-b937-4c8d657e718f {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.821009] env[61957]: DEBUG oslo_vmware.api [None req-a12d1497-5e6b-4b01-8612-e7e1d2fcba68 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] Task: {'id': task-1277327, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 624.826013] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fc1a05e-e6a1-412d-91c5-743d56683d93 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.842085] env[61957]: DEBUG nova.compute.provider_tree [None req-61bcf42d-e42b-40a8-98f8-af8cb211ebed tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 625.164391] env[61957]: DEBUG nova.compute.manager [req-c151688b-6562-46da-8635-63cbd6c7d4d3 req-a4f70d19-7d48-462c-81c6-c58df4ace907 service nova] [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] Received event network-vif-deleted-3e020214-1678-4e89-930e-bb37e8f3e31b {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 625.235527] env[61957]: DEBUG oslo_concurrency.lockutils [None req-1b94927c-57d2-4b99-982c-53a5e69214ea tempest-VolumesAssistedSnapshotsTest-190365324 tempest-VolumesAssistedSnapshotsTest-190365324-project-member] Releasing lock "refresh_cache-ee479d3a-239f-426c-8bdb-484fdf8dcd76" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 625.236309] env[61957]: DEBUG nova.compute.manager [None req-1b94927c-57d2-4b99-982c-53a5e69214ea tempest-VolumesAssistedSnapshotsTest-190365324 tempest-VolumesAssistedSnapshotsTest-190365324-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61957) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 625.236587] env[61957]: DEBUG nova.compute.manager [None req-1b94927c-57d2-4b99-982c-53a5e69214ea tempest-VolumesAssistedSnapshotsTest-190365324 tempest-VolumesAssistedSnapshotsTest-190365324-project-member] [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 625.236828] env[61957]: DEBUG nova.network.neutron [None req-1b94927c-57d2-4b99-982c-53a5e69214ea tempest-VolumesAssistedSnapshotsTest-190365324 tempest-VolumesAssistedSnapshotsTest-190365324-project-member] [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 625.267709] env[61957]: DEBUG nova.network.neutron [None req-1b94927c-57d2-4b99-982c-53a5e69214ea tempest-VolumesAssistedSnapshotsTest-190365324 tempest-VolumesAssistedSnapshotsTest-190365324-project-member] [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 625.303994] env[61957]: DEBUG oslo_vmware.api [None req-a12d1497-5e6b-4b01-8612-e7e1d2fcba68 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] Task: {'id': task-1277327, 'name': PowerOffVM_Task, 'duration_secs': 0.12191} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 625.304357] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-a12d1497-5e6b-4b01-8612-e7e1d2fcba68 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] [instance: 4ad746d0-4076-4c97-8ea9-20e49fda59ca] Powered off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 625.304620] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-a12d1497-5e6b-4b01-8612-e7e1d2fcba68 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] [instance: 4ad746d0-4076-4c97-8ea9-20e49fda59ca] Unregistering the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 625.305016] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7cd37613-daef-424c-b051-608f85f27541 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.332569] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-a12d1497-5e6b-4b01-8612-e7e1d2fcba68 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] [instance: 4ad746d0-4076-4c97-8ea9-20e49fda59ca] Unregistered the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 625.332569] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-a12d1497-5e6b-4b01-8612-e7e1d2fcba68 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] [instance: 4ad746d0-4076-4c97-8ea9-20e49fda59ca] Deleting contents of the VM from datastore datastore1 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 625.332569] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-a12d1497-5e6b-4b01-8612-e7e1d2fcba68 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] Deleting the datastore file [datastore1] 4ad746d0-4076-4c97-8ea9-20e49fda59ca {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 625.332569] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-223dd1b2-ad5c-4fa5-8ade-c2e4797a2f9b {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.342926] env[61957]: DEBUG oslo_vmware.api [None req-a12d1497-5e6b-4b01-8612-e7e1d2fcba68 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] Waiting for the task: (returnval){ [ 625.342926] env[61957]: value = "task-1277329" [ 625.342926] env[61957]: _type = "Task" [ 625.342926] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 625.346995] env[61957]: DEBUG nova.scheduler.client.report [None req-61bcf42d-e42b-40a8-98f8-af8cb211ebed tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 625.357023] env[61957]: DEBUG oslo_vmware.api [None req-a12d1497-5e6b-4b01-8612-e7e1d2fcba68 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] Task: {'id': task-1277329, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 625.777178] env[61957]: DEBUG nova.network.neutron [None req-1b94927c-57d2-4b99-982c-53a5e69214ea tempest-VolumesAssistedSnapshotsTest-190365324 tempest-VolumesAssistedSnapshotsTest-190365324-project-member] [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 625.854778] env[61957]: DEBUG oslo_vmware.api [None req-a12d1497-5e6b-4b01-8612-e7e1d2fcba68 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] Task: {'id': task-1277329, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.087075} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 625.854778] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-a12d1497-5e6b-4b01-8612-e7e1d2fcba68 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] Deleted the datastore file {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 625.854914] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-a12d1497-5e6b-4b01-8612-e7e1d2fcba68 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] [instance: 4ad746d0-4076-4c97-8ea9-20e49fda59ca] Deleted contents of the VM from datastore datastore1 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 625.855118] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-a12d1497-5e6b-4b01-8612-e7e1d2fcba68 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] [instance: 4ad746d0-4076-4c97-8ea9-20e49fda59ca] Instance destroyed {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 625.855340] env[61957]: INFO nova.compute.manager [None req-a12d1497-5e6b-4b01-8612-e7e1d2fcba68 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] [instance: 4ad746d0-4076-4c97-8ea9-20e49fda59ca] Took 1.10 seconds to destroy the instance on the hypervisor. [ 625.855599] env[61957]: DEBUG oslo.service.loopingcall [None req-a12d1497-5e6b-4b01-8612-e7e1d2fcba68 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 625.855788] env[61957]: DEBUG nova.compute.manager [-] [instance: 4ad746d0-4076-4c97-8ea9-20e49fda59ca] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 625.855896] env[61957]: DEBUG nova.network.neutron [-] [instance: 4ad746d0-4076-4c97-8ea9-20e49fda59ca] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 625.859864] env[61957]: DEBUG oslo_concurrency.lockutils [None req-61bcf42d-e42b-40a8-98f8-af8cb211ebed tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.999s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 625.860563] env[61957]: ERROR nova.compute.manager [None req-61bcf42d-e42b-40a8-98f8-af8cb211ebed tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 6a3b4c95-304d-4644-ba91-67780e2d82ae, please check neutron logs for more information. [ 625.860563] env[61957]: ERROR nova.compute.manager [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] Traceback (most recent call last): [ 625.860563] env[61957]: ERROR nova.compute.manager [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 625.860563] env[61957]: ERROR nova.compute.manager [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] self.driver.spawn(context, instance, image_meta, [ 625.860563] env[61957]: ERROR nova.compute.manager [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 625.860563] env[61957]: ERROR nova.compute.manager [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] self._vmops.spawn(context, instance, image_meta, injected_files, [ 625.860563] env[61957]: ERROR nova.compute.manager [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 625.860563] env[61957]: ERROR nova.compute.manager [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] vm_ref = self.build_virtual_machine(instance, [ 625.860563] env[61957]: ERROR nova.compute.manager [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 625.860563] env[61957]: ERROR nova.compute.manager [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] vif_infos = vmwarevif.get_vif_info(self._session, [ 625.860563] env[61957]: ERROR nova.compute.manager [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 625.860937] env[61957]: ERROR nova.compute.manager [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] for vif in network_info: [ 625.860937] env[61957]: ERROR nova.compute.manager [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 625.860937] env[61957]: ERROR nova.compute.manager [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] return self._sync_wrapper(fn, *args, **kwargs) [ 625.860937] env[61957]: ERROR nova.compute.manager [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 625.860937] env[61957]: ERROR nova.compute.manager [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] self.wait() [ 625.860937] env[61957]: ERROR nova.compute.manager [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 625.860937] env[61957]: ERROR nova.compute.manager [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] self[:] = self._gt.wait() [ 625.860937] env[61957]: ERROR nova.compute.manager [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 625.860937] env[61957]: ERROR nova.compute.manager [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] return self._exit_event.wait() [ 625.860937] env[61957]: ERROR nova.compute.manager [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 625.860937] env[61957]: ERROR nova.compute.manager [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] result = hub.switch() [ 625.860937] env[61957]: ERROR nova.compute.manager [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 625.860937] env[61957]: ERROR nova.compute.manager [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] return self.greenlet.switch() [ 625.861448] env[61957]: ERROR nova.compute.manager [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 625.861448] env[61957]: ERROR nova.compute.manager [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] result = function(*args, **kwargs) [ 625.861448] env[61957]: ERROR nova.compute.manager [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 625.861448] env[61957]: ERROR nova.compute.manager [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] return func(*args, **kwargs) [ 625.861448] env[61957]: ERROR nova.compute.manager [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 625.861448] env[61957]: ERROR nova.compute.manager [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] raise e [ 625.861448] env[61957]: ERROR nova.compute.manager [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 625.861448] env[61957]: ERROR nova.compute.manager [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] nwinfo = self.network_api.allocate_for_instance( [ 625.861448] env[61957]: ERROR nova.compute.manager [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 625.861448] env[61957]: ERROR nova.compute.manager [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] created_port_ids = self._update_ports_for_instance( [ 625.861448] env[61957]: ERROR nova.compute.manager [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 625.861448] env[61957]: ERROR nova.compute.manager [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] with excutils.save_and_reraise_exception(): [ 625.861448] env[61957]: ERROR nova.compute.manager [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 625.861828] env[61957]: ERROR nova.compute.manager [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] self.force_reraise() [ 625.861828] env[61957]: ERROR nova.compute.manager [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 625.861828] env[61957]: ERROR nova.compute.manager [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] raise self.value [ 625.861828] env[61957]: ERROR nova.compute.manager [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 625.861828] env[61957]: ERROR nova.compute.manager [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] updated_port = self._update_port( [ 625.861828] env[61957]: ERROR nova.compute.manager [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 625.861828] env[61957]: ERROR nova.compute.manager [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] _ensure_no_port_binding_failure(port) [ 625.861828] env[61957]: ERROR nova.compute.manager [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 625.861828] env[61957]: ERROR nova.compute.manager [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] raise exception.PortBindingFailed(port_id=port['id']) [ 625.861828] env[61957]: ERROR nova.compute.manager [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] nova.exception.PortBindingFailed: Binding failed for port 6a3b4c95-304d-4644-ba91-67780e2d82ae, please check neutron logs for more information. [ 625.861828] env[61957]: ERROR nova.compute.manager [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] [ 625.862184] env[61957]: DEBUG nova.compute.utils [None req-61bcf42d-e42b-40a8-98f8-af8cb211ebed tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] Binding failed for port 6a3b4c95-304d-4644-ba91-67780e2d82ae, please check neutron logs for more information. {{(pid=61957) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 625.863538] env[61957]: DEBUG oslo_concurrency.lockutils [None req-4e7f7a28-6560-4b73-affe-356aa8d11b0b tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.942s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 625.865479] env[61957]: INFO nova.compute.claims [None req-4e7f7a28-6560-4b73-affe-356aa8d11b0b tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 625.868847] env[61957]: DEBUG nova.compute.manager [None req-61bcf42d-e42b-40a8-98f8-af8cb211ebed tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] Build of instance 96c6b747-7293-4252-a1d0-b0d7684c5529 was re-scheduled: Binding failed for port 6a3b4c95-304d-4644-ba91-67780e2d82ae, please check neutron logs for more information. {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 625.869298] env[61957]: DEBUG nova.compute.manager [None req-61bcf42d-e42b-40a8-98f8-af8cb211ebed tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] Unplugging VIFs for instance {{(pid=61957) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 625.869603] env[61957]: DEBUG oslo_concurrency.lockutils [None req-61bcf42d-e42b-40a8-98f8-af8cb211ebed tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] Acquiring lock "refresh_cache-96c6b747-7293-4252-a1d0-b0d7684c5529" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 625.869784] env[61957]: DEBUG oslo_concurrency.lockutils [None req-61bcf42d-e42b-40a8-98f8-af8cb211ebed tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] Acquired lock "refresh_cache-96c6b747-7293-4252-a1d0-b0d7684c5529" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 625.870026] env[61957]: DEBUG nova.network.neutron [None req-61bcf42d-e42b-40a8-98f8-af8cb211ebed tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 625.904509] env[61957]: DEBUG nova.network.neutron [-] [instance: 4ad746d0-4076-4c97-8ea9-20e49fda59ca] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 625.954042] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ca1e8d1e-8c18-4652-924a-ab1cdb751a05 tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] Acquiring lock "825130d4-b1f6-434e-be3e-dca952abe930" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 625.954281] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ca1e8d1e-8c18-4652-924a-ab1cdb751a05 tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] Lock "825130d4-b1f6-434e-be3e-dca952abe930" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 626.283359] env[61957]: INFO nova.compute.manager [None req-1b94927c-57d2-4b99-982c-53a5e69214ea tempest-VolumesAssistedSnapshotsTest-190365324 tempest-VolumesAssistedSnapshotsTest-190365324-project-member] [instance: ee479d3a-239f-426c-8bdb-484fdf8dcd76] Took 1.04 seconds to deallocate network for instance. [ 626.408886] env[61957]: DEBUG nova.network.neutron [-] [instance: 4ad746d0-4076-4c97-8ea9-20e49fda59ca] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 626.433806] env[61957]: DEBUG nova.network.neutron [None req-61bcf42d-e42b-40a8-98f8-af8cb211ebed tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 626.668706] env[61957]: DEBUG nova.network.neutron [None req-61bcf42d-e42b-40a8-98f8-af8cb211ebed tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 626.913884] env[61957]: INFO nova.compute.manager [-] [instance: 4ad746d0-4076-4c97-8ea9-20e49fda59ca] Took 1.06 seconds to deallocate network for instance. [ 627.171683] env[61957]: DEBUG oslo_concurrency.lockutils [None req-61bcf42d-e42b-40a8-98f8-af8cb211ebed tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] Releasing lock "refresh_cache-96c6b747-7293-4252-a1d0-b0d7684c5529" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 627.171928] env[61957]: DEBUG nova.compute.manager [None req-61bcf42d-e42b-40a8-98f8-af8cb211ebed tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61957) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 627.175251] env[61957]: DEBUG nova.compute.manager [None req-61bcf42d-e42b-40a8-98f8-af8cb211ebed tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 627.175251] env[61957]: DEBUG nova.network.neutron [None req-61bcf42d-e42b-40a8-98f8-af8cb211ebed tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 627.324531] env[61957]: INFO nova.scheduler.client.report [None req-1b94927c-57d2-4b99-982c-53a5e69214ea tempest-VolumesAssistedSnapshotsTest-190365324 tempest-VolumesAssistedSnapshotsTest-190365324-project-member] Deleted allocations for instance ee479d3a-239f-426c-8bdb-484fdf8dcd76 [ 627.367206] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e8e0346-cc5f-4f64-9367-4ee0706355f3 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.376837] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28a8f535-7517-456f-b8ad-699dc39bb630 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.411791] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e2feb04-09f9-4a1b-90f3-02532b3f586f {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.418406] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a223d98d-73e0-4ebd-bebb-f3bdce97580b {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.423506] env[61957]: DEBUG oslo_concurrency.lockutils [None req-a12d1497-5e6b-4b01-8612-e7e1d2fcba68 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 627.424858] env[61957]: DEBUG nova.network.neutron [None req-61bcf42d-e42b-40a8-98f8-af8cb211ebed tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 627.434875] env[61957]: DEBUG nova.compute.provider_tree [None req-4e7f7a28-6560-4b73-affe-356aa8d11b0b tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 627.436939] env[61957]: DEBUG nova.network.neutron [None req-61bcf42d-e42b-40a8-98f8-af8cb211ebed tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 627.761623] env[61957]: DEBUG oslo_concurrency.lockutils [None req-58fe0400-f1d0-4e1d-a459-7541ba7aeeed tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] Acquiring lock "c369d148-2093-4724-82cc-d80a67131fea" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 627.761850] env[61957]: DEBUG oslo_concurrency.lockutils [None req-58fe0400-f1d0-4e1d-a459-7541ba7aeeed tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] Lock "c369d148-2093-4724-82cc-d80a67131fea" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 627.838360] env[61957]: DEBUG oslo_concurrency.lockutils [None req-1b94927c-57d2-4b99-982c-53a5e69214ea tempest-VolumesAssistedSnapshotsTest-190365324 tempest-VolumesAssistedSnapshotsTest-190365324-project-member] Lock "ee479d3a-239f-426c-8bdb-484fdf8dcd76" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 47.790s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 627.939678] env[61957]: INFO nova.compute.manager [None req-61bcf42d-e42b-40a8-98f8-af8cb211ebed tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] [instance: 96c6b747-7293-4252-a1d0-b0d7684c5529] Took 0.77 seconds to deallocate network for instance. [ 627.943895] env[61957]: DEBUG nova.scheduler.client.report [None req-4e7f7a28-6560-4b73-affe-356aa8d11b0b tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 628.076415] env[61957]: DEBUG nova.compute.manager [req-8dd16680-425f-4d0c-8259-ea9bc1b5dbd7 req-2cf74538-b0bd-49c6-9d18-d7e462211000 service nova] [instance: 3c099888-d5dc-4cee-9729-311808694625] Received event network-changed-cd92211b-fb72-4034-a229-b18203e54c1e {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 628.076613] env[61957]: DEBUG nova.compute.manager [req-8dd16680-425f-4d0c-8259-ea9bc1b5dbd7 req-2cf74538-b0bd-49c6-9d18-d7e462211000 service nova] [instance: 3c099888-d5dc-4cee-9729-311808694625] Refreshing instance network info cache due to event network-changed-cd92211b-fb72-4034-a229-b18203e54c1e. {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 628.076817] env[61957]: DEBUG oslo_concurrency.lockutils [req-8dd16680-425f-4d0c-8259-ea9bc1b5dbd7 req-2cf74538-b0bd-49c6-9d18-d7e462211000 service nova] Acquiring lock "refresh_cache-3c099888-d5dc-4cee-9729-311808694625" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 628.076953] env[61957]: DEBUG oslo_concurrency.lockutils [req-8dd16680-425f-4d0c-8259-ea9bc1b5dbd7 req-2cf74538-b0bd-49c6-9d18-d7e462211000 service nova] Acquired lock "refresh_cache-3c099888-d5dc-4cee-9729-311808694625" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 628.078256] env[61957]: DEBUG nova.network.neutron [req-8dd16680-425f-4d0c-8259-ea9bc1b5dbd7 req-2cf74538-b0bd-49c6-9d18-d7e462211000 service nova] [instance: 3c099888-d5dc-4cee-9729-311808694625] Refreshing network info cache for port cd92211b-fb72-4034-a229-b18203e54c1e {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 628.102350] env[61957]: ERROR nova.compute.manager [None req-adf5e3c1-3919-48f3-8d1a-68a60974255e tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port cd92211b-fb72-4034-a229-b18203e54c1e, please check neutron logs for more information. [ 628.102350] env[61957]: ERROR nova.compute.manager Traceback (most recent call last): [ 628.102350] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 628.102350] env[61957]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 628.102350] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 628.102350] env[61957]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 628.102350] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 628.102350] env[61957]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 628.102350] env[61957]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 628.102350] env[61957]: ERROR nova.compute.manager self.force_reraise() [ 628.102350] env[61957]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 628.102350] env[61957]: ERROR nova.compute.manager raise self.value [ 628.102350] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 628.102350] env[61957]: ERROR nova.compute.manager updated_port = self._update_port( [ 628.102350] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 628.102350] env[61957]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 628.103073] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 628.103073] env[61957]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 628.103073] env[61957]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port cd92211b-fb72-4034-a229-b18203e54c1e, please check neutron logs for more information. [ 628.103073] env[61957]: ERROR nova.compute.manager [ 628.103073] env[61957]: Traceback (most recent call last): [ 628.103073] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 628.103073] env[61957]: listener.cb(fileno) [ 628.103073] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 628.103073] env[61957]: result = function(*args, **kwargs) [ 628.103073] env[61957]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 628.103073] env[61957]: return func(*args, **kwargs) [ 628.103073] env[61957]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 628.103073] env[61957]: raise e [ 628.103073] env[61957]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 628.103073] env[61957]: nwinfo = self.network_api.allocate_for_instance( [ 628.103073] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 628.103073] env[61957]: created_port_ids = self._update_ports_for_instance( [ 628.103073] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 628.103073] env[61957]: with excutils.save_and_reraise_exception(): [ 628.103073] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 628.103073] env[61957]: self.force_reraise() [ 628.103073] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 628.103073] env[61957]: raise self.value [ 628.103073] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 628.103073] env[61957]: updated_port = self._update_port( [ 628.103073] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 628.103073] env[61957]: _ensure_no_port_binding_failure(port) [ 628.103073] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 628.103073] env[61957]: raise exception.PortBindingFailed(port_id=port['id']) [ 628.104048] env[61957]: nova.exception.PortBindingFailed: Binding failed for port cd92211b-fb72-4034-a229-b18203e54c1e, please check neutron logs for more information. [ 628.104048] env[61957]: Removing descriptor: 15 [ 628.104048] env[61957]: ERROR nova.compute.manager [None req-adf5e3c1-3919-48f3-8d1a-68a60974255e tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] [instance: 3c099888-d5dc-4cee-9729-311808694625] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port cd92211b-fb72-4034-a229-b18203e54c1e, please check neutron logs for more information. [ 628.104048] env[61957]: ERROR nova.compute.manager [instance: 3c099888-d5dc-4cee-9729-311808694625] Traceback (most recent call last): [ 628.104048] env[61957]: ERROR nova.compute.manager [instance: 3c099888-d5dc-4cee-9729-311808694625] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 628.104048] env[61957]: ERROR nova.compute.manager [instance: 3c099888-d5dc-4cee-9729-311808694625] yield resources [ 628.104048] env[61957]: ERROR nova.compute.manager [instance: 3c099888-d5dc-4cee-9729-311808694625] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 628.104048] env[61957]: ERROR nova.compute.manager [instance: 3c099888-d5dc-4cee-9729-311808694625] self.driver.spawn(context, instance, image_meta, [ 628.104048] env[61957]: ERROR nova.compute.manager [instance: 3c099888-d5dc-4cee-9729-311808694625] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 628.104048] env[61957]: ERROR nova.compute.manager [instance: 3c099888-d5dc-4cee-9729-311808694625] self._vmops.spawn(context, instance, image_meta, injected_files, [ 628.104048] env[61957]: ERROR nova.compute.manager [instance: 3c099888-d5dc-4cee-9729-311808694625] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 628.104048] env[61957]: ERROR nova.compute.manager [instance: 3c099888-d5dc-4cee-9729-311808694625] vm_ref = self.build_virtual_machine(instance, [ 628.104475] env[61957]: ERROR nova.compute.manager [instance: 3c099888-d5dc-4cee-9729-311808694625] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 628.104475] env[61957]: ERROR nova.compute.manager [instance: 3c099888-d5dc-4cee-9729-311808694625] vif_infos = vmwarevif.get_vif_info(self._session, [ 628.104475] env[61957]: ERROR nova.compute.manager [instance: 3c099888-d5dc-4cee-9729-311808694625] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 628.104475] env[61957]: ERROR nova.compute.manager [instance: 3c099888-d5dc-4cee-9729-311808694625] for vif in network_info: [ 628.104475] env[61957]: ERROR nova.compute.manager [instance: 3c099888-d5dc-4cee-9729-311808694625] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 628.104475] env[61957]: ERROR nova.compute.manager [instance: 3c099888-d5dc-4cee-9729-311808694625] return self._sync_wrapper(fn, *args, **kwargs) [ 628.104475] env[61957]: ERROR nova.compute.manager [instance: 3c099888-d5dc-4cee-9729-311808694625] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 628.104475] env[61957]: ERROR nova.compute.manager [instance: 3c099888-d5dc-4cee-9729-311808694625] self.wait() [ 628.104475] env[61957]: ERROR nova.compute.manager [instance: 3c099888-d5dc-4cee-9729-311808694625] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 628.104475] env[61957]: ERROR nova.compute.manager [instance: 3c099888-d5dc-4cee-9729-311808694625] self[:] = self._gt.wait() [ 628.104475] env[61957]: ERROR nova.compute.manager [instance: 3c099888-d5dc-4cee-9729-311808694625] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 628.104475] env[61957]: ERROR nova.compute.manager [instance: 3c099888-d5dc-4cee-9729-311808694625] return self._exit_event.wait() [ 628.104475] env[61957]: ERROR nova.compute.manager [instance: 3c099888-d5dc-4cee-9729-311808694625] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 628.104880] env[61957]: ERROR nova.compute.manager [instance: 3c099888-d5dc-4cee-9729-311808694625] result = hub.switch() [ 628.104880] env[61957]: ERROR nova.compute.manager [instance: 3c099888-d5dc-4cee-9729-311808694625] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 628.104880] env[61957]: ERROR nova.compute.manager [instance: 3c099888-d5dc-4cee-9729-311808694625] return self.greenlet.switch() [ 628.104880] env[61957]: ERROR nova.compute.manager [instance: 3c099888-d5dc-4cee-9729-311808694625] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 628.104880] env[61957]: ERROR nova.compute.manager [instance: 3c099888-d5dc-4cee-9729-311808694625] result = function(*args, **kwargs) [ 628.104880] env[61957]: ERROR nova.compute.manager [instance: 3c099888-d5dc-4cee-9729-311808694625] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 628.104880] env[61957]: ERROR nova.compute.manager [instance: 3c099888-d5dc-4cee-9729-311808694625] return func(*args, **kwargs) [ 628.104880] env[61957]: ERROR nova.compute.manager [instance: 3c099888-d5dc-4cee-9729-311808694625] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 628.104880] env[61957]: ERROR nova.compute.manager [instance: 3c099888-d5dc-4cee-9729-311808694625] raise e [ 628.104880] env[61957]: ERROR nova.compute.manager [instance: 3c099888-d5dc-4cee-9729-311808694625] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 628.104880] env[61957]: ERROR nova.compute.manager [instance: 3c099888-d5dc-4cee-9729-311808694625] nwinfo = self.network_api.allocate_for_instance( [ 628.104880] env[61957]: ERROR nova.compute.manager [instance: 3c099888-d5dc-4cee-9729-311808694625] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 628.104880] env[61957]: ERROR nova.compute.manager [instance: 3c099888-d5dc-4cee-9729-311808694625] created_port_ids = self._update_ports_for_instance( [ 628.105330] env[61957]: ERROR nova.compute.manager [instance: 3c099888-d5dc-4cee-9729-311808694625] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 628.105330] env[61957]: ERROR nova.compute.manager [instance: 3c099888-d5dc-4cee-9729-311808694625] with excutils.save_and_reraise_exception(): [ 628.105330] env[61957]: ERROR nova.compute.manager [instance: 3c099888-d5dc-4cee-9729-311808694625] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 628.105330] env[61957]: ERROR nova.compute.manager [instance: 3c099888-d5dc-4cee-9729-311808694625] self.force_reraise() [ 628.105330] env[61957]: ERROR nova.compute.manager [instance: 3c099888-d5dc-4cee-9729-311808694625] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 628.105330] env[61957]: ERROR nova.compute.manager [instance: 3c099888-d5dc-4cee-9729-311808694625] raise self.value [ 628.105330] env[61957]: ERROR nova.compute.manager [instance: 3c099888-d5dc-4cee-9729-311808694625] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 628.105330] env[61957]: ERROR nova.compute.manager [instance: 3c099888-d5dc-4cee-9729-311808694625] updated_port = self._update_port( [ 628.105330] env[61957]: ERROR nova.compute.manager [instance: 3c099888-d5dc-4cee-9729-311808694625] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 628.105330] env[61957]: ERROR nova.compute.manager [instance: 3c099888-d5dc-4cee-9729-311808694625] _ensure_no_port_binding_failure(port) [ 628.105330] env[61957]: ERROR nova.compute.manager [instance: 3c099888-d5dc-4cee-9729-311808694625] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 628.105330] env[61957]: ERROR nova.compute.manager [instance: 3c099888-d5dc-4cee-9729-311808694625] raise exception.PortBindingFailed(port_id=port['id']) [ 628.105809] env[61957]: ERROR nova.compute.manager [instance: 3c099888-d5dc-4cee-9729-311808694625] nova.exception.PortBindingFailed: Binding failed for port cd92211b-fb72-4034-a229-b18203e54c1e, please check neutron logs for more information. [ 628.105809] env[61957]: ERROR nova.compute.manager [instance: 3c099888-d5dc-4cee-9729-311808694625] [ 628.105809] env[61957]: INFO nova.compute.manager [None req-adf5e3c1-3919-48f3-8d1a-68a60974255e tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] [instance: 3c099888-d5dc-4cee-9729-311808694625] Terminating instance [ 628.106491] env[61957]: DEBUG oslo_concurrency.lockutils [None req-adf5e3c1-3919-48f3-8d1a-68a60974255e tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] Acquiring lock "refresh_cache-3c099888-d5dc-4cee-9729-311808694625" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 628.344035] env[61957]: DEBUG nova.compute.manager [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] Starting instance... {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 628.457403] env[61957]: DEBUG oslo_concurrency.lockutils [None req-4e7f7a28-6560-4b73-affe-356aa8d11b0b tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.591s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 628.457403] env[61957]: DEBUG nova.compute.manager [None req-4e7f7a28-6560-4b73-affe-356aa8d11b0b tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] Start building networks asynchronously for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 628.458948] env[61957]: DEBUG oslo_concurrency.lockutils [None req-9c3c6101-e50c-439a-91f6-f293e059d0d2 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.833s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 628.462114] env[61957]: INFO nova.compute.claims [None req-9c3c6101-e50c-439a-91f6-f293e059d0d2 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 628.621080] env[61957]: DEBUG nova.network.neutron [req-8dd16680-425f-4d0c-8259-ea9bc1b5dbd7 req-2cf74538-b0bd-49c6-9d18-d7e462211000 service nova] [instance: 3c099888-d5dc-4cee-9729-311808694625] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 628.796766] env[61957]: DEBUG oslo_concurrency.lockutils [None req-707bfa11-3857-45e5-a4be-581ffade0e04 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Acquiring lock "2252ef6e-19ec-4356-8221-33168c01bef0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 628.796991] env[61957]: DEBUG oslo_concurrency.lockutils [None req-707bfa11-3857-45e5-a4be-581ffade0e04 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Lock "2252ef6e-19ec-4356-8221-33168c01bef0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 628.808484] env[61957]: DEBUG nova.network.neutron [req-8dd16680-425f-4d0c-8259-ea9bc1b5dbd7 req-2cf74538-b0bd-49c6-9d18-d7e462211000 service nova] [instance: 3c099888-d5dc-4cee-9729-311808694625] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 628.867810] env[61957]: DEBUG oslo_concurrency.lockutils [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 628.972121] env[61957]: DEBUG nova.compute.utils [None req-4e7f7a28-6560-4b73-affe-356aa8d11b0b tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] Using /dev/sd instead of None {{(pid=61957) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 628.975115] env[61957]: DEBUG nova.compute.manager [None req-4e7f7a28-6560-4b73-affe-356aa8d11b0b tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] Allocating IP information in the background. {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 628.976746] env[61957]: DEBUG nova.network.neutron [None req-4e7f7a28-6560-4b73-affe-356aa8d11b0b tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] allocate_for_instance() {{(pid=61957) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 628.998248] env[61957]: INFO nova.scheduler.client.report [None req-61bcf42d-e42b-40a8-98f8-af8cb211ebed tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] Deleted allocations for instance 96c6b747-7293-4252-a1d0-b0d7684c5529 [ 629.213512] env[61957]: DEBUG nova.policy [None req-4e7f7a28-6560-4b73-affe-356aa8d11b0b tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '73ccf8befeac4f6f8eac11ebdb700684', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3d2369599ed547b9964e02d3549958dd', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61957) authorize /opt/stack/nova/nova/policy.py:203}} [ 629.312078] env[61957]: DEBUG oslo_concurrency.lockutils [req-8dd16680-425f-4d0c-8259-ea9bc1b5dbd7 req-2cf74538-b0bd-49c6-9d18-d7e462211000 service nova] Releasing lock "refresh_cache-3c099888-d5dc-4cee-9729-311808694625" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 629.315495] env[61957]: DEBUG oslo_concurrency.lockutils [None req-adf5e3c1-3919-48f3-8d1a-68a60974255e tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] Acquired lock "refresh_cache-3c099888-d5dc-4cee-9729-311808694625" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 629.315495] env[61957]: DEBUG nova.network.neutron [None req-adf5e3c1-3919-48f3-8d1a-68a60974255e tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] [instance: 3c099888-d5dc-4cee-9729-311808694625] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 629.476339] env[61957]: DEBUG nova.compute.manager [None req-4e7f7a28-6560-4b73-affe-356aa8d11b0b tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] Start building block device mappings for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 629.487522] env[61957]: DEBUG oslo_concurrency.lockutils [None req-2b9492ff-1cb8-4ba9-ba5e-b513c4b108ad tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] Acquiring lock "f8235178-7ae0-4733-bb05-3515d5489a5b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 629.487787] env[61957]: DEBUG oslo_concurrency.lockutils [None req-2b9492ff-1cb8-4ba9-ba5e-b513c4b108ad tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] Lock "f8235178-7ae0-4733-bb05-3515d5489a5b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 629.510406] env[61957]: DEBUG oslo_concurrency.lockutils [None req-61bcf42d-e42b-40a8-98f8-af8cb211ebed tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] Lock "96c6b747-7293-4252-a1d0-b0d7684c5529" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 48.711s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 629.863383] env[61957]: DEBUG nova.network.neutron [None req-adf5e3c1-3919-48f3-8d1a-68a60974255e tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] [instance: 3c099888-d5dc-4cee-9729-311808694625] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 629.891553] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01bcca0c-61b8-4946-866d-be070e644c3b {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.900122] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57741d5c-9b96-44f0-8ad6-deb0f530d43a {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.937180] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01e2260f-02b2-4f4f-9c00-53f9109259fc {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.946464] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-778b20c1-ad8d-442d-9349-aea62f577695 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.962544] env[61957]: DEBUG nova.compute.provider_tree [None req-9c3c6101-e50c-439a-91f6-f293e059d0d2 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 630.022156] env[61957]: DEBUG nova.compute.manager [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] [instance: 394b1955-c448-42c2-a718-28df7bd366e9] Starting instance... {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 630.051670] env[61957]: DEBUG nova.network.neutron [None req-adf5e3c1-3919-48f3-8d1a-68a60974255e tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] [instance: 3c099888-d5dc-4cee-9729-311808694625] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 630.163287] env[61957]: DEBUG nova.network.neutron [None req-4e7f7a28-6560-4b73-affe-356aa8d11b0b tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] Successfully created port: 09ed1ca2-59f3-48fa-b1a0-4d61495c3c4c {{(pid=61957) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 630.464710] env[61957]: DEBUG nova.scheduler.client.report [None req-9c3c6101-e50c-439a-91f6-f293e059d0d2 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 630.494447] env[61957]: DEBUG nova.compute.manager [None req-4e7f7a28-6560-4b73-affe-356aa8d11b0b tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] Start spawning the instance on the hypervisor. {{(pid=61957) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 630.535911] env[61957]: DEBUG nova.virt.hardware [None req-4e7f7a28-6560-4b73-affe-356aa8d11b0b tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T17:22:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T17:21:50Z,direct_url=,disk_format='vmdk',id=11c76a2c-f705-470a-ba9d-4657858bab38,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='81d3096d0f094373913fc3dc8f5c3c6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T17:21:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 630.536748] env[61957]: DEBUG nova.virt.hardware [None req-4e7f7a28-6560-4b73-affe-356aa8d11b0b tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] Flavor limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 630.536748] env[61957]: DEBUG nova.virt.hardware [None req-4e7f7a28-6560-4b73-affe-356aa8d11b0b tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] Image limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 630.536748] env[61957]: DEBUG nova.virt.hardware [None req-4e7f7a28-6560-4b73-affe-356aa8d11b0b tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] Flavor pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 630.536748] env[61957]: DEBUG nova.virt.hardware [None req-4e7f7a28-6560-4b73-affe-356aa8d11b0b tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] Image pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 630.536946] env[61957]: DEBUG nova.virt.hardware [None req-4e7f7a28-6560-4b73-affe-356aa8d11b0b tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 630.537065] env[61957]: DEBUG nova.virt.hardware [None req-4e7f7a28-6560-4b73-affe-356aa8d11b0b tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 630.537218] env[61957]: DEBUG nova.virt.hardware [None req-4e7f7a28-6560-4b73-affe-356aa8d11b0b tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 630.537439] env[61957]: DEBUG nova.virt.hardware [None req-4e7f7a28-6560-4b73-affe-356aa8d11b0b tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] Got 1 possible topologies {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 630.537623] env[61957]: DEBUG nova.virt.hardware [None req-4e7f7a28-6560-4b73-affe-356aa8d11b0b tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 630.537798] env[61957]: DEBUG nova.virt.hardware [None req-4e7f7a28-6560-4b73-affe-356aa8d11b0b tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 630.542061] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6db831ca-4143-461f-9b9b-84a5728b55aa {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.555763] env[61957]: DEBUG oslo_concurrency.lockutils [None req-adf5e3c1-3919-48f3-8d1a-68a60974255e tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] Releasing lock "refresh_cache-3c099888-d5dc-4cee-9729-311808694625" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 630.560359] env[61957]: DEBUG nova.compute.manager [None req-adf5e3c1-3919-48f3-8d1a-68a60974255e tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] [instance: 3c099888-d5dc-4cee-9729-311808694625] Start destroying the instance on the hypervisor. {{(pid=61957) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 630.561265] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-adf5e3c1-3919-48f3-8d1a-68a60974255e tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] [instance: 3c099888-d5dc-4cee-9729-311808694625] Destroying instance {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 630.561265] env[61957]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2555529b-8603-477a-a2dd-2e46a2136564 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.568058] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8104f8de-69a0-408f-af71-2a8e63d3c387 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.572804] env[61957]: DEBUG oslo_concurrency.lockutils [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 630.591898] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d774b66c-840c-4626-9325-983576632918 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.616850] env[61957]: WARNING nova.virt.vmwareapi.vmops [None req-adf5e3c1-3919-48f3-8d1a-68a60974255e tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] [instance: 3c099888-d5dc-4cee-9729-311808694625] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 3c099888-d5dc-4cee-9729-311808694625 could not be found. [ 630.616850] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-adf5e3c1-3919-48f3-8d1a-68a60974255e tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] [instance: 3c099888-d5dc-4cee-9729-311808694625] Instance destroyed {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 630.616850] env[61957]: INFO nova.compute.manager [None req-adf5e3c1-3919-48f3-8d1a-68a60974255e tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] [instance: 3c099888-d5dc-4cee-9729-311808694625] Took 0.06 seconds to destroy the instance on the hypervisor. [ 630.617378] env[61957]: DEBUG oslo.service.loopingcall [None req-adf5e3c1-3919-48f3-8d1a-68a60974255e tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 630.617908] env[61957]: DEBUG nova.compute.manager [-] [instance: 3c099888-d5dc-4cee-9729-311808694625] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 630.618136] env[61957]: DEBUG nova.network.neutron [-] [instance: 3c099888-d5dc-4cee-9729-311808694625] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 630.672157] env[61957]: DEBUG nova.network.neutron [-] [instance: 3c099888-d5dc-4cee-9729-311808694625] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 630.976198] env[61957]: DEBUG oslo_concurrency.lockutils [None req-9c3c6101-e50c-439a-91f6-f293e059d0d2 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.516s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 630.976198] env[61957]: DEBUG nova.compute.manager [None req-9c3c6101-e50c-439a-91f6-f293e059d0d2 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] Start building networks asynchronously for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 630.980765] env[61957]: DEBUG oslo_concurrency.lockutils [None req-1b6e3bde-e95e-4682-9e04-c3f69f3be74f tempest-ServersWithSpecificFlavorTestJSON-568377525 tempest-ServersWithSpecificFlavorTestJSON-568377525-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.234s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 630.988465] env[61957]: DEBUG nova.compute.manager [req-80266cc4-2e52-420f-805c-d50a57818588 req-064e8ac8-49f4-4b7c-a443-5a6e3a1d62f0 service nova] [instance: 3c099888-d5dc-4cee-9729-311808694625] Received event network-vif-deleted-cd92211b-fb72-4034-a229-b18203e54c1e {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 631.043670] env[61957]: DEBUG oslo_concurrency.lockutils [None req-f462e606-4c03-40e0-ac91-60efbd401559 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Acquiring lock "115c1bb4-0e6b-4c6c-9d38-9584debea01f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 631.043928] env[61957]: DEBUG oslo_concurrency.lockutils [None req-f462e606-4c03-40e0-ac91-60efbd401559 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Lock "115c1bb4-0e6b-4c6c-9d38-9584debea01f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 631.183761] env[61957]: DEBUG nova.network.neutron [-] [instance: 3c099888-d5dc-4cee-9729-311808694625] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 631.493730] env[61957]: DEBUG nova.compute.utils [None req-9c3c6101-e50c-439a-91f6-f293e059d0d2 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] Using /dev/sd instead of None {{(pid=61957) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 631.499371] env[61957]: DEBUG nova.compute.manager [None req-9c3c6101-e50c-439a-91f6-f293e059d0d2 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] Allocating IP information in the background. {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 631.499553] env[61957]: DEBUG nova.network.neutron [None req-9c3c6101-e50c-439a-91f6-f293e059d0d2 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] allocate_for_instance() {{(pid=61957) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 631.680547] env[61957]: DEBUG nova.policy [None req-9c3c6101-e50c-439a-91f6-f293e059d0d2 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'dbec97161e3249759c8743b15ea72873', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e4b7e4d62d074c1f9d66b40b328e5373', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61957) authorize /opt/stack/nova/nova/policy.py:203}} [ 631.691092] env[61957]: INFO nova.compute.manager [-] [instance: 3c099888-d5dc-4cee-9729-311808694625] Took 1.07 seconds to deallocate network for instance. [ 631.695192] env[61957]: DEBUG nova.compute.claims [None req-adf5e3c1-3919-48f3-8d1a-68a60974255e tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] [instance: 3c099888-d5dc-4cee-9729-311808694625] Aborting claim: {{(pid=61957) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 631.695428] env[61957]: DEBUG oslo_concurrency.lockutils [None req-adf5e3c1-3919-48f3-8d1a-68a60974255e tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 631.934292] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a57f4e9a-1642-4a97-8be3-758e1c83d0bf {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.948708] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd946250-6008-4593-9c02-62954b5b16b2 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.985755] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e84714a-a3b5-4b81-b346-8c2289618238 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.993562] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ea49382-2755-47a6-b3d8-68d5915cd44d {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.009970] env[61957]: DEBUG nova.compute.manager [None req-9c3c6101-e50c-439a-91f6-f293e059d0d2 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] Start building block device mappings for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 632.012524] env[61957]: DEBUG nova.compute.provider_tree [None req-1b6e3bde-e95e-4682-9e04-c3f69f3be74f tempest-ServersWithSpecificFlavorTestJSON-568377525 tempest-ServersWithSpecificFlavorTestJSON-568377525-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 632.524455] env[61957]: DEBUG nova.scheduler.client.report [None req-1b6e3bde-e95e-4682-9e04-c3f69f3be74f tempest-ServersWithSpecificFlavorTestJSON-568377525 tempest-ServersWithSpecificFlavorTestJSON-568377525-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 632.676111] env[61957]: DEBUG oslo_concurrency.lockutils [None req-5ca5fdda-dbef-4c48-b38f-9d2aa78e9841 tempest-FloatingIPsAssociationNegativeTestJSON-572957506 tempest-FloatingIPsAssociationNegativeTestJSON-572957506-project-member] Acquiring lock "348ea580-2268-4082-a5f1-33c954063e4d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 632.676294] env[61957]: DEBUG oslo_concurrency.lockutils [None req-5ca5fdda-dbef-4c48-b38f-9d2aa78e9841 tempest-FloatingIPsAssociationNegativeTestJSON-572957506 tempest-FloatingIPsAssociationNegativeTestJSON-572957506-project-member] Lock "348ea580-2268-4082-a5f1-33c954063e4d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 632.840304] env[61957]: DEBUG nova.network.neutron [None req-9c3c6101-e50c-439a-91f6-f293e059d0d2 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] Successfully created port: ea3b3b9c-21bf-4374-a7c5-d040c2ca5f7a {{(pid=61957) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 633.028026] env[61957]: DEBUG nova.compute.manager [None req-9c3c6101-e50c-439a-91f6-f293e059d0d2 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] Start spawning the instance on the hypervisor. {{(pid=61957) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 633.035063] env[61957]: DEBUG oslo_concurrency.lockutils [None req-1b6e3bde-e95e-4682-9e04-c3f69f3be74f tempest-ServersWithSpecificFlavorTestJSON-568377525 tempest-ServersWithSpecificFlavorTestJSON-568377525-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.052s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 633.035063] env[61957]: ERROR nova.compute.manager [None req-1b6e3bde-e95e-4682-9e04-c3f69f3be74f tempest-ServersWithSpecificFlavorTestJSON-568377525 tempest-ServersWithSpecificFlavorTestJSON-568377525-project-member] [instance: f660c725-fc06-42e4-8a34-5d80568b535e] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 511dfb63-c95d-4f13-9a94-1bbc532cbb38, please check neutron logs for more information. [ 633.035063] env[61957]: ERROR nova.compute.manager [instance: f660c725-fc06-42e4-8a34-5d80568b535e] Traceback (most recent call last): [ 633.035063] env[61957]: ERROR nova.compute.manager [instance: f660c725-fc06-42e4-8a34-5d80568b535e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 633.035063] env[61957]: ERROR nova.compute.manager [instance: f660c725-fc06-42e4-8a34-5d80568b535e] self.driver.spawn(context, instance, image_meta, [ 633.035063] env[61957]: ERROR nova.compute.manager [instance: f660c725-fc06-42e4-8a34-5d80568b535e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 633.035063] env[61957]: ERROR nova.compute.manager [instance: f660c725-fc06-42e4-8a34-5d80568b535e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 633.035063] env[61957]: ERROR nova.compute.manager [instance: f660c725-fc06-42e4-8a34-5d80568b535e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 633.035063] env[61957]: ERROR nova.compute.manager [instance: f660c725-fc06-42e4-8a34-5d80568b535e] vm_ref = self.build_virtual_machine(instance, [ 633.035730] env[61957]: ERROR nova.compute.manager [instance: f660c725-fc06-42e4-8a34-5d80568b535e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 633.035730] env[61957]: ERROR nova.compute.manager [instance: f660c725-fc06-42e4-8a34-5d80568b535e] vif_infos = vmwarevif.get_vif_info(self._session, [ 633.035730] env[61957]: ERROR nova.compute.manager [instance: f660c725-fc06-42e4-8a34-5d80568b535e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 633.035730] env[61957]: ERROR nova.compute.manager [instance: f660c725-fc06-42e4-8a34-5d80568b535e] for vif in network_info: [ 633.035730] env[61957]: ERROR nova.compute.manager [instance: f660c725-fc06-42e4-8a34-5d80568b535e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 633.035730] env[61957]: ERROR nova.compute.manager [instance: f660c725-fc06-42e4-8a34-5d80568b535e] return self._sync_wrapper(fn, *args, **kwargs) [ 633.035730] env[61957]: ERROR nova.compute.manager [instance: f660c725-fc06-42e4-8a34-5d80568b535e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 633.035730] env[61957]: ERROR nova.compute.manager [instance: f660c725-fc06-42e4-8a34-5d80568b535e] self.wait() [ 633.035730] env[61957]: ERROR nova.compute.manager [instance: f660c725-fc06-42e4-8a34-5d80568b535e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 633.035730] env[61957]: ERROR nova.compute.manager [instance: f660c725-fc06-42e4-8a34-5d80568b535e] self[:] = self._gt.wait() [ 633.035730] env[61957]: ERROR nova.compute.manager [instance: f660c725-fc06-42e4-8a34-5d80568b535e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 633.035730] env[61957]: ERROR nova.compute.manager [instance: f660c725-fc06-42e4-8a34-5d80568b535e] return self._exit_event.wait() [ 633.035730] env[61957]: ERROR nova.compute.manager [instance: f660c725-fc06-42e4-8a34-5d80568b535e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 633.036119] env[61957]: ERROR nova.compute.manager [instance: f660c725-fc06-42e4-8a34-5d80568b535e] result = hub.switch() [ 633.036119] env[61957]: ERROR nova.compute.manager [instance: f660c725-fc06-42e4-8a34-5d80568b535e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 633.036119] env[61957]: ERROR nova.compute.manager [instance: f660c725-fc06-42e4-8a34-5d80568b535e] return self.greenlet.switch() [ 633.036119] env[61957]: ERROR nova.compute.manager [instance: f660c725-fc06-42e4-8a34-5d80568b535e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 633.036119] env[61957]: ERROR nova.compute.manager [instance: f660c725-fc06-42e4-8a34-5d80568b535e] result = function(*args, **kwargs) [ 633.036119] env[61957]: ERROR nova.compute.manager [instance: f660c725-fc06-42e4-8a34-5d80568b535e] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 633.036119] env[61957]: ERROR nova.compute.manager [instance: f660c725-fc06-42e4-8a34-5d80568b535e] return func(*args, **kwargs) [ 633.036119] env[61957]: ERROR nova.compute.manager [instance: f660c725-fc06-42e4-8a34-5d80568b535e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 633.036119] env[61957]: ERROR nova.compute.manager [instance: f660c725-fc06-42e4-8a34-5d80568b535e] raise e [ 633.036119] env[61957]: ERROR nova.compute.manager [instance: f660c725-fc06-42e4-8a34-5d80568b535e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 633.036119] env[61957]: ERROR nova.compute.manager [instance: f660c725-fc06-42e4-8a34-5d80568b535e] nwinfo = self.network_api.allocate_for_instance( [ 633.036119] env[61957]: ERROR nova.compute.manager [instance: f660c725-fc06-42e4-8a34-5d80568b535e] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 633.036119] env[61957]: ERROR nova.compute.manager [instance: f660c725-fc06-42e4-8a34-5d80568b535e] created_port_ids = self._update_ports_for_instance( [ 633.037556] env[61957]: ERROR nova.compute.manager [instance: f660c725-fc06-42e4-8a34-5d80568b535e] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 633.037556] env[61957]: ERROR nova.compute.manager [instance: f660c725-fc06-42e4-8a34-5d80568b535e] with excutils.save_and_reraise_exception(): [ 633.037556] env[61957]: ERROR nova.compute.manager [instance: f660c725-fc06-42e4-8a34-5d80568b535e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 633.037556] env[61957]: ERROR nova.compute.manager [instance: f660c725-fc06-42e4-8a34-5d80568b535e] self.force_reraise() [ 633.037556] env[61957]: ERROR nova.compute.manager [instance: f660c725-fc06-42e4-8a34-5d80568b535e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 633.037556] env[61957]: ERROR nova.compute.manager [instance: f660c725-fc06-42e4-8a34-5d80568b535e] raise self.value [ 633.037556] env[61957]: ERROR nova.compute.manager [instance: f660c725-fc06-42e4-8a34-5d80568b535e] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 633.037556] env[61957]: ERROR nova.compute.manager [instance: f660c725-fc06-42e4-8a34-5d80568b535e] updated_port = self._update_port( [ 633.037556] env[61957]: ERROR nova.compute.manager [instance: f660c725-fc06-42e4-8a34-5d80568b535e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 633.037556] env[61957]: ERROR nova.compute.manager [instance: f660c725-fc06-42e4-8a34-5d80568b535e] _ensure_no_port_binding_failure(port) [ 633.037556] env[61957]: ERROR nova.compute.manager [instance: f660c725-fc06-42e4-8a34-5d80568b535e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 633.037556] env[61957]: ERROR nova.compute.manager [instance: f660c725-fc06-42e4-8a34-5d80568b535e] raise exception.PortBindingFailed(port_id=port['id']) [ 633.038539] env[61957]: ERROR nova.compute.manager [instance: f660c725-fc06-42e4-8a34-5d80568b535e] nova.exception.PortBindingFailed: Binding failed for port 511dfb63-c95d-4f13-9a94-1bbc532cbb38, please check neutron logs for more information. [ 633.038539] env[61957]: ERROR nova.compute.manager [instance: f660c725-fc06-42e4-8a34-5d80568b535e] [ 633.039299] env[61957]: DEBUG nova.compute.utils [None req-1b6e3bde-e95e-4682-9e04-c3f69f3be74f tempest-ServersWithSpecificFlavorTestJSON-568377525 tempest-ServersWithSpecificFlavorTestJSON-568377525-project-member] [instance: f660c725-fc06-42e4-8a34-5d80568b535e] Binding failed for port 511dfb63-c95d-4f13-9a94-1bbc532cbb38, please check neutron logs for more information. {{(pid=61957) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 633.044353] env[61957]: DEBUG oslo_concurrency.lockutils [None req-3e8259a8-69fb-4692-917a-5e7f03f3483c tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 13.932s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 633.045030] env[61957]: DEBUG nova.objects.instance [None req-3e8259a8-69fb-4692-917a-5e7f03f3483c tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] Lazy-loading 'resources' on Instance uuid c5af147e-3526-4014-98bc-7ad163ae89d9 {{(pid=61957) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 633.046552] env[61957]: DEBUG nova.compute.manager [None req-1b6e3bde-e95e-4682-9e04-c3f69f3be74f tempest-ServersWithSpecificFlavorTestJSON-568377525 tempest-ServersWithSpecificFlavorTestJSON-568377525-project-member] [instance: f660c725-fc06-42e4-8a34-5d80568b535e] Build of instance f660c725-fc06-42e4-8a34-5d80568b535e was re-scheduled: Binding failed for port 511dfb63-c95d-4f13-9a94-1bbc532cbb38, please check neutron logs for more information. {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 633.047852] env[61957]: DEBUG nova.compute.manager [None req-1b6e3bde-e95e-4682-9e04-c3f69f3be74f tempest-ServersWithSpecificFlavorTestJSON-568377525 tempest-ServersWithSpecificFlavorTestJSON-568377525-project-member] [instance: f660c725-fc06-42e4-8a34-5d80568b535e] Unplugging VIFs for instance {{(pid=61957) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 633.049511] env[61957]: DEBUG oslo_concurrency.lockutils [None req-1b6e3bde-e95e-4682-9e04-c3f69f3be74f tempest-ServersWithSpecificFlavorTestJSON-568377525 tempest-ServersWithSpecificFlavorTestJSON-568377525-project-member] Acquiring lock "refresh_cache-f660c725-fc06-42e4-8a34-5d80568b535e" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 633.049511] env[61957]: DEBUG oslo_concurrency.lockutils [None req-1b6e3bde-e95e-4682-9e04-c3f69f3be74f tempest-ServersWithSpecificFlavorTestJSON-568377525 tempest-ServersWithSpecificFlavorTestJSON-568377525-project-member] Acquired lock "refresh_cache-f660c725-fc06-42e4-8a34-5d80568b535e" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 633.049511] env[61957]: DEBUG nova.network.neutron [None req-1b6e3bde-e95e-4682-9e04-c3f69f3be74f tempest-ServersWithSpecificFlavorTestJSON-568377525 tempest-ServersWithSpecificFlavorTestJSON-568377525-project-member] [instance: f660c725-fc06-42e4-8a34-5d80568b535e] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 633.066591] env[61957]: DEBUG nova.virt.hardware [None req-9c3c6101-e50c-439a-91f6-f293e059d0d2 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T17:22:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T17:21:50Z,direct_url=,disk_format='vmdk',id=11c76a2c-f705-470a-ba9d-4657858bab38,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='81d3096d0f094373913fc3dc8f5c3c6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T17:21:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 633.066835] env[61957]: DEBUG nova.virt.hardware [None req-9c3c6101-e50c-439a-91f6-f293e059d0d2 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] Flavor limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 633.066988] env[61957]: DEBUG nova.virt.hardware [None req-9c3c6101-e50c-439a-91f6-f293e059d0d2 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] Image limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 633.067518] env[61957]: DEBUG nova.virt.hardware [None req-9c3c6101-e50c-439a-91f6-f293e059d0d2 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] Flavor pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 633.067518] env[61957]: DEBUG nova.virt.hardware [None req-9c3c6101-e50c-439a-91f6-f293e059d0d2 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] Image pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 633.067518] env[61957]: DEBUG nova.virt.hardware [None req-9c3c6101-e50c-439a-91f6-f293e059d0d2 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 633.067820] env[61957]: DEBUG nova.virt.hardware [None req-9c3c6101-e50c-439a-91f6-f293e059d0d2 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 633.067906] env[61957]: DEBUG nova.virt.hardware [None req-9c3c6101-e50c-439a-91f6-f293e059d0d2 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 633.069279] env[61957]: DEBUG nova.virt.hardware [None req-9c3c6101-e50c-439a-91f6-f293e059d0d2 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] Got 1 possible topologies {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 633.069862] env[61957]: DEBUG nova.virt.hardware [None req-9c3c6101-e50c-439a-91f6-f293e059d0d2 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 633.069862] env[61957]: DEBUG nova.virt.hardware [None req-9c3c6101-e50c-439a-91f6-f293e059d0d2 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 633.071030] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36c35415-dbf8-43bc-b18c-b37f64befc4f {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.082107] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b70b270c-036b-452c-a5a7-592c78421ada {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.652565] env[61957]: DEBUG nova.network.neutron [None req-1b6e3bde-e95e-4682-9e04-c3f69f3be74f tempest-ServersWithSpecificFlavorTestJSON-568377525 tempest-ServersWithSpecificFlavorTestJSON-568377525-project-member] [instance: f660c725-fc06-42e4-8a34-5d80568b535e] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 634.035263] env[61957]: DEBUG nova.network.neutron [None req-1b6e3bde-e95e-4682-9e04-c3f69f3be74f tempest-ServersWithSpecificFlavorTestJSON-568377525 tempest-ServersWithSpecificFlavorTestJSON-568377525-project-member] [instance: f660c725-fc06-42e4-8a34-5d80568b535e] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 634.109245] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9e93daa-b20d-4a3b-a134-0bc86af02550 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.118235] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a3ca61a-5a9b-475d-b341-3d3ec7e5f486 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.154227] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82290b7f-9c24-4249-9efd-cfb8006f2e84 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.162567] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2b03b27-e3d1-42bc-9c74-0b8a0f69d43a {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.179424] env[61957]: DEBUG nova.compute.provider_tree [None req-3e8259a8-69fb-4692-917a-5e7f03f3483c tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 634.188551] env[61957]: DEBUG oslo_concurrency.lockutils [None req-838e8ad7-4b6c-4dbc-8d4b-8380b652e559 tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] Acquiring lock "042425a0-5398-47a6-9842-380d5b3299c0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 634.188775] env[61957]: DEBUG oslo_concurrency.lockutils [None req-838e8ad7-4b6c-4dbc-8d4b-8380b652e559 tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] Lock "042425a0-5398-47a6-9842-380d5b3299c0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 634.193159] env[61957]: DEBUG oslo_concurrency.lockutils [None req-47718518-520f-4177-b41f-8f0cae78b6f6 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Acquiring lock "7f3850c0-a455-4e5b-a329-fb3b02f9725f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 634.193159] env[61957]: DEBUG oslo_concurrency.lockutils [None req-47718518-520f-4177-b41f-8f0cae78b6f6 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Lock "7f3850c0-a455-4e5b-a329-fb3b02f9725f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 634.542303] env[61957]: DEBUG oslo_concurrency.lockutils [None req-1b6e3bde-e95e-4682-9e04-c3f69f3be74f tempest-ServersWithSpecificFlavorTestJSON-568377525 tempest-ServersWithSpecificFlavorTestJSON-568377525-project-member] Releasing lock "refresh_cache-f660c725-fc06-42e4-8a34-5d80568b535e" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 634.542303] env[61957]: DEBUG nova.compute.manager [None req-1b6e3bde-e95e-4682-9e04-c3f69f3be74f tempest-ServersWithSpecificFlavorTestJSON-568377525 tempest-ServersWithSpecificFlavorTestJSON-568377525-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61957) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 634.542303] env[61957]: DEBUG nova.compute.manager [None req-1b6e3bde-e95e-4682-9e04-c3f69f3be74f tempest-ServersWithSpecificFlavorTestJSON-568377525 tempest-ServersWithSpecificFlavorTestJSON-568377525-project-member] [instance: f660c725-fc06-42e4-8a34-5d80568b535e] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 634.542303] env[61957]: DEBUG nova.network.neutron [None req-1b6e3bde-e95e-4682-9e04-c3f69f3be74f tempest-ServersWithSpecificFlavorTestJSON-568377525 tempest-ServersWithSpecificFlavorTestJSON-568377525-project-member] [instance: f660c725-fc06-42e4-8a34-5d80568b535e] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 634.580981] env[61957]: DEBUG nova.network.neutron [None req-1b6e3bde-e95e-4682-9e04-c3f69f3be74f tempest-ServersWithSpecificFlavorTestJSON-568377525 tempest-ServersWithSpecificFlavorTestJSON-568377525-project-member] [instance: f660c725-fc06-42e4-8a34-5d80568b535e] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 634.616447] env[61957]: DEBUG nova.compute.manager [req-7ec4b66a-2002-4858-9f4e-516de8f7e877 req-c8463d0b-306d-4243-9772-b89c3c1c3d84 service nova] [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] Received event network-changed-09ed1ca2-59f3-48fa-b1a0-4d61495c3c4c {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 634.616959] env[61957]: DEBUG nova.compute.manager [req-7ec4b66a-2002-4858-9f4e-516de8f7e877 req-c8463d0b-306d-4243-9772-b89c3c1c3d84 service nova] [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] Refreshing instance network info cache due to event network-changed-09ed1ca2-59f3-48fa-b1a0-4d61495c3c4c. {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 634.616959] env[61957]: DEBUG oslo_concurrency.lockutils [req-7ec4b66a-2002-4858-9f4e-516de8f7e877 req-c8463d0b-306d-4243-9772-b89c3c1c3d84 service nova] Acquiring lock "refresh_cache-763d87bb-652c-43e0-ba39-135bae2cc368" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 634.617126] env[61957]: DEBUG oslo_concurrency.lockutils [req-7ec4b66a-2002-4858-9f4e-516de8f7e877 req-c8463d0b-306d-4243-9772-b89c3c1c3d84 service nova] Acquired lock "refresh_cache-763d87bb-652c-43e0-ba39-135bae2cc368" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 634.617363] env[61957]: DEBUG nova.network.neutron [req-7ec4b66a-2002-4858-9f4e-516de8f7e877 req-c8463d0b-306d-4243-9772-b89c3c1c3d84 service nova] [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] Refreshing network info cache for port 09ed1ca2-59f3-48fa-b1a0-4d61495c3c4c {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 634.684199] env[61957]: DEBUG nova.scheduler.client.report [None req-3e8259a8-69fb-4692-917a-5e7f03f3483c tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 634.691144] env[61957]: ERROR nova.compute.manager [None req-4e7f7a28-6560-4b73-affe-356aa8d11b0b tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 09ed1ca2-59f3-48fa-b1a0-4d61495c3c4c, please check neutron logs for more information. [ 634.691144] env[61957]: ERROR nova.compute.manager Traceback (most recent call last): [ 634.691144] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 634.691144] env[61957]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 634.691144] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 634.691144] env[61957]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 634.691144] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 634.691144] env[61957]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 634.691144] env[61957]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 634.691144] env[61957]: ERROR nova.compute.manager self.force_reraise() [ 634.691144] env[61957]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 634.691144] env[61957]: ERROR nova.compute.manager raise self.value [ 634.691144] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 634.691144] env[61957]: ERROR nova.compute.manager updated_port = self._update_port( [ 634.691144] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 634.691144] env[61957]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 634.691756] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 634.691756] env[61957]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 634.691756] env[61957]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 09ed1ca2-59f3-48fa-b1a0-4d61495c3c4c, please check neutron logs for more information. [ 634.691756] env[61957]: ERROR nova.compute.manager [ 634.691756] env[61957]: Traceback (most recent call last): [ 634.691756] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 634.691756] env[61957]: listener.cb(fileno) [ 634.691756] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 634.691756] env[61957]: result = function(*args, **kwargs) [ 634.691756] env[61957]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 634.691756] env[61957]: return func(*args, **kwargs) [ 634.691756] env[61957]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 634.691756] env[61957]: raise e [ 634.691756] env[61957]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 634.691756] env[61957]: nwinfo = self.network_api.allocate_for_instance( [ 634.691756] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 634.691756] env[61957]: created_port_ids = self._update_ports_for_instance( [ 634.691756] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 634.691756] env[61957]: with excutils.save_and_reraise_exception(): [ 634.691756] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 634.691756] env[61957]: self.force_reraise() [ 634.691756] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 634.691756] env[61957]: raise self.value [ 634.691756] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 634.691756] env[61957]: updated_port = self._update_port( [ 634.691756] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 634.691756] env[61957]: _ensure_no_port_binding_failure(port) [ 634.691756] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 634.691756] env[61957]: raise exception.PortBindingFailed(port_id=port['id']) [ 634.692883] env[61957]: nova.exception.PortBindingFailed: Binding failed for port 09ed1ca2-59f3-48fa-b1a0-4d61495c3c4c, please check neutron logs for more information. [ 634.692883] env[61957]: Removing descriptor: 15 [ 634.692883] env[61957]: ERROR nova.compute.manager [None req-4e7f7a28-6560-4b73-affe-356aa8d11b0b tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 09ed1ca2-59f3-48fa-b1a0-4d61495c3c4c, please check neutron logs for more information. [ 634.692883] env[61957]: ERROR nova.compute.manager [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] Traceback (most recent call last): [ 634.692883] env[61957]: ERROR nova.compute.manager [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 634.692883] env[61957]: ERROR nova.compute.manager [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] yield resources [ 634.692883] env[61957]: ERROR nova.compute.manager [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 634.692883] env[61957]: ERROR nova.compute.manager [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] self.driver.spawn(context, instance, image_meta, [ 634.692883] env[61957]: ERROR nova.compute.manager [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 634.692883] env[61957]: ERROR nova.compute.manager [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] self._vmops.spawn(context, instance, image_meta, injected_files, [ 634.692883] env[61957]: ERROR nova.compute.manager [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 634.692883] env[61957]: ERROR nova.compute.manager [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] vm_ref = self.build_virtual_machine(instance, [ 634.694046] env[61957]: ERROR nova.compute.manager [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 634.694046] env[61957]: ERROR nova.compute.manager [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] vif_infos = vmwarevif.get_vif_info(self._session, [ 634.694046] env[61957]: ERROR nova.compute.manager [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 634.694046] env[61957]: ERROR nova.compute.manager [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] for vif in network_info: [ 634.694046] env[61957]: ERROR nova.compute.manager [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 634.694046] env[61957]: ERROR nova.compute.manager [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] return self._sync_wrapper(fn, *args, **kwargs) [ 634.694046] env[61957]: ERROR nova.compute.manager [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 634.694046] env[61957]: ERROR nova.compute.manager [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] self.wait() [ 634.694046] env[61957]: ERROR nova.compute.manager [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 634.694046] env[61957]: ERROR nova.compute.manager [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] self[:] = self._gt.wait() [ 634.694046] env[61957]: ERROR nova.compute.manager [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 634.694046] env[61957]: ERROR nova.compute.manager [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] return self._exit_event.wait() [ 634.694046] env[61957]: ERROR nova.compute.manager [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 634.698330] env[61957]: ERROR nova.compute.manager [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] result = hub.switch() [ 634.698330] env[61957]: ERROR nova.compute.manager [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 634.698330] env[61957]: ERROR nova.compute.manager [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] return self.greenlet.switch() [ 634.698330] env[61957]: ERROR nova.compute.manager [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 634.698330] env[61957]: ERROR nova.compute.manager [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] result = function(*args, **kwargs) [ 634.698330] env[61957]: ERROR nova.compute.manager [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 634.698330] env[61957]: ERROR nova.compute.manager [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] return func(*args, **kwargs) [ 634.698330] env[61957]: ERROR nova.compute.manager [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 634.698330] env[61957]: ERROR nova.compute.manager [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] raise e [ 634.698330] env[61957]: ERROR nova.compute.manager [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 634.698330] env[61957]: ERROR nova.compute.manager [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] nwinfo = self.network_api.allocate_for_instance( [ 634.698330] env[61957]: ERROR nova.compute.manager [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 634.698330] env[61957]: ERROR nova.compute.manager [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] created_port_ids = self._update_ports_for_instance( [ 634.699233] env[61957]: ERROR nova.compute.manager [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 634.699233] env[61957]: ERROR nova.compute.manager [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] with excutils.save_and_reraise_exception(): [ 634.699233] env[61957]: ERROR nova.compute.manager [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 634.699233] env[61957]: ERROR nova.compute.manager [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] self.force_reraise() [ 634.699233] env[61957]: ERROR nova.compute.manager [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 634.699233] env[61957]: ERROR nova.compute.manager [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] raise self.value [ 634.699233] env[61957]: ERROR nova.compute.manager [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 634.699233] env[61957]: ERROR nova.compute.manager [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] updated_port = self._update_port( [ 634.699233] env[61957]: ERROR nova.compute.manager [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 634.699233] env[61957]: ERROR nova.compute.manager [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] _ensure_no_port_binding_failure(port) [ 634.699233] env[61957]: ERROR nova.compute.manager [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 634.699233] env[61957]: ERROR nova.compute.manager [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] raise exception.PortBindingFailed(port_id=port['id']) [ 634.699618] env[61957]: ERROR nova.compute.manager [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] nova.exception.PortBindingFailed: Binding failed for port 09ed1ca2-59f3-48fa-b1a0-4d61495c3c4c, please check neutron logs for more information. [ 634.699618] env[61957]: ERROR nova.compute.manager [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] [ 634.699618] env[61957]: INFO nova.compute.manager [None req-4e7f7a28-6560-4b73-affe-356aa8d11b0b tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] Terminating instance [ 634.699618] env[61957]: DEBUG oslo_concurrency.lockutils [None req-4e7f7a28-6560-4b73-affe-356aa8d11b0b tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] Acquiring lock "refresh_cache-763d87bb-652c-43e0-ba39-135bae2cc368" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 634.867728] env[61957]: DEBUG oslo_concurrency.lockutils [None req-09f1eb26-3ef8-431d-92bc-9c62b712a196 tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] Acquiring lock "60ec4271-2614-470d-bf40-47b9955f544c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 634.867949] env[61957]: DEBUG oslo_concurrency.lockutils [None req-09f1eb26-3ef8-431d-92bc-9c62b712a196 tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] Lock "60ec4271-2614-470d-bf40-47b9955f544c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 634.946565] env[61957]: DEBUG oslo_concurrency.lockutils [None req-6b91bce8-d0f8-4c40-855e-ed34f3eb7e64 tempest-ServerActionsTestJSON-1567241756 tempest-ServerActionsTestJSON-1567241756-project-member] Acquiring lock "e69a7833-97b1-4063-a6e8-c4558980ee1e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 634.946565] env[61957]: DEBUG oslo_concurrency.lockutils [None req-6b91bce8-d0f8-4c40-855e-ed34f3eb7e64 tempest-ServerActionsTestJSON-1567241756 tempest-ServerActionsTestJSON-1567241756-project-member] Lock "e69a7833-97b1-4063-a6e8-c4558980ee1e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 635.084680] env[61957]: DEBUG nova.network.neutron [None req-1b6e3bde-e95e-4682-9e04-c3f69f3be74f tempest-ServersWithSpecificFlavorTestJSON-568377525 tempest-ServersWithSpecificFlavorTestJSON-568377525-project-member] [instance: f660c725-fc06-42e4-8a34-5d80568b535e] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 635.157523] env[61957]: DEBUG nova.network.neutron [req-7ec4b66a-2002-4858-9f4e-516de8f7e877 req-c8463d0b-306d-4243-9772-b89c3c1c3d84 service nova] [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 635.191558] env[61957]: DEBUG oslo_concurrency.lockutils [None req-3e8259a8-69fb-4692-917a-5e7f03f3483c tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.150s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 635.194339] env[61957]: DEBUG oslo_concurrency.lockutils [None req-af971b81-6348-4338-b0a9-b06675d34db5 tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.025s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 635.195889] env[61957]: INFO nova.compute.claims [None req-af971b81-6348-4338-b0a9-b06675d34db5 tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 635.215207] env[61957]: INFO nova.scheduler.client.report [None req-3e8259a8-69fb-4692-917a-5e7f03f3483c tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] Deleted allocations for instance c5af147e-3526-4014-98bc-7ad163ae89d9 [ 635.325252] env[61957]: DEBUG nova.network.neutron [req-7ec4b66a-2002-4858-9f4e-516de8f7e877 req-c8463d0b-306d-4243-9772-b89c3c1c3d84 service nova] [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 635.589687] env[61957]: INFO nova.compute.manager [None req-1b6e3bde-e95e-4682-9e04-c3f69f3be74f tempest-ServersWithSpecificFlavorTestJSON-568377525 tempest-ServersWithSpecificFlavorTestJSON-568377525-project-member] [instance: f660c725-fc06-42e4-8a34-5d80568b535e] Took 1.05 seconds to deallocate network for instance. [ 635.725773] env[61957]: DEBUG oslo_concurrency.lockutils [None req-3e8259a8-69fb-4692-917a-5e7f03f3483c tempest-ServerDiagnosticsV248Test-1614944228 tempest-ServerDiagnosticsV248Test-1614944228-project-member] Lock "c5af147e-3526-4014-98bc-7ad163ae89d9" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 20.432s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 635.829643] env[61957]: DEBUG oslo_concurrency.lockutils [req-7ec4b66a-2002-4858-9f4e-516de8f7e877 req-c8463d0b-306d-4243-9772-b89c3c1c3d84 service nova] Releasing lock "refresh_cache-763d87bb-652c-43e0-ba39-135bae2cc368" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 635.830853] env[61957]: DEBUG oslo_concurrency.lockutils [None req-4e7f7a28-6560-4b73-affe-356aa8d11b0b tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] Acquired lock "refresh_cache-763d87bb-652c-43e0-ba39-135bae2cc368" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 635.830853] env[61957]: DEBUG nova.network.neutron [None req-4e7f7a28-6560-4b73-affe-356aa8d11b0b tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 636.042135] env[61957]: ERROR nova.compute.manager [None req-9c3c6101-e50c-439a-91f6-f293e059d0d2 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port ea3b3b9c-21bf-4374-a7c5-d040c2ca5f7a, please check neutron logs for more information. [ 636.042135] env[61957]: ERROR nova.compute.manager Traceback (most recent call last): [ 636.042135] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 636.042135] env[61957]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 636.042135] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 636.042135] env[61957]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 636.042135] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 636.042135] env[61957]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 636.042135] env[61957]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 636.042135] env[61957]: ERROR nova.compute.manager self.force_reraise() [ 636.042135] env[61957]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 636.042135] env[61957]: ERROR nova.compute.manager raise self.value [ 636.042135] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 636.042135] env[61957]: ERROR nova.compute.manager updated_port = self._update_port( [ 636.042135] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 636.042135] env[61957]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 636.042633] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 636.042633] env[61957]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 636.042633] env[61957]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port ea3b3b9c-21bf-4374-a7c5-d040c2ca5f7a, please check neutron logs for more information. [ 636.042633] env[61957]: ERROR nova.compute.manager [ 636.046154] env[61957]: Traceback (most recent call last): [ 636.046154] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 636.046154] env[61957]: listener.cb(fileno) [ 636.046154] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 636.046154] env[61957]: result = function(*args, **kwargs) [ 636.046154] env[61957]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 636.046154] env[61957]: return func(*args, **kwargs) [ 636.046154] env[61957]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 636.046154] env[61957]: raise e [ 636.046154] env[61957]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 636.046154] env[61957]: nwinfo = self.network_api.allocate_for_instance( [ 636.046154] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 636.046154] env[61957]: created_port_ids = self._update_ports_for_instance( [ 636.046154] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 636.046154] env[61957]: with excutils.save_and_reraise_exception(): [ 636.046154] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 636.046154] env[61957]: self.force_reraise() [ 636.046154] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 636.046154] env[61957]: raise self.value [ 636.046154] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 636.046154] env[61957]: updated_port = self._update_port( [ 636.046154] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 636.046154] env[61957]: _ensure_no_port_binding_failure(port) [ 636.046154] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 636.046154] env[61957]: raise exception.PortBindingFailed(port_id=port['id']) [ 636.046154] env[61957]: nova.exception.PortBindingFailed: Binding failed for port ea3b3b9c-21bf-4374-a7c5-d040c2ca5f7a, please check neutron logs for more information. [ 636.046154] env[61957]: Removing descriptor: 17 [ 636.047069] env[61957]: ERROR nova.compute.manager [None req-9c3c6101-e50c-439a-91f6-f293e059d0d2 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port ea3b3b9c-21bf-4374-a7c5-d040c2ca5f7a, please check neutron logs for more information. [ 636.047069] env[61957]: ERROR nova.compute.manager [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] Traceback (most recent call last): [ 636.047069] env[61957]: ERROR nova.compute.manager [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 636.047069] env[61957]: ERROR nova.compute.manager [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] yield resources [ 636.047069] env[61957]: ERROR nova.compute.manager [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 636.047069] env[61957]: ERROR nova.compute.manager [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] self.driver.spawn(context, instance, image_meta, [ 636.047069] env[61957]: ERROR nova.compute.manager [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 636.047069] env[61957]: ERROR nova.compute.manager [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 636.047069] env[61957]: ERROR nova.compute.manager [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 636.047069] env[61957]: ERROR nova.compute.manager [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] vm_ref = self.build_virtual_machine(instance, [ 636.047069] env[61957]: ERROR nova.compute.manager [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 636.047375] env[61957]: ERROR nova.compute.manager [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] vif_infos = vmwarevif.get_vif_info(self._session, [ 636.047375] env[61957]: ERROR nova.compute.manager [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 636.047375] env[61957]: ERROR nova.compute.manager [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] for vif in network_info: [ 636.047375] env[61957]: ERROR nova.compute.manager [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 636.047375] env[61957]: ERROR nova.compute.manager [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] return self._sync_wrapper(fn, *args, **kwargs) [ 636.047375] env[61957]: ERROR nova.compute.manager [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 636.047375] env[61957]: ERROR nova.compute.manager [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] self.wait() [ 636.047375] env[61957]: ERROR nova.compute.manager [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 636.047375] env[61957]: ERROR nova.compute.manager [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] self[:] = self._gt.wait() [ 636.047375] env[61957]: ERROR nova.compute.manager [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 636.047375] env[61957]: ERROR nova.compute.manager [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] return self._exit_event.wait() [ 636.047375] env[61957]: ERROR nova.compute.manager [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 636.047375] env[61957]: ERROR nova.compute.manager [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] result = hub.switch() [ 636.047782] env[61957]: ERROR nova.compute.manager [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 636.047782] env[61957]: ERROR nova.compute.manager [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] return self.greenlet.switch() [ 636.047782] env[61957]: ERROR nova.compute.manager [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 636.047782] env[61957]: ERROR nova.compute.manager [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] result = function(*args, **kwargs) [ 636.047782] env[61957]: ERROR nova.compute.manager [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 636.047782] env[61957]: ERROR nova.compute.manager [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] return func(*args, **kwargs) [ 636.047782] env[61957]: ERROR nova.compute.manager [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 636.047782] env[61957]: ERROR nova.compute.manager [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] raise e [ 636.047782] env[61957]: ERROR nova.compute.manager [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 636.047782] env[61957]: ERROR nova.compute.manager [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] nwinfo = self.network_api.allocate_for_instance( [ 636.047782] env[61957]: ERROR nova.compute.manager [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 636.047782] env[61957]: ERROR nova.compute.manager [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] created_port_ids = self._update_ports_for_instance( [ 636.047782] env[61957]: ERROR nova.compute.manager [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 636.048207] env[61957]: ERROR nova.compute.manager [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] with excutils.save_and_reraise_exception(): [ 636.048207] env[61957]: ERROR nova.compute.manager [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 636.048207] env[61957]: ERROR nova.compute.manager [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] self.force_reraise() [ 636.048207] env[61957]: ERROR nova.compute.manager [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 636.048207] env[61957]: ERROR nova.compute.manager [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] raise self.value [ 636.048207] env[61957]: ERROR nova.compute.manager [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 636.048207] env[61957]: ERROR nova.compute.manager [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] updated_port = self._update_port( [ 636.048207] env[61957]: ERROR nova.compute.manager [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 636.048207] env[61957]: ERROR nova.compute.manager [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] _ensure_no_port_binding_failure(port) [ 636.048207] env[61957]: ERROR nova.compute.manager [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 636.048207] env[61957]: ERROR nova.compute.manager [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] raise exception.PortBindingFailed(port_id=port['id']) [ 636.048207] env[61957]: ERROR nova.compute.manager [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] nova.exception.PortBindingFailed: Binding failed for port ea3b3b9c-21bf-4374-a7c5-d040c2ca5f7a, please check neutron logs for more information. [ 636.048207] env[61957]: ERROR nova.compute.manager [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] [ 636.048666] env[61957]: INFO nova.compute.manager [None req-9c3c6101-e50c-439a-91f6-f293e059d0d2 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] Terminating instance [ 636.051580] env[61957]: DEBUG oslo_concurrency.lockutils [None req-9c3c6101-e50c-439a-91f6-f293e059d0d2 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] Acquiring lock "refresh_cache-4be3eabd-4e58-48b8-b14a-798ba419655e" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 636.051822] env[61957]: DEBUG oslo_concurrency.lockutils [None req-9c3c6101-e50c-439a-91f6-f293e059d0d2 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] Acquired lock "refresh_cache-4be3eabd-4e58-48b8-b14a-798ba419655e" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 636.051915] env[61957]: DEBUG nova.network.neutron [None req-9c3c6101-e50c-439a-91f6-f293e059d0d2 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 636.351387] env[61957]: DEBUG nova.network.neutron [None req-4e7f7a28-6560-4b73-affe-356aa8d11b0b tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 636.432860] env[61957]: DEBUG nova.network.neutron [None req-4e7f7a28-6560-4b73-affe-356aa8d11b0b tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 636.548727] env[61957]: DEBUG oslo_service.periodic_task [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61957) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 636.548939] env[61957]: DEBUG oslo_service.periodic_task [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61957) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 636.609585] env[61957]: DEBUG nova.network.neutron [None req-9c3c6101-e50c-439a-91f6-f293e059d0d2 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 636.633026] env[61957]: INFO nova.scheduler.client.report [None req-1b6e3bde-e95e-4682-9e04-c3f69f3be74f tempest-ServersWithSpecificFlavorTestJSON-568377525 tempest-ServersWithSpecificFlavorTestJSON-568377525-project-member] Deleted allocations for instance f660c725-fc06-42e4-8a34-5d80568b535e [ 636.737555] env[61957]: DEBUG nova.network.neutron [None req-9c3c6101-e50c-439a-91f6-f293e059d0d2 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 636.769019] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87ba1147-cb67-48aa-82ce-5cb0ef98ca10 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.776094] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55508349-a73d-497b-a644-8e46530ff81c {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.811790] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e22b541-4f6a-422d-a43d-2f7088bb68ac {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.815984] env[61957]: DEBUG nova.compute.manager [req-7a481e1c-dfe4-443a-89bb-a0788372cda9 req-a84e1fa2-c027-4eb5-a63c-9076544aae5e service nova] [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] Received event network-vif-deleted-09ed1ca2-59f3-48fa-b1a0-4d61495c3c4c {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 636.816340] env[61957]: DEBUG nova.compute.manager [req-7a481e1c-dfe4-443a-89bb-a0788372cda9 req-a84e1fa2-c027-4eb5-a63c-9076544aae5e service nova] [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] Received event network-changed-ea3b3b9c-21bf-4374-a7c5-d040c2ca5f7a {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 636.816877] env[61957]: DEBUG nova.compute.manager [req-7a481e1c-dfe4-443a-89bb-a0788372cda9 req-a84e1fa2-c027-4eb5-a63c-9076544aae5e service nova] [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] Refreshing instance network info cache due to event network-changed-ea3b3b9c-21bf-4374-a7c5-d040c2ca5f7a. {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 636.817490] env[61957]: DEBUG oslo_concurrency.lockutils [req-7a481e1c-dfe4-443a-89bb-a0788372cda9 req-a84e1fa2-c027-4eb5-a63c-9076544aae5e service nova] Acquiring lock "refresh_cache-4be3eabd-4e58-48b8-b14a-798ba419655e" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 636.824620] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f909a4b6-3fce-416f-80d3-15992174399c {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.839340] env[61957]: DEBUG nova.compute.provider_tree [None req-af971b81-6348-4338-b0a9-b06675d34db5 tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 636.935417] env[61957]: DEBUG oslo_concurrency.lockutils [None req-4e7f7a28-6560-4b73-affe-356aa8d11b0b tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] Releasing lock "refresh_cache-763d87bb-652c-43e0-ba39-135bae2cc368" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 636.935883] env[61957]: DEBUG nova.compute.manager [None req-4e7f7a28-6560-4b73-affe-356aa8d11b0b tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] Start destroying the instance on the hypervisor. {{(pid=61957) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 636.936055] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-4e7f7a28-6560-4b73-affe-356aa8d11b0b tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] Destroying instance {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 636.936366] env[61957]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c9f445f9-d13f-4d05-aeb7-6966b09a95e6 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.946483] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4a0bd64-fc1a-4fdc-8957-d869ea068150 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.971969] env[61957]: WARNING nova.virt.vmwareapi.vmops [None req-4e7f7a28-6560-4b73-affe-356aa8d11b0b tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 763d87bb-652c-43e0-ba39-135bae2cc368 could not be found. [ 636.972235] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-4e7f7a28-6560-4b73-affe-356aa8d11b0b tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] Instance destroyed {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 636.972494] env[61957]: INFO nova.compute.manager [None req-4e7f7a28-6560-4b73-affe-356aa8d11b0b tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] Took 0.04 seconds to destroy the instance on the hypervisor. [ 636.976016] env[61957]: DEBUG oslo.service.loopingcall [None req-4e7f7a28-6560-4b73-affe-356aa8d11b0b tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 636.976016] env[61957]: DEBUG nova.compute.manager [-] [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 636.976016] env[61957]: DEBUG nova.network.neutron [-] [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 637.032318] env[61957]: DEBUG nova.network.neutron [-] [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 637.060522] env[61957]: DEBUG oslo_service.periodic_task [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61957) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 637.060706] env[61957]: DEBUG nova.compute.manager [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Starting heal instance info cache {{(pid=61957) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 637.060817] env[61957]: DEBUG nova.compute.manager [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Rebuilding the list of instances to heal {{(pid=61957) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9930}} [ 637.150136] env[61957]: DEBUG oslo_concurrency.lockutils [None req-1b6e3bde-e95e-4682-9e04-c3f69f3be74f tempest-ServersWithSpecificFlavorTestJSON-568377525 tempest-ServersWithSpecificFlavorTestJSON-568377525-project-member] Lock "f660c725-fc06-42e4-8a34-5d80568b535e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 54.980s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 637.238269] env[61957]: DEBUG oslo_concurrency.lockutils [None req-9c3c6101-e50c-439a-91f6-f293e059d0d2 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] Releasing lock "refresh_cache-4be3eabd-4e58-48b8-b14a-798ba419655e" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 637.239360] env[61957]: DEBUG nova.compute.manager [None req-9c3c6101-e50c-439a-91f6-f293e059d0d2 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] Start destroying the instance on the hypervisor. {{(pid=61957) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 637.239483] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-9c3c6101-e50c-439a-91f6-f293e059d0d2 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] Destroying instance {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 637.239895] env[61957]: DEBUG oslo_concurrency.lockutils [req-7a481e1c-dfe4-443a-89bb-a0788372cda9 req-a84e1fa2-c027-4eb5-a63c-9076544aae5e service nova] Acquired lock "refresh_cache-4be3eabd-4e58-48b8-b14a-798ba419655e" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 637.239895] env[61957]: DEBUG nova.network.neutron [req-7a481e1c-dfe4-443a-89bb-a0788372cda9 req-a84e1fa2-c027-4eb5-a63c-9076544aae5e service nova] [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] Refreshing network info cache for port ea3b3b9c-21bf-4374-a7c5-d040c2ca5f7a {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 637.241010] env[61957]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-63dd7004-abda-4fce-888e-9709bd372dbd {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.250214] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee74d7b4-492e-44c0-9f72-bed41d243453 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.276877] env[61957]: WARNING nova.virt.vmwareapi.vmops [None req-9c3c6101-e50c-439a-91f6-f293e059d0d2 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 4be3eabd-4e58-48b8-b14a-798ba419655e could not be found. [ 637.277127] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-9c3c6101-e50c-439a-91f6-f293e059d0d2 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] Instance destroyed {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 637.277305] env[61957]: INFO nova.compute.manager [None req-9c3c6101-e50c-439a-91f6-f293e059d0d2 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] Took 0.04 seconds to destroy the instance on the hypervisor. [ 637.277565] env[61957]: DEBUG oslo.service.loopingcall [None req-9c3c6101-e50c-439a-91f6-f293e059d0d2 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 637.277783] env[61957]: DEBUG nova.compute.manager [-] [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 637.278132] env[61957]: DEBUG nova.network.neutron [-] [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 637.301777] env[61957]: DEBUG nova.network.neutron [-] [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 637.346114] env[61957]: DEBUG nova.scheduler.client.report [None req-af971b81-6348-4338-b0a9-b06675d34db5 tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 637.535024] env[61957]: DEBUG nova.network.neutron [-] [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 637.564694] env[61957]: DEBUG nova.compute.manager [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] Skipping network cache update for instance because it is Building. {{(pid=61957) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 637.564901] env[61957]: DEBUG nova.compute.manager [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] Skipping network cache update for instance because it is Building. {{(pid=61957) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 637.565041] env[61957]: DEBUG nova.compute.manager [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] [instance: 3c099888-d5dc-4cee-9729-311808694625] Skipping network cache update for instance because it is Building. {{(pid=61957) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 637.565173] env[61957]: DEBUG nova.compute.manager [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] Skipping network cache update for instance because it is Building. {{(pid=61957) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 637.565290] env[61957]: DEBUG nova.compute.manager [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] Skipping network cache update for instance because it is Building. {{(pid=61957) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 637.565936] env[61957]: DEBUG nova.compute.manager [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] Skipping network cache update for instance because it is Building. {{(pid=61957) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 637.602421] env[61957]: DEBUG oslo_concurrency.lockutils [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Acquiring lock "refresh_cache-4ad746d0-4076-4c97-8ea9-20e49fda59ca" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 637.604565] env[61957]: DEBUG oslo_concurrency.lockutils [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Acquired lock "refresh_cache-4ad746d0-4076-4c97-8ea9-20e49fda59ca" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 637.604565] env[61957]: DEBUG nova.network.neutron [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] [instance: 4ad746d0-4076-4c97-8ea9-20e49fda59ca] Forcefully refreshing network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 637.604565] env[61957]: DEBUG nova.objects.instance [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Lazy-loading 'info_cache' on Instance uuid 4ad746d0-4076-4c97-8ea9-20e49fda59ca {{(pid=61957) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 637.650024] env[61957]: DEBUG oslo_concurrency.lockutils [None req-660095f0-9d5a-4797-b621-e0585ffddbba tempest-ServerActionsTestOtherB-303558607 tempest-ServerActionsTestOtherB-303558607-project-member] Acquiring lock "d77276ba-6d37-440b-a113-c4f62b2e7946" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 637.650254] env[61957]: DEBUG oslo_concurrency.lockutils [None req-660095f0-9d5a-4797-b621-e0585ffddbba tempest-ServerActionsTestOtherB-303558607 tempest-ServerActionsTestOtherB-303558607-project-member] Lock "d77276ba-6d37-440b-a113-c4f62b2e7946" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 637.652922] env[61957]: DEBUG nova.compute.manager [None req-b1f62379-e574-4175-a099-f938af5ddaa8 tempest-ServerPasswordTestJSON-1543821420 tempest-ServerPasswordTestJSON-1543821420-project-member] [instance: 49e99297-ffb9-4104-bd06-911243908828] Starting instance... {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 637.761925] env[61957]: DEBUG nova.network.neutron [req-7a481e1c-dfe4-443a-89bb-a0788372cda9 req-a84e1fa2-c027-4eb5-a63c-9076544aae5e service nova] [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 637.805582] env[61957]: DEBUG nova.network.neutron [-] [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 637.851149] env[61957]: DEBUG nova.network.neutron [req-7a481e1c-dfe4-443a-89bb-a0788372cda9 req-a84e1fa2-c027-4eb5-a63c-9076544aae5e service nova] [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 637.852751] env[61957]: DEBUG oslo_concurrency.lockutils [None req-af971b81-6348-4338-b0a9-b06675d34db5 tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.658s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 637.853323] env[61957]: DEBUG nova.compute.manager [None req-af971b81-6348-4338-b0a9-b06675d34db5 tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] Start building networks asynchronously for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 637.856273] env[61957]: DEBUG oslo_concurrency.lockutils [None req-42e67e16-6377-41f9-9501-426c4335c678 tempest-ServersTestBootFromVolume-482289819 tempest-ServersTestBootFromVolume-482289819-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.678s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 638.039958] env[61957]: INFO nova.compute.manager [-] [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] Took 1.07 seconds to deallocate network for instance. [ 638.041935] env[61957]: DEBUG nova.compute.claims [None req-4e7f7a28-6560-4b73-affe-356aa8d11b0b tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] Aborting claim: {{(pid=61957) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 638.042132] env[61957]: DEBUG oslo_concurrency.lockutils [None req-4e7f7a28-6560-4b73-affe-356aa8d11b0b tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 638.182295] env[61957]: DEBUG oslo_concurrency.lockutils [None req-b1f62379-e574-4175-a099-f938af5ddaa8 tempest-ServerPasswordTestJSON-1543821420 tempest-ServerPasswordTestJSON-1543821420-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 638.309214] env[61957]: INFO nova.compute.manager [-] [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] Took 1.03 seconds to deallocate network for instance. [ 638.313764] env[61957]: DEBUG nova.compute.claims [None req-9c3c6101-e50c-439a-91f6-f293e059d0d2 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] Aborting claim: {{(pid=61957) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 638.313861] env[61957]: DEBUG oslo_concurrency.lockutils [None req-9c3c6101-e50c-439a-91f6-f293e059d0d2 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 638.357624] env[61957]: DEBUG nova.compute.utils [None req-af971b81-6348-4338-b0a9-b06675d34db5 tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] Using /dev/sd instead of None {{(pid=61957) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 638.362022] env[61957]: DEBUG oslo_concurrency.lockutils [req-7a481e1c-dfe4-443a-89bb-a0788372cda9 req-a84e1fa2-c027-4eb5-a63c-9076544aae5e service nova] Releasing lock "refresh_cache-4be3eabd-4e58-48b8-b14a-798ba419655e" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 638.362022] env[61957]: DEBUG nova.compute.manager [req-7a481e1c-dfe4-443a-89bb-a0788372cda9 req-a84e1fa2-c027-4eb5-a63c-9076544aae5e service nova] [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] Received event network-vif-deleted-ea3b3b9c-21bf-4374-a7c5-d040c2ca5f7a {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 638.362022] env[61957]: DEBUG nova.compute.manager [None req-af971b81-6348-4338-b0a9-b06675d34db5 tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] Allocating IP information in the background. {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 638.362022] env[61957]: DEBUG nova.network.neutron [None req-af971b81-6348-4338-b0a9-b06675d34db5 tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] allocate_for_instance() {{(pid=61957) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 638.462472] env[61957]: DEBUG nova.policy [None req-af971b81-6348-4338-b0a9-b06675d34db5 tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2414df50921145e98bb4e5cade9c724b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8363e21fa47b46b8b2538b741564cb60', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61957) authorize /opt/stack/nova/nova/policy.py:203}} [ 638.647733] env[61957]: DEBUG nova.network.neutron [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] [instance: 4ad746d0-4076-4c97-8ea9-20e49fda59ca] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 638.863448] env[61957]: DEBUG nova.compute.manager [None req-af971b81-6348-4338-b0a9-b06675d34db5 tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] Start building block device mappings for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 638.883496] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d748a032-ef66-4d0b-815e-14efd9aa6edb {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.891936] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0a2be1c-f5c2-4a5d-b6d1-15d181483e5a {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.930370] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-887a627f-08c0-4166-865d-3cd6e33fc200 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.938772] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae2b596e-096e-4ceb-b285-1cddc14d8d9a {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.954947] env[61957]: DEBUG nova.compute.provider_tree [None req-42e67e16-6377-41f9-9501-426c4335c678 tempest-ServersTestBootFromVolume-482289819 tempest-ServersTestBootFromVolume-482289819-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 639.181090] env[61957]: DEBUG nova.network.neutron [None req-af971b81-6348-4338-b0a9-b06675d34db5 tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] Successfully created port: b29b2b7c-e839-4ae7-a55b-fc34bee248b2 {{(pid=61957) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 639.298520] env[61957]: DEBUG nova.network.neutron [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] [instance: 4ad746d0-4076-4c97-8ea9-20e49fda59ca] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 639.457479] env[61957]: DEBUG nova.scheduler.client.report [None req-42e67e16-6377-41f9-9501-426c4335c678 tempest-ServersTestBootFromVolume-482289819 tempest-ServersTestBootFromVolume-482289819-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 639.805859] env[61957]: DEBUG oslo_concurrency.lockutils [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Releasing lock "refresh_cache-4ad746d0-4076-4c97-8ea9-20e49fda59ca" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 639.805859] env[61957]: DEBUG nova.compute.manager [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] [instance: 4ad746d0-4076-4c97-8ea9-20e49fda59ca] Updated the network info_cache for instance {{(pid=61957) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9997}} [ 639.805859] env[61957]: DEBUG oslo_service.periodic_task [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61957) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 639.805859] env[61957]: DEBUG oslo_service.periodic_task [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61957) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 639.806302] env[61957]: DEBUG oslo_service.periodic_task [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61957) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 639.806302] env[61957]: DEBUG oslo_service.periodic_task [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61957) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 639.806302] env[61957]: DEBUG oslo_service.periodic_task [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61957) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 639.806423] env[61957]: DEBUG oslo_service.periodic_task [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61957) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 639.806541] env[61957]: DEBUG nova.compute.manager [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61957) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 639.806669] env[61957]: DEBUG oslo_service.periodic_task [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61957) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 639.874229] env[61957]: DEBUG nova.compute.manager [None req-af971b81-6348-4338-b0a9-b06675d34db5 tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] Start spawning the instance on the hypervisor. {{(pid=61957) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 639.900484] env[61957]: DEBUG nova.virt.hardware [None req-af971b81-6348-4338-b0a9-b06675d34db5 tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T17:22:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T17:21:50Z,direct_url=,disk_format='vmdk',id=11c76a2c-f705-470a-ba9d-4657858bab38,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='81d3096d0f094373913fc3dc8f5c3c6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T17:21:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 639.900918] env[61957]: DEBUG nova.virt.hardware [None req-af971b81-6348-4338-b0a9-b06675d34db5 tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] Flavor limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 639.901105] env[61957]: DEBUG nova.virt.hardware [None req-af971b81-6348-4338-b0a9-b06675d34db5 tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] Image limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 639.901296] env[61957]: DEBUG nova.virt.hardware [None req-af971b81-6348-4338-b0a9-b06675d34db5 tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] Flavor pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 639.901439] env[61957]: DEBUG nova.virt.hardware [None req-af971b81-6348-4338-b0a9-b06675d34db5 tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] Image pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 639.901583] env[61957]: DEBUG nova.virt.hardware [None req-af971b81-6348-4338-b0a9-b06675d34db5 tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 639.901790] env[61957]: DEBUG nova.virt.hardware [None req-af971b81-6348-4338-b0a9-b06675d34db5 tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 639.901944] env[61957]: DEBUG nova.virt.hardware [None req-af971b81-6348-4338-b0a9-b06675d34db5 tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 639.902125] env[61957]: DEBUG nova.virt.hardware [None req-af971b81-6348-4338-b0a9-b06675d34db5 tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] Got 1 possible topologies {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 639.902288] env[61957]: DEBUG nova.virt.hardware [None req-af971b81-6348-4338-b0a9-b06675d34db5 tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 639.902456] env[61957]: DEBUG nova.virt.hardware [None req-af971b81-6348-4338-b0a9-b06675d34db5 tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 639.903355] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9216244f-06eb-41a7-8a79-84404535c796 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.913178] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d25805c-2857-4a16-8161-9cbe52171f3e {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.962892] env[61957]: DEBUG oslo_concurrency.lockutils [None req-42e67e16-6377-41f9-9501-426c4335c678 tempest-ServersTestBootFromVolume-482289819 tempest-ServersTestBootFromVolume-482289819-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.106s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 639.963536] env[61957]: ERROR nova.compute.manager [None req-42e67e16-6377-41f9-9501-426c4335c678 tempest-ServersTestBootFromVolume-482289819 tempest-ServersTestBootFromVolume-482289819-project-member] [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port bb5c53cf-6a2a-4efe-9be1-07da3dd400d6, please check neutron logs for more information. [ 639.963536] env[61957]: ERROR nova.compute.manager [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] Traceback (most recent call last): [ 639.963536] env[61957]: ERROR nova.compute.manager [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 639.963536] env[61957]: ERROR nova.compute.manager [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] self.driver.spawn(context, instance, image_meta, [ 639.963536] env[61957]: ERROR nova.compute.manager [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 639.963536] env[61957]: ERROR nova.compute.manager [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 639.963536] env[61957]: ERROR nova.compute.manager [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 639.963536] env[61957]: ERROR nova.compute.manager [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] vm_ref = self.build_virtual_machine(instance, [ 639.963536] env[61957]: ERROR nova.compute.manager [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 639.963536] env[61957]: ERROR nova.compute.manager [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] vif_infos = vmwarevif.get_vif_info(self._session, [ 639.963536] env[61957]: ERROR nova.compute.manager [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 639.963827] env[61957]: ERROR nova.compute.manager [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] for vif in network_info: [ 639.963827] env[61957]: ERROR nova.compute.manager [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 639.963827] env[61957]: ERROR nova.compute.manager [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] return self._sync_wrapper(fn, *args, **kwargs) [ 639.963827] env[61957]: ERROR nova.compute.manager [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 639.963827] env[61957]: ERROR nova.compute.manager [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] self.wait() [ 639.963827] env[61957]: ERROR nova.compute.manager [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 639.963827] env[61957]: ERROR nova.compute.manager [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] self[:] = self._gt.wait() [ 639.963827] env[61957]: ERROR nova.compute.manager [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 639.963827] env[61957]: ERROR nova.compute.manager [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] return self._exit_event.wait() [ 639.963827] env[61957]: ERROR nova.compute.manager [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 639.963827] env[61957]: ERROR nova.compute.manager [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] current.throw(*self._exc) [ 639.963827] env[61957]: ERROR nova.compute.manager [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 639.963827] env[61957]: ERROR nova.compute.manager [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] result = function(*args, **kwargs) [ 639.964124] env[61957]: ERROR nova.compute.manager [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 639.964124] env[61957]: ERROR nova.compute.manager [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] return func(*args, **kwargs) [ 639.964124] env[61957]: ERROR nova.compute.manager [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 639.964124] env[61957]: ERROR nova.compute.manager [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] raise e [ 639.964124] env[61957]: ERROR nova.compute.manager [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 639.964124] env[61957]: ERROR nova.compute.manager [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] nwinfo = self.network_api.allocate_for_instance( [ 639.964124] env[61957]: ERROR nova.compute.manager [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 639.964124] env[61957]: ERROR nova.compute.manager [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] created_port_ids = self._update_ports_for_instance( [ 639.964124] env[61957]: ERROR nova.compute.manager [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 639.964124] env[61957]: ERROR nova.compute.manager [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] with excutils.save_and_reraise_exception(): [ 639.964124] env[61957]: ERROR nova.compute.manager [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 639.964124] env[61957]: ERROR nova.compute.manager [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] self.force_reraise() [ 639.964124] env[61957]: ERROR nova.compute.manager [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 639.964420] env[61957]: ERROR nova.compute.manager [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] raise self.value [ 639.964420] env[61957]: ERROR nova.compute.manager [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 639.964420] env[61957]: ERROR nova.compute.manager [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] updated_port = self._update_port( [ 639.964420] env[61957]: ERROR nova.compute.manager [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 639.964420] env[61957]: ERROR nova.compute.manager [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] _ensure_no_port_binding_failure(port) [ 639.964420] env[61957]: ERROR nova.compute.manager [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 639.964420] env[61957]: ERROR nova.compute.manager [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] raise exception.PortBindingFailed(port_id=port['id']) [ 639.964420] env[61957]: ERROR nova.compute.manager [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] nova.exception.PortBindingFailed: Binding failed for port bb5c53cf-6a2a-4efe-9be1-07da3dd400d6, please check neutron logs for more information. [ 639.964420] env[61957]: ERROR nova.compute.manager [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] [ 639.964420] env[61957]: DEBUG nova.compute.utils [None req-42e67e16-6377-41f9-9501-426c4335c678 tempest-ServersTestBootFromVolume-482289819 tempest-ServersTestBootFromVolume-482289819-project-member] [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] Binding failed for port bb5c53cf-6a2a-4efe-9be1-07da3dd400d6, please check neutron logs for more information. {{(pid=61957) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 639.965650] env[61957]: DEBUG oslo_concurrency.lockutils [None req-260de5ae-9540-404d-a23e-9facd81096ee tempest-AttachInterfacesUnderV243Test-1229304814 tempest-AttachInterfacesUnderV243Test-1229304814-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.788s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 639.968880] env[61957]: DEBUG nova.compute.manager [None req-42e67e16-6377-41f9-9501-426c4335c678 tempest-ServersTestBootFromVolume-482289819 tempest-ServersTestBootFromVolume-482289819-project-member] [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] Build of instance 32e56693-5100-4ed8-b26a-7d377f9ab7b6 was re-scheduled: Binding failed for port bb5c53cf-6a2a-4efe-9be1-07da3dd400d6, please check neutron logs for more information. {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 639.969318] env[61957]: DEBUG nova.compute.manager [None req-42e67e16-6377-41f9-9501-426c4335c678 tempest-ServersTestBootFromVolume-482289819 tempest-ServersTestBootFromVolume-482289819-project-member] [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] Unplugging VIFs for instance {{(pid=61957) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 639.969538] env[61957]: DEBUG oslo_concurrency.lockutils [None req-42e67e16-6377-41f9-9501-426c4335c678 tempest-ServersTestBootFromVolume-482289819 tempest-ServersTestBootFromVolume-482289819-project-member] Acquiring lock "refresh_cache-32e56693-5100-4ed8-b26a-7d377f9ab7b6" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 639.969681] env[61957]: DEBUG oslo_concurrency.lockutils [None req-42e67e16-6377-41f9-9501-426c4335c678 tempest-ServersTestBootFromVolume-482289819 tempest-ServersTestBootFromVolume-482289819-project-member] Acquired lock "refresh_cache-32e56693-5100-4ed8-b26a-7d377f9ab7b6" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 639.969856] env[61957]: DEBUG nova.network.neutron [None req-42e67e16-6377-41f9-9501-426c4335c678 tempest-ServersTestBootFromVolume-482289819 tempest-ServersTestBootFromVolume-482289819-project-member] [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 640.019943] env[61957]: DEBUG nova.network.neutron [None req-af971b81-6348-4338-b0a9-b06675d34db5 tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] Successfully created port: f8773edc-0900-4504-b052-14b135a25054 {{(pid=61957) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 640.310464] env[61957]: DEBUG oslo_concurrency.lockutils [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 640.505462] env[61957]: DEBUG nova.network.neutron [None req-42e67e16-6377-41f9-9501-426c4335c678 tempest-ServersTestBootFromVolume-482289819 tempest-ServersTestBootFromVolume-482289819-project-member] [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 640.651456] env[61957]: DEBUG nova.network.neutron [None req-af971b81-6348-4338-b0a9-b06675d34db5 tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] Successfully created port: 16b24aba-ccce-4f7b-b5ae-7dc3cc6a8974 {{(pid=61957) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 640.716345] env[61957]: DEBUG nova.network.neutron [None req-42e67e16-6377-41f9-9501-426c4335c678 tempest-ServersTestBootFromVolume-482289819 tempest-ServersTestBootFromVolume-482289819-project-member] [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 641.025408] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74423de5-f105-49ab-b03c-920f51b781b4 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.037172] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54ddc3fd-51a4-49bb-9a9f-644ff719c6d0 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.072875] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba956877-63ee-4c04-a04c-b2c6471c7b7f {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.080334] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d206bb93-ccbc-4db7-a63b-b87414707f9c {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.094102] env[61957]: DEBUG nova.compute.provider_tree [None req-260de5ae-9540-404d-a23e-9facd81096ee tempest-AttachInterfacesUnderV243Test-1229304814 tempest-AttachInterfacesUnderV243Test-1229304814-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 641.223142] env[61957]: DEBUG oslo_concurrency.lockutils [None req-42e67e16-6377-41f9-9501-426c4335c678 tempest-ServersTestBootFromVolume-482289819 tempest-ServersTestBootFromVolume-482289819-project-member] Releasing lock "refresh_cache-32e56693-5100-4ed8-b26a-7d377f9ab7b6" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 641.223394] env[61957]: DEBUG nova.compute.manager [None req-42e67e16-6377-41f9-9501-426c4335c678 tempest-ServersTestBootFromVolume-482289819 tempest-ServersTestBootFromVolume-482289819-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61957) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 641.223623] env[61957]: DEBUG nova.compute.manager [None req-42e67e16-6377-41f9-9501-426c4335c678 tempest-ServersTestBootFromVolume-482289819 tempest-ServersTestBootFromVolume-482289819-project-member] [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 641.223745] env[61957]: DEBUG nova.network.neutron [None req-42e67e16-6377-41f9-9501-426c4335c678 tempest-ServersTestBootFromVolume-482289819 tempest-ServersTestBootFromVolume-482289819-project-member] [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 641.263556] env[61957]: DEBUG nova.network.neutron [None req-42e67e16-6377-41f9-9501-426c4335c678 tempest-ServersTestBootFromVolume-482289819 tempest-ServersTestBootFromVolume-482289819-project-member] [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 641.601414] env[61957]: DEBUG nova.scheduler.client.report [None req-260de5ae-9540-404d-a23e-9facd81096ee tempest-AttachInterfacesUnderV243Test-1229304814 tempest-AttachInterfacesUnderV243Test-1229304814-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 641.765867] env[61957]: DEBUG nova.network.neutron [None req-42e67e16-6377-41f9-9501-426c4335c678 tempest-ServersTestBootFromVolume-482289819 tempest-ServersTestBootFromVolume-482289819-project-member] [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 642.108375] env[61957]: DEBUG oslo_concurrency.lockutils [None req-260de5ae-9540-404d-a23e-9facd81096ee tempest-AttachInterfacesUnderV243Test-1229304814 tempest-AttachInterfacesUnderV243Test-1229304814-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.143s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 642.109274] env[61957]: ERROR nova.compute.manager [None req-260de5ae-9540-404d-a23e-9facd81096ee tempest-AttachInterfacesUnderV243Test-1229304814 tempest-AttachInterfacesUnderV243Test-1229304814-project-member] [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 3e020214-1678-4e89-930e-bb37e8f3e31b, please check neutron logs for more information. [ 642.109274] env[61957]: ERROR nova.compute.manager [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] Traceback (most recent call last): [ 642.109274] env[61957]: ERROR nova.compute.manager [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 642.109274] env[61957]: ERROR nova.compute.manager [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] self.driver.spawn(context, instance, image_meta, [ 642.109274] env[61957]: ERROR nova.compute.manager [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 642.109274] env[61957]: ERROR nova.compute.manager [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] self._vmops.spawn(context, instance, image_meta, injected_files, [ 642.109274] env[61957]: ERROR nova.compute.manager [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 642.109274] env[61957]: ERROR nova.compute.manager [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] vm_ref = self.build_virtual_machine(instance, [ 642.109274] env[61957]: ERROR nova.compute.manager [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 642.109274] env[61957]: ERROR nova.compute.manager [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] vif_infos = vmwarevif.get_vif_info(self._session, [ 642.109274] env[61957]: ERROR nova.compute.manager [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 642.109786] env[61957]: ERROR nova.compute.manager [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] for vif in network_info: [ 642.109786] env[61957]: ERROR nova.compute.manager [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 642.109786] env[61957]: ERROR nova.compute.manager [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] return self._sync_wrapper(fn, *args, **kwargs) [ 642.109786] env[61957]: ERROR nova.compute.manager [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 642.109786] env[61957]: ERROR nova.compute.manager [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] self.wait() [ 642.109786] env[61957]: ERROR nova.compute.manager [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 642.109786] env[61957]: ERROR nova.compute.manager [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] self[:] = self._gt.wait() [ 642.109786] env[61957]: ERROR nova.compute.manager [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 642.109786] env[61957]: ERROR nova.compute.manager [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] return self._exit_event.wait() [ 642.109786] env[61957]: ERROR nova.compute.manager [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 642.109786] env[61957]: ERROR nova.compute.manager [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] result = hub.switch() [ 642.109786] env[61957]: ERROR nova.compute.manager [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 642.109786] env[61957]: ERROR nova.compute.manager [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] return self.greenlet.switch() [ 642.110141] env[61957]: ERROR nova.compute.manager [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 642.110141] env[61957]: ERROR nova.compute.manager [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] result = function(*args, **kwargs) [ 642.110141] env[61957]: ERROR nova.compute.manager [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 642.110141] env[61957]: ERROR nova.compute.manager [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] return func(*args, **kwargs) [ 642.110141] env[61957]: ERROR nova.compute.manager [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 642.110141] env[61957]: ERROR nova.compute.manager [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] raise e [ 642.110141] env[61957]: ERROR nova.compute.manager [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 642.110141] env[61957]: ERROR nova.compute.manager [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] nwinfo = self.network_api.allocate_for_instance( [ 642.110141] env[61957]: ERROR nova.compute.manager [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 642.110141] env[61957]: ERROR nova.compute.manager [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] created_port_ids = self._update_ports_for_instance( [ 642.110141] env[61957]: ERROR nova.compute.manager [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 642.110141] env[61957]: ERROR nova.compute.manager [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] with excutils.save_and_reraise_exception(): [ 642.110141] env[61957]: ERROR nova.compute.manager [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 642.110466] env[61957]: ERROR nova.compute.manager [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] self.force_reraise() [ 642.110466] env[61957]: ERROR nova.compute.manager [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 642.110466] env[61957]: ERROR nova.compute.manager [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] raise self.value [ 642.110466] env[61957]: ERROR nova.compute.manager [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 642.110466] env[61957]: ERROR nova.compute.manager [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] updated_port = self._update_port( [ 642.110466] env[61957]: ERROR nova.compute.manager [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 642.110466] env[61957]: ERROR nova.compute.manager [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] _ensure_no_port_binding_failure(port) [ 642.110466] env[61957]: ERROR nova.compute.manager [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 642.110466] env[61957]: ERROR nova.compute.manager [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] raise exception.PortBindingFailed(port_id=port['id']) [ 642.110466] env[61957]: ERROR nova.compute.manager [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] nova.exception.PortBindingFailed: Binding failed for port 3e020214-1678-4e89-930e-bb37e8f3e31b, please check neutron logs for more information. [ 642.110466] env[61957]: ERROR nova.compute.manager [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] [ 642.110741] env[61957]: DEBUG nova.compute.utils [None req-260de5ae-9540-404d-a23e-9facd81096ee tempest-AttachInterfacesUnderV243Test-1229304814 tempest-AttachInterfacesUnderV243Test-1229304814-project-member] [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] Binding failed for port 3e020214-1678-4e89-930e-bb37e8f3e31b, please check neutron logs for more information. {{(pid=61957) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 642.111055] env[61957]: DEBUG oslo_concurrency.lockutils [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.433s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 642.112526] env[61957]: INFO nova.compute.claims [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 642.117010] env[61957]: DEBUG nova.compute.manager [None req-260de5ae-9540-404d-a23e-9facd81096ee tempest-AttachInterfacesUnderV243Test-1229304814 tempest-AttachInterfacesUnderV243Test-1229304814-project-member] [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] Build of instance 02d215d8-dd42-49cc-9509-08335c7bafe3 was re-scheduled: Binding failed for port 3e020214-1678-4e89-930e-bb37e8f3e31b, please check neutron logs for more information. {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 642.117486] env[61957]: DEBUG nova.compute.manager [None req-260de5ae-9540-404d-a23e-9facd81096ee tempest-AttachInterfacesUnderV243Test-1229304814 tempest-AttachInterfacesUnderV243Test-1229304814-project-member] [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] Unplugging VIFs for instance {{(pid=61957) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 642.117753] env[61957]: DEBUG oslo_concurrency.lockutils [None req-260de5ae-9540-404d-a23e-9facd81096ee tempest-AttachInterfacesUnderV243Test-1229304814 tempest-AttachInterfacesUnderV243Test-1229304814-project-member] Acquiring lock "refresh_cache-02d215d8-dd42-49cc-9509-08335c7bafe3" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 642.117905] env[61957]: DEBUG oslo_concurrency.lockutils [None req-260de5ae-9540-404d-a23e-9facd81096ee tempest-AttachInterfacesUnderV243Test-1229304814 tempest-AttachInterfacesUnderV243Test-1229304814-project-member] Acquired lock "refresh_cache-02d215d8-dd42-49cc-9509-08335c7bafe3" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 642.118076] env[61957]: DEBUG nova.network.neutron [None req-260de5ae-9540-404d-a23e-9facd81096ee tempest-AttachInterfacesUnderV243Test-1229304814 tempest-AttachInterfacesUnderV243Test-1229304814-project-member] [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 642.268812] env[61957]: INFO nova.compute.manager [None req-42e67e16-6377-41f9-9501-426c4335c678 tempest-ServersTestBootFromVolume-482289819 tempest-ServersTestBootFromVolume-482289819-project-member] [instance: 32e56693-5100-4ed8-b26a-7d377f9ab7b6] Took 1.04 seconds to deallocate network for instance. [ 642.417646] env[61957]: DEBUG oslo_concurrency.lockutils [None req-3d8587e3-7ad5-40a6-88af-e8ed44c49baa tempest-ServerActionsV293TestJSON-1269989829 tempest-ServerActionsV293TestJSON-1269989829-project-member] Acquiring lock "5dc993e9-5970-43ae-90be-f3a53d1f090a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 642.417646] env[61957]: DEBUG oslo_concurrency.lockutils [None req-3d8587e3-7ad5-40a6-88af-e8ed44c49baa tempest-ServerActionsV293TestJSON-1269989829 tempest-ServerActionsV293TestJSON-1269989829-project-member] Lock "5dc993e9-5970-43ae-90be-f3a53d1f090a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 642.654640] env[61957]: DEBUG nova.network.neutron [None req-260de5ae-9540-404d-a23e-9facd81096ee tempest-AttachInterfacesUnderV243Test-1229304814 tempest-AttachInterfacesUnderV243Test-1229304814-project-member] [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 642.686079] env[61957]: DEBUG nova.compute.manager [req-1d30ae70-8027-4b22-9ce5-1f3e9270a602 req-430707b4-be8b-4285-9bf3-48c729ec30c9 service nova] [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] Received event network-changed-b29b2b7c-e839-4ae7-a55b-fc34bee248b2 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 642.686079] env[61957]: DEBUG nova.compute.manager [req-1d30ae70-8027-4b22-9ce5-1f3e9270a602 req-430707b4-be8b-4285-9bf3-48c729ec30c9 service nova] [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] Refreshing instance network info cache due to event network-changed-b29b2b7c-e839-4ae7-a55b-fc34bee248b2. {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 642.686079] env[61957]: DEBUG oslo_concurrency.lockutils [req-1d30ae70-8027-4b22-9ce5-1f3e9270a602 req-430707b4-be8b-4285-9bf3-48c729ec30c9 service nova] Acquiring lock "refresh_cache-967056dd-1679-421c-b4fd-9d663b1411e2" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 642.686079] env[61957]: DEBUG oslo_concurrency.lockutils [req-1d30ae70-8027-4b22-9ce5-1f3e9270a602 req-430707b4-be8b-4285-9bf3-48c729ec30c9 service nova] Acquired lock "refresh_cache-967056dd-1679-421c-b4fd-9d663b1411e2" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 642.686079] env[61957]: DEBUG nova.network.neutron [req-1d30ae70-8027-4b22-9ce5-1f3e9270a602 req-430707b4-be8b-4285-9bf3-48c729ec30c9 service nova] [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] Refreshing network info cache for port b29b2b7c-e839-4ae7-a55b-fc34bee248b2 {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 642.865321] env[61957]: DEBUG nova.network.neutron [None req-260de5ae-9540-404d-a23e-9facd81096ee tempest-AttachInterfacesUnderV243Test-1229304814 tempest-AttachInterfacesUnderV243Test-1229304814-project-member] [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 643.040283] env[61957]: ERROR nova.compute.manager [None req-af971b81-6348-4338-b0a9-b06675d34db5 tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port b29b2b7c-e839-4ae7-a55b-fc34bee248b2, please check neutron logs for more information. [ 643.040283] env[61957]: ERROR nova.compute.manager Traceback (most recent call last): [ 643.040283] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 643.040283] env[61957]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 643.040283] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 643.040283] env[61957]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 643.040283] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 643.040283] env[61957]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 643.040283] env[61957]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 643.040283] env[61957]: ERROR nova.compute.manager self.force_reraise() [ 643.040283] env[61957]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 643.040283] env[61957]: ERROR nova.compute.manager raise self.value [ 643.040283] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 643.040283] env[61957]: ERROR nova.compute.manager updated_port = self._update_port( [ 643.040283] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 643.040283] env[61957]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 643.040639] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 643.040639] env[61957]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 643.040639] env[61957]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port b29b2b7c-e839-4ae7-a55b-fc34bee248b2, please check neutron logs for more information. [ 643.040639] env[61957]: ERROR nova.compute.manager [ 643.040639] env[61957]: Traceback (most recent call last): [ 643.040639] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 643.040639] env[61957]: listener.cb(fileno) [ 643.040639] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 643.040639] env[61957]: result = function(*args, **kwargs) [ 643.040639] env[61957]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 643.040639] env[61957]: return func(*args, **kwargs) [ 643.040639] env[61957]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 643.040639] env[61957]: raise e [ 643.040639] env[61957]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 643.040639] env[61957]: nwinfo = self.network_api.allocate_for_instance( [ 643.040639] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 643.040639] env[61957]: created_port_ids = self._update_ports_for_instance( [ 643.040639] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 643.040639] env[61957]: with excutils.save_and_reraise_exception(): [ 643.040639] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 643.040639] env[61957]: self.force_reraise() [ 643.040639] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 643.040639] env[61957]: raise self.value [ 643.040639] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 643.040639] env[61957]: updated_port = self._update_port( [ 643.040639] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 643.040639] env[61957]: _ensure_no_port_binding_failure(port) [ 643.040639] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 643.040639] env[61957]: raise exception.PortBindingFailed(port_id=port['id']) [ 643.041219] env[61957]: nova.exception.PortBindingFailed: Binding failed for port b29b2b7c-e839-4ae7-a55b-fc34bee248b2, please check neutron logs for more information. [ 643.041219] env[61957]: Removing descriptor: 17 [ 643.041219] env[61957]: ERROR nova.compute.manager [None req-af971b81-6348-4338-b0a9-b06675d34db5 tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port b29b2b7c-e839-4ae7-a55b-fc34bee248b2, please check neutron logs for more information. [ 643.041219] env[61957]: ERROR nova.compute.manager [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] Traceback (most recent call last): [ 643.041219] env[61957]: ERROR nova.compute.manager [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 643.041219] env[61957]: ERROR nova.compute.manager [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] yield resources [ 643.041219] env[61957]: ERROR nova.compute.manager [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 643.041219] env[61957]: ERROR nova.compute.manager [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] self.driver.spawn(context, instance, image_meta, [ 643.041219] env[61957]: ERROR nova.compute.manager [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 643.041219] env[61957]: ERROR nova.compute.manager [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] self._vmops.spawn(context, instance, image_meta, injected_files, [ 643.041219] env[61957]: ERROR nova.compute.manager [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 643.041219] env[61957]: ERROR nova.compute.manager [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] vm_ref = self.build_virtual_machine(instance, [ 643.041470] env[61957]: ERROR nova.compute.manager [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 643.041470] env[61957]: ERROR nova.compute.manager [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] vif_infos = vmwarevif.get_vif_info(self._session, [ 643.041470] env[61957]: ERROR nova.compute.manager [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 643.041470] env[61957]: ERROR nova.compute.manager [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] for vif in network_info: [ 643.041470] env[61957]: ERROR nova.compute.manager [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 643.041470] env[61957]: ERROR nova.compute.manager [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] return self._sync_wrapper(fn, *args, **kwargs) [ 643.041470] env[61957]: ERROR nova.compute.manager [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 643.041470] env[61957]: ERROR nova.compute.manager [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] self.wait() [ 643.041470] env[61957]: ERROR nova.compute.manager [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 643.041470] env[61957]: ERROR nova.compute.manager [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] self[:] = self._gt.wait() [ 643.041470] env[61957]: ERROR nova.compute.manager [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 643.041470] env[61957]: ERROR nova.compute.manager [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] return self._exit_event.wait() [ 643.041470] env[61957]: ERROR nova.compute.manager [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 643.041725] env[61957]: ERROR nova.compute.manager [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] result = hub.switch() [ 643.041725] env[61957]: ERROR nova.compute.manager [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 643.041725] env[61957]: ERROR nova.compute.manager [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] return self.greenlet.switch() [ 643.041725] env[61957]: ERROR nova.compute.manager [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 643.041725] env[61957]: ERROR nova.compute.manager [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] result = function(*args, **kwargs) [ 643.041725] env[61957]: ERROR nova.compute.manager [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 643.041725] env[61957]: ERROR nova.compute.manager [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] return func(*args, **kwargs) [ 643.041725] env[61957]: ERROR nova.compute.manager [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 643.041725] env[61957]: ERROR nova.compute.manager [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] raise e [ 643.041725] env[61957]: ERROR nova.compute.manager [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 643.041725] env[61957]: ERROR nova.compute.manager [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] nwinfo = self.network_api.allocate_for_instance( [ 643.041725] env[61957]: ERROR nova.compute.manager [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 643.041725] env[61957]: ERROR nova.compute.manager [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] created_port_ids = self._update_ports_for_instance( [ 643.041982] env[61957]: ERROR nova.compute.manager [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 643.041982] env[61957]: ERROR nova.compute.manager [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] with excutils.save_and_reraise_exception(): [ 643.041982] env[61957]: ERROR nova.compute.manager [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 643.041982] env[61957]: ERROR nova.compute.manager [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] self.force_reraise() [ 643.041982] env[61957]: ERROR nova.compute.manager [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 643.041982] env[61957]: ERROR nova.compute.manager [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] raise self.value [ 643.041982] env[61957]: ERROR nova.compute.manager [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 643.041982] env[61957]: ERROR nova.compute.manager [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] updated_port = self._update_port( [ 643.041982] env[61957]: ERROR nova.compute.manager [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 643.041982] env[61957]: ERROR nova.compute.manager [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] _ensure_no_port_binding_failure(port) [ 643.041982] env[61957]: ERROR nova.compute.manager [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 643.041982] env[61957]: ERROR nova.compute.manager [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] raise exception.PortBindingFailed(port_id=port['id']) [ 643.042277] env[61957]: ERROR nova.compute.manager [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] nova.exception.PortBindingFailed: Binding failed for port b29b2b7c-e839-4ae7-a55b-fc34bee248b2, please check neutron logs for more information. [ 643.042277] env[61957]: ERROR nova.compute.manager [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] [ 643.042277] env[61957]: INFO nova.compute.manager [None req-af971b81-6348-4338-b0a9-b06675d34db5 tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] Terminating instance [ 643.044922] env[61957]: DEBUG oslo_concurrency.lockutils [None req-af971b81-6348-4338-b0a9-b06675d34db5 tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] Acquiring lock "refresh_cache-967056dd-1679-421c-b4fd-9d663b1411e2" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 643.253687] env[61957]: DEBUG nova.network.neutron [req-1d30ae70-8027-4b22-9ce5-1f3e9270a602 req-430707b4-be8b-4285-9bf3-48c729ec30c9 service nova] [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 643.305827] env[61957]: INFO nova.scheduler.client.report [None req-42e67e16-6377-41f9-9501-426c4335c678 tempest-ServersTestBootFromVolume-482289819 tempest-ServersTestBootFromVolume-482289819-project-member] Deleted allocations for instance 32e56693-5100-4ed8-b26a-7d377f9ab7b6 [ 643.367561] env[61957]: DEBUG oslo_concurrency.lockutils [None req-260de5ae-9540-404d-a23e-9facd81096ee tempest-AttachInterfacesUnderV243Test-1229304814 tempest-AttachInterfacesUnderV243Test-1229304814-project-member] Releasing lock "refresh_cache-02d215d8-dd42-49cc-9509-08335c7bafe3" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 643.367857] env[61957]: DEBUG nova.compute.manager [None req-260de5ae-9540-404d-a23e-9facd81096ee tempest-AttachInterfacesUnderV243Test-1229304814 tempest-AttachInterfacesUnderV243Test-1229304814-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61957) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 643.367968] env[61957]: DEBUG nova.compute.manager [None req-260de5ae-9540-404d-a23e-9facd81096ee tempest-AttachInterfacesUnderV243Test-1229304814 tempest-AttachInterfacesUnderV243Test-1229304814-project-member] [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 643.368144] env[61957]: DEBUG nova.network.neutron [None req-260de5ae-9540-404d-a23e-9facd81096ee tempest-AttachInterfacesUnderV243Test-1229304814 tempest-AttachInterfacesUnderV243Test-1229304814-project-member] [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 643.402591] env[61957]: DEBUG nova.network.neutron [None req-260de5ae-9540-404d-a23e-9facd81096ee tempest-AttachInterfacesUnderV243Test-1229304814 tempest-AttachInterfacesUnderV243Test-1229304814-project-member] [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 643.442854] env[61957]: DEBUG nova.network.neutron [req-1d30ae70-8027-4b22-9ce5-1f3e9270a602 req-430707b4-be8b-4285-9bf3-48c729ec30c9 service nova] [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 643.530455] env[61957]: DEBUG oslo_concurrency.lockutils [None req-394366c7-3cf4-481b-bb9f-4da67892af40 tempest-ServersTestManualDisk-495750068 tempest-ServersTestManualDisk-495750068-project-member] Acquiring lock "517eeb71-e9e3-4284-a144-dc4b63319d2d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 643.531082] env[61957]: DEBUG oslo_concurrency.lockutils [None req-394366c7-3cf4-481b-bb9f-4da67892af40 tempest-ServersTestManualDisk-495750068 tempest-ServersTestManualDisk-495750068-project-member] Lock "517eeb71-e9e3-4284-a144-dc4b63319d2d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 643.648479] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-208d2096-6317-44a8-8674-b4752be108a5 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.657373] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86d5f22c-8a23-44d6-ae8f-4bf1f3894ab4 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.695506] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f2087b8-d558-40c3-ae5e-f5bb0a0bd4c8 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.705789] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b210fc77-dd31-4285-b80d-30c107ff72eb {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.722143] env[61957]: DEBUG nova.compute.provider_tree [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 643.817614] env[61957]: DEBUG oslo_concurrency.lockutils [None req-42e67e16-6377-41f9-9501-426c4335c678 tempest-ServersTestBootFromVolume-482289819 tempest-ServersTestBootFromVolume-482289819-project-member] Lock "32e56693-5100-4ed8-b26a-7d377f9ab7b6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 61.435s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 643.905186] env[61957]: DEBUG nova.network.neutron [None req-260de5ae-9540-404d-a23e-9facd81096ee tempest-AttachInterfacesUnderV243Test-1229304814 tempest-AttachInterfacesUnderV243Test-1229304814-project-member] [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 643.945648] env[61957]: DEBUG oslo_concurrency.lockutils [req-1d30ae70-8027-4b22-9ce5-1f3e9270a602 req-430707b4-be8b-4285-9bf3-48c729ec30c9 service nova] Releasing lock "refresh_cache-967056dd-1679-421c-b4fd-9d663b1411e2" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 643.946312] env[61957]: DEBUG oslo_concurrency.lockutils [None req-af971b81-6348-4338-b0a9-b06675d34db5 tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] Acquired lock "refresh_cache-967056dd-1679-421c-b4fd-9d663b1411e2" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 643.946312] env[61957]: DEBUG nova.network.neutron [None req-af971b81-6348-4338-b0a9-b06675d34db5 tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 644.227204] env[61957]: DEBUG nova.scheduler.client.report [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 644.322667] env[61957]: DEBUG nova.compute.manager [None req-fc6b77c8-b5cd-4608-beab-d2cb5c9f7440 tempest-ImagesNegativeTestJSON-1548152881 tempest-ImagesNegativeTestJSON-1548152881-project-member] [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] Starting instance... {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 644.408109] env[61957]: INFO nova.compute.manager [None req-260de5ae-9540-404d-a23e-9facd81096ee tempest-AttachInterfacesUnderV243Test-1229304814 tempest-AttachInterfacesUnderV243Test-1229304814-project-member] [instance: 02d215d8-dd42-49cc-9509-08335c7bafe3] Took 1.04 seconds to deallocate network for instance. [ 644.509019] env[61957]: DEBUG nova.network.neutron [None req-af971b81-6348-4338-b0a9-b06675d34db5 tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 644.620178] env[61957]: DEBUG nova.network.neutron [None req-af971b81-6348-4338-b0a9-b06675d34db5 tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 644.733562] env[61957]: DEBUG oslo_concurrency.lockutils [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.622s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 644.734160] env[61957]: DEBUG nova.compute.manager [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] Start building networks asynchronously for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 644.736700] env[61957]: DEBUG oslo_concurrency.lockutils [None req-a12d1497-5e6b-4b01-8612-e7e1d2fcba68 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 17.313s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 644.736916] env[61957]: DEBUG nova.objects.instance [None req-a12d1497-5e6b-4b01-8612-e7e1d2fcba68 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] Lazy-loading 'resources' on Instance uuid 4ad746d0-4076-4c97-8ea9-20e49fda59ca {{(pid=61957) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 644.769252] env[61957]: DEBUG nova.compute.manager [req-53d0bd5e-5ffd-4501-b74b-11944d9baddc req-0251865f-6626-4da2-823c-14c952c8c76f service nova] [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] Received event network-vif-deleted-b29b2b7c-e839-4ae7-a55b-fc34bee248b2 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 644.849140] env[61957]: DEBUG oslo_concurrency.lockutils [None req-fc6b77c8-b5cd-4608-beab-d2cb5c9f7440 tempest-ImagesNegativeTestJSON-1548152881 tempest-ImagesNegativeTestJSON-1548152881-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 645.125040] env[61957]: DEBUG oslo_concurrency.lockutils [None req-af971b81-6348-4338-b0a9-b06675d34db5 tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] Releasing lock "refresh_cache-967056dd-1679-421c-b4fd-9d663b1411e2" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 645.125040] env[61957]: DEBUG nova.compute.manager [None req-af971b81-6348-4338-b0a9-b06675d34db5 tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] Start destroying the instance on the hypervisor. {{(pid=61957) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 645.125040] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-af971b81-6348-4338-b0a9-b06675d34db5 tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] Destroying instance {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 645.125040] env[61957]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-335156d0-3353-48f3-8940-a1d829183f62 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.137991] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e69b356-729a-49ba-8f70-f3c19b1d69de {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.164986] env[61957]: WARNING nova.virt.vmwareapi.vmops [None req-af971b81-6348-4338-b0a9-b06675d34db5 tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 967056dd-1679-421c-b4fd-9d663b1411e2 could not be found. [ 645.165089] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-af971b81-6348-4338-b0a9-b06675d34db5 tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] Instance destroyed {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 645.165372] env[61957]: INFO nova.compute.manager [None req-af971b81-6348-4338-b0a9-b06675d34db5 tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] Took 0.04 seconds to destroy the instance on the hypervisor. [ 645.165501] env[61957]: DEBUG oslo.service.loopingcall [None req-af971b81-6348-4338-b0a9-b06675d34db5 tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 645.167355] env[61957]: DEBUG nova.compute.manager [-] [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 645.167482] env[61957]: DEBUG nova.network.neutron [-] [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 645.242516] env[61957]: DEBUG nova.compute.utils [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] Using /dev/sd instead of None {{(pid=61957) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 645.249748] env[61957]: DEBUG nova.compute.manager [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] Allocating IP information in the background. {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 645.249748] env[61957]: DEBUG nova.network.neutron [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] allocate_for_instance() {{(pid=61957) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 645.255975] env[61957]: DEBUG nova.network.neutron [-] [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 645.325483] env[61957]: DEBUG nova.policy [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5283a75779d74aa7aee51e6b066769d4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '37ffd9c6cf05433e87a3739433cc7248', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61957) authorize /opt/stack/nova/nova/policy.py:203}} [ 645.443868] env[61957]: INFO nova.scheduler.client.report [None req-260de5ae-9540-404d-a23e-9facd81096ee tempest-AttachInterfacesUnderV243Test-1229304814 tempest-AttachInterfacesUnderV243Test-1229304814-project-member] Deleted allocations for instance 02d215d8-dd42-49cc-9509-08335c7bafe3 [ 645.752897] env[61957]: DEBUG nova.compute.manager [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] Start building block device mappings for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 645.758276] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-880c9bb1-3f6b-4ea7-bef4-5db5c37fb782 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.770982] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8f0c24c-eb40-4a25-bbda-1796d95f2d9d {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.811852] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f0a837d-8e0a-4552-b005-ee4cb03bd873 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.819735] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e834e15a-bece-403c-a83c-ea6a67134db1 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.833902] env[61957]: DEBUG nova.compute.provider_tree [None req-a12d1497-5e6b-4b01-8612-e7e1d2fcba68 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 645.952301] env[61957]: DEBUG nova.network.neutron [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] Successfully created port: 9758f59c-2ebd-4314-8d87-c51d41bbbcbe {{(pid=61957) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 645.956318] env[61957]: DEBUG oslo_concurrency.lockutils [None req-260de5ae-9540-404d-a23e-9facd81096ee tempest-AttachInterfacesUnderV243Test-1229304814 tempest-AttachInterfacesUnderV243Test-1229304814-project-member] Lock "02d215d8-dd42-49cc-9509-08335c7bafe3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 61.473s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 646.337072] env[61957]: DEBUG nova.scheduler.client.report [None req-a12d1497-5e6b-4b01-8612-e7e1d2fcba68 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 646.457064] env[61957]: DEBUG nova.compute.manager [None req-f19b0e78-2c52-4a40-8c7e-832f687868ae tempest-ServersListShow296Test-1066136130 tempest-ServersListShow296Test-1066136130-project-member] [instance: af0a4907-4206-467f-98b9-a13b99fdf432] Starting instance... {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 646.537284] env[61957]: DEBUG nova.network.neutron [-] [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 646.770651] env[61957]: DEBUG nova.compute.manager [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] Start spawning the instance on the hypervisor. {{(pid=61957) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 646.802695] env[61957]: DEBUG nova.virt.hardware [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T17:22:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T17:21:50Z,direct_url=,disk_format='vmdk',id=11c76a2c-f705-470a-ba9d-4657858bab38,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='81d3096d0f094373913fc3dc8f5c3c6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T17:21:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 646.802835] env[61957]: DEBUG nova.virt.hardware [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] Flavor limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 646.803018] env[61957]: DEBUG nova.virt.hardware [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] Image limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 646.803212] env[61957]: DEBUG nova.virt.hardware [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] Flavor pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 646.803367] env[61957]: DEBUG nova.virt.hardware [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] Image pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 646.803489] env[61957]: DEBUG nova.virt.hardware [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 646.803699] env[61957]: DEBUG nova.virt.hardware [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 646.804097] env[61957]: DEBUG nova.virt.hardware [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 646.804097] env[61957]: DEBUG nova.virt.hardware [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] Got 1 possible topologies {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 646.806316] env[61957]: DEBUG nova.virt.hardware [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 646.806454] env[61957]: DEBUG nova.virt.hardware [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 646.807924] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-409252e7-47b7-446e-99fa-c786e6680005 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.817273] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b64ca2a-e49e-41c9-a335-1998a513f3d6 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.842619] env[61957]: DEBUG oslo_concurrency.lockutils [None req-a12d1497-5e6b-4b01-8612-e7e1d2fcba68 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.105s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 646.844913] env[61957]: DEBUG oslo_concurrency.lockutils [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.977s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 646.846453] env[61957]: INFO nova.compute.claims [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 646.892086] env[61957]: INFO nova.scheduler.client.report [None req-a12d1497-5e6b-4b01-8612-e7e1d2fcba68 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] Deleted allocations for instance 4ad746d0-4076-4c97-8ea9-20e49fda59ca [ 646.961367] env[61957]: DEBUG nova.compute.manager [None req-f19b0e78-2c52-4a40-8c7e-832f687868ae tempest-ServersListShow296Test-1066136130 tempest-ServersListShow296Test-1066136130-project-member] [instance: af0a4907-4206-467f-98b9-a13b99fdf432] Instance disappeared before build. {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 647.045859] env[61957]: INFO nova.compute.manager [-] [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] Took 1.88 seconds to deallocate network for instance. [ 647.048660] env[61957]: DEBUG nova.compute.claims [None req-af971b81-6348-4338-b0a9-b06675d34db5 tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] Aborting claim: {{(pid=61957) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 647.048660] env[61957]: DEBUG oslo_concurrency.lockutils [None req-af971b81-6348-4338-b0a9-b06675d34db5 tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 647.402248] env[61957]: DEBUG oslo_concurrency.lockutils [None req-a12d1497-5e6b-4b01-8612-e7e1d2fcba68 tempest-ServersAaction247Test-35473783 tempest-ServersAaction247Test-35473783-project-member] Lock "4ad746d0-4076-4c97-8ea9-20e49fda59ca" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 23.871s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 647.482030] env[61957]: DEBUG oslo_concurrency.lockutils [None req-f19b0e78-2c52-4a40-8c7e-832f687868ae tempest-ServersListShow296Test-1066136130 tempest-ServersListShow296Test-1066136130-project-member] Lock "af0a4907-4206-467f-98b9-a13b99fdf432" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 46.995s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 647.758458] env[61957]: DEBUG nova.compute.manager [req-deca7961-0902-4f78-b615-850253c47146 req-6b044266-b78c-42ea-a633-c4892c290876 service nova] [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] Received event network-changed-9758f59c-2ebd-4314-8d87-c51d41bbbcbe {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 647.758678] env[61957]: DEBUG nova.compute.manager [req-deca7961-0902-4f78-b615-850253c47146 req-6b044266-b78c-42ea-a633-c4892c290876 service nova] [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] Refreshing instance network info cache due to event network-changed-9758f59c-2ebd-4314-8d87-c51d41bbbcbe. {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 647.759021] env[61957]: DEBUG oslo_concurrency.lockutils [req-deca7961-0902-4f78-b615-850253c47146 req-6b044266-b78c-42ea-a633-c4892c290876 service nova] Acquiring lock "refresh_cache-74d3de77-1272-449d-8b64-75e21fff8d7f" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 647.759083] env[61957]: DEBUG oslo_concurrency.lockutils [req-deca7961-0902-4f78-b615-850253c47146 req-6b044266-b78c-42ea-a633-c4892c290876 service nova] Acquired lock "refresh_cache-74d3de77-1272-449d-8b64-75e21fff8d7f" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 647.759216] env[61957]: DEBUG nova.network.neutron [req-deca7961-0902-4f78-b615-850253c47146 req-6b044266-b78c-42ea-a633-c4892c290876 service nova] [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] Refreshing network info cache for port 9758f59c-2ebd-4314-8d87-c51d41bbbcbe {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 647.989163] env[61957]: DEBUG nova.compute.manager [None req-3d56c405-76b3-4bdb-9787-ecf239aeeb0d tempest-ImagesOneServerNegativeTestJSON-1581220397 tempest-ImagesOneServerNegativeTestJSON-1581220397-project-member] [instance: 2b151571-348f-4543-a0c5-afe6458b1973] Starting instance... {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 648.146069] env[61957]: ERROR nova.compute.manager [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 9758f59c-2ebd-4314-8d87-c51d41bbbcbe, please check neutron logs for more information. [ 648.146069] env[61957]: ERROR nova.compute.manager Traceback (most recent call last): [ 648.146069] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 648.146069] env[61957]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 648.146069] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 648.146069] env[61957]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 648.146069] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 648.146069] env[61957]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 648.146069] env[61957]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 648.146069] env[61957]: ERROR nova.compute.manager self.force_reraise() [ 648.146069] env[61957]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 648.146069] env[61957]: ERROR nova.compute.manager raise self.value [ 648.146069] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 648.146069] env[61957]: ERROR nova.compute.manager updated_port = self._update_port( [ 648.146069] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 648.146069] env[61957]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 648.146469] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 648.146469] env[61957]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 648.146469] env[61957]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 9758f59c-2ebd-4314-8d87-c51d41bbbcbe, please check neutron logs for more information. [ 648.146469] env[61957]: ERROR nova.compute.manager [ 648.146469] env[61957]: Traceback (most recent call last): [ 648.146469] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 648.146469] env[61957]: listener.cb(fileno) [ 648.146469] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 648.146469] env[61957]: result = function(*args, **kwargs) [ 648.146469] env[61957]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 648.146469] env[61957]: return func(*args, **kwargs) [ 648.146469] env[61957]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 648.146469] env[61957]: raise e [ 648.146469] env[61957]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 648.146469] env[61957]: nwinfo = self.network_api.allocate_for_instance( [ 648.146469] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 648.146469] env[61957]: created_port_ids = self._update_ports_for_instance( [ 648.146469] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 648.146469] env[61957]: with excutils.save_and_reraise_exception(): [ 648.146469] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 648.146469] env[61957]: self.force_reraise() [ 648.146469] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 648.146469] env[61957]: raise self.value [ 648.146469] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 648.146469] env[61957]: updated_port = self._update_port( [ 648.146469] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 648.146469] env[61957]: _ensure_no_port_binding_failure(port) [ 648.146469] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 648.146469] env[61957]: raise exception.PortBindingFailed(port_id=port['id']) [ 648.147140] env[61957]: nova.exception.PortBindingFailed: Binding failed for port 9758f59c-2ebd-4314-8d87-c51d41bbbcbe, please check neutron logs for more information. [ 648.147140] env[61957]: Removing descriptor: 15 [ 648.147140] env[61957]: ERROR nova.compute.manager [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 9758f59c-2ebd-4314-8d87-c51d41bbbcbe, please check neutron logs for more information. [ 648.147140] env[61957]: ERROR nova.compute.manager [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] Traceback (most recent call last): [ 648.147140] env[61957]: ERROR nova.compute.manager [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 648.147140] env[61957]: ERROR nova.compute.manager [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] yield resources [ 648.147140] env[61957]: ERROR nova.compute.manager [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 648.147140] env[61957]: ERROR nova.compute.manager [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] self.driver.spawn(context, instance, image_meta, [ 648.147140] env[61957]: ERROR nova.compute.manager [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 648.147140] env[61957]: ERROR nova.compute.manager [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 648.147140] env[61957]: ERROR nova.compute.manager [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 648.147140] env[61957]: ERROR nova.compute.manager [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] vm_ref = self.build_virtual_machine(instance, [ 648.147406] env[61957]: ERROR nova.compute.manager [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 648.147406] env[61957]: ERROR nova.compute.manager [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] vif_infos = vmwarevif.get_vif_info(self._session, [ 648.147406] env[61957]: ERROR nova.compute.manager [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 648.147406] env[61957]: ERROR nova.compute.manager [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] for vif in network_info: [ 648.147406] env[61957]: ERROR nova.compute.manager [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 648.147406] env[61957]: ERROR nova.compute.manager [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] return self._sync_wrapper(fn, *args, **kwargs) [ 648.147406] env[61957]: ERROR nova.compute.manager [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 648.147406] env[61957]: ERROR nova.compute.manager [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] self.wait() [ 648.147406] env[61957]: ERROR nova.compute.manager [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 648.147406] env[61957]: ERROR nova.compute.manager [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] self[:] = self._gt.wait() [ 648.147406] env[61957]: ERROR nova.compute.manager [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 648.147406] env[61957]: ERROR nova.compute.manager [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] return self._exit_event.wait() [ 648.147406] env[61957]: ERROR nova.compute.manager [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 648.147711] env[61957]: ERROR nova.compute.manager [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] result = hub.switch() [ 648.147711] env[61957]: ERROR nova.compute.manager [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 648.147711] env[61957]: ERROR nova.compute.manager [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] return self.greenlet.switch() [ 648.147711] env[61957]: ERROR nova.compute.manager [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 648.147711] env[61957]: ERROR nova.compute.manager [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] result = function(*args, **kwargs) [ 648.147711] env[61957]: ERROR nova.compute.manager [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 648.147711] env[61957]: ERROR nova.compute.manager [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] return func(*args, **kwargs) [ 648.147711] env[61957]: ERROR nova.compute.manager [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 648.147711] env[61957]: ERROR nova.compute.manager [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] raise e [ 648.147711] env[61957]: ERROR nova.compute.manager [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 648.147711] env[61957]: ERROR nova.compute.manager [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] nwinfo = self.network_api.allocate_for_instance( [ 648.147711] env[61957]: ERROR nova.compute.manager [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 648.147711] env[61957]: ERROR nova.compute.manager [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] created_port_ids = self._update_ports_for_instance( [ 648.147988] env[61957]: ERROR nova.compute.manager [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 648.147988] env[61957]: ERROR nova.compute.manager [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] with excutils.save_and_reraise_exception(): [ 648.147988] env[61957]: ERROR nova.compute.manager [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 648.147988] env[61957]: ERROR nova.compute.manager [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] self.force_reraise() [ 648.147988] env[61957]: ERROR nova.compute.manager [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 648.147988] env[61957]: ERROR nova.compute.manager [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] raise self.value [ 648.147988] env[61957]: ERROR nova.compute.manager [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 648.147988] env[61957]: ERROR nova.compute.manager [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] updated_port = self._update_port( [ 648.147988] env[61957]: ERROR nova.compute.manager [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 648.147988] env[61957]: ERROR nova.compute.manager [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] _ensure_no_port_binding_failure(port) [ 648.147988] env[61957]: ERROR nova.compute.manager [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 648.147988] env[61957]: ERROR nova.compute.manager [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] raise exception.PortBindingFailed(port_id=port['id']) [ 648.148263] env[61957]: ERROR nova.compute.manager [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] nova.exception.PortBindingFailed: Binding failed for port 9758f59c-2ebd-4314-8d87-c51d41bbbcbe, please check neutron logs for more information. [ 648.148263] env[61957]: ERROR nova.compute.manager [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] [ 648.148263] env[61957]: INFO nova.compute.manager [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] Terminating instance [ 648.150301] env[61957]: DEBUG oslo_concurrency.lockutils [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] Acquiring lock "refresh_cache-74d3de77-1272-449d-8b64-75e21fff8d7f" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 648.290629] env[61957]: DEBUG nova.network.neutron [req-deca7961-0902-4f78-b615-850253c47146 req-6b044266-b78c-42ea-a633-c4892c290876 service nova] [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 648.350415] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4c44da8-20d1-4426-b69b-cf6ca99c8d1a {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.362822] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e2d74b0-8c5c-47aa-8f3f-6b7a99e0b5a8 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.395933] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1073ccba-eaa4-463a-a8aa-a9d8efb59377 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.403740] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de122566-556f-4ca5-9dff-226a8508bd6d {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.420722] env[61957]: DEBUG nova.compute.provider_tree [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 648.424135] env[61957]: DEBUG nova.network.neutron [req-deca7961-0902-4f78-b615-850253c47146 req-6b044266-b78c-42ea-a633-c4892c290876 service nova] [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 648.515379] env[61957]: DEBUG oslo_concurrency.lockutils [None req-3d56c405-76b3-4bdb-9787-ecf239aeeb0d tempest-ImagesOneServerNegativeTestJSON-1581220397 tempest-ImagesOneServerNegativeTestJSON-1581220397-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 648.926719] env[61957]: DEBUG nova.scheduler.client.report [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 648.930730] env[61957]: DEBUG oslo_concurrency.lockutils [req-deca7961-0902-4f78-b615-850253c47146 req-6b044266-b78c-42ea-a633-c4892c290876 service nova] Releasing lock "refresh_cache-74d3de77-1272-449d-8b64-75e21fff8d7f" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 648.933409] env[61957]: DEBUG oslo_concurrency.lockutils [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] Acquired lock "refresh_cache-74d3de77-1272-449d-8b64-75e21fff8d7f" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 648.933632] env[61957]: DEBUG nova.network.neutron [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 649.438991] env[61957]: DEBUG oslo_concurrency.lockutils [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.594s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 649.439718] env[61957]: DEBUG nova.compute.manager [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] Start building networks asynchronously for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 649.449253] env[61957]: DEBUG oslo_concurrency.lockutils [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.876s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 649.450909] env[61957]: INFO nova.compute.claims [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] [instance: 394b1955-c448-42c2-a718-28df7bd366e9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 649.487875] env[61957]: DEBUG nova.network.neutron [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 649.681512] env[61957]: DEBUG nova.network.neutron [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 649.892198] env[61957]: DEBUG nova.compute.manager [req-4af897bb-2e24-4d31-9232-7994e4d295f6 req-25461eeb-3479-4199-8e23-3177d1bcdacf service nova] [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] Received event network-vif-deleted-9758f59c-2ebd-4314-8d87-c51d41bbbcbe {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 649.957257] env[61957]: DEBUG nova.compute.utils [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] Using /dev/sd instead of None {{(pid=61957) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 649.962245] env[61957]: DEBUG nova.compute.manager [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] Allocating IP information in the background. {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 649.962245] env[61957]: DEBUG nova.network.neutron [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] allocate_for_instance() {{(pid=61957) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 650.184856] env[61957]: DEBUG oslo_concurrency.lockutils [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] Releasing lock "refresh_cache-74d3de77-1272-449d-8b64-75e21fff8d7f" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 650.185326] env[61957]: DEBUG nova.compute.manager [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] Start destroying the instance on the hypervisor. {{(pid=61957) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 650.185531] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] Destroying instance {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 650.185840] env[61957]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d26d5df9-3b45-44bc-b173-7fa8d69e821a {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.195051] env[61957]: DEBUG nova.policy [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5283a75779d74aa7aee51e6b066769d4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '37ffd9c6cf05433e87a3739433cc7248', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61957) authorize /opt/stack/nova/nova/policy.py:203}} [ 650.201817] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b00f8f71-8612-49de-9ee6-41aa52248679 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.232804] env[61957]: WARNING nova.virt.vmwareapi.vmops [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 74d3de77-1272-449d-8b64-75e21fff8d7f could not be found. [ 650.233066] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] Instance destroyed {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 650.233294] env[61957]: INFO nova.compute.manager [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] Took 0.05 seconds to destroy the instance on the hypervisor. [ 650.233541] env[61957]: DEBUG oslo.service.loopingcall [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 650.234496] env[61957]: DEBUG nova.compute.manager [-] [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 650.234496] env[61957]: DEBUG nova.network.neutron [-] [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 650.267847] env[61957]: DEBUG nova.network.neutron [-] [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 650.464729] env[61957]: DEBUG nova.compute.manager [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] Start building block device mappings for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 650.780082] env[61957]: DEBUG nova.network.neutron [-] [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 650.940527] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-882be459-cb95-4751-899f-158d555c1910 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.948588] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76ccd1df-2e56-4331-9763-b16a155cbaca {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.984022] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01653917-4c7e-47b4-ab92-d9a4e5fb78bf {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.990350] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54126425-c5b9-4bb1-98bb-2f734a48f3ea {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.003986] env[61957]: DEBUG nova.compute.provider_tree [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 651.020261] env[61957]: DEBUG nova.network.neutron [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] Successfully created port: ebb67595-2f07-4312-a3d9-11a91467c152 {{(pid=61957) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 651.282830] env[61957]: INFO nova.compute.manager [-] [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] Took 1.05 seconds to deallocate network for instance. [ 651.285525] env[61957]: DEBUG nova.compute.claims [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] Aborting claim: {{(pid=61957) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 651.285683] env[61957]: DEBUG oslo_concurrency.lockutils [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 651.487257] env[61957]: DEBUG nova.compute.manager [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] Start spawning the instance on the hypervisor. {{(pid=61957) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 651.507821] env[61957]: DEBUG nova.scheduler.client.report [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 651.526575] env[61957]: DEBUG nova.virt.hardware [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T17:22:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T17:21:50Z,direct_url=,disk_format='vmdk',id=11c76a2c-f705-470a-ba9d-4657858bab38,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='81d3096d0f094373913fc3dc8f5c3c6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T17:21:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 651.526718] env[61957]: DEBUG nova.virt.hardware [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] Flavor limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 651.526777] env[61957]: DEBUG nova.virt.hardware [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] Image limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 651.526958] env[61957]: DEBUG nova.virt.hardware [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] Flavor pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 651.527174] env[61957]: DEBUG nova.virt.hardware [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] Image pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 651.527253] env[61957]: DEBUG nova.virt.hardware [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 651.527451] env[61957]: DEBUG nova.virt.hardware [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 651.527633] env[61957]: DEBUG nova.virt.hardware [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 651.527815] env[61957]: DEBUG nova.virt.hardware [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] Got 1 possible topologies {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 651.528071] env[61957]: DEBUG nova.virt.hardware [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 651.528736] env[61957]: DEBUG nova.virt.hardware [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 651.530246] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d71c2a8a-d433-4360-a598-0089668c25a0 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.540180] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9308523-64c2-4f58-96e3-77e3c63aec39 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.111899] env[61957]: DEBUG oslo_concurrency.lockutils [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.570s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 652.111899] env[61957]: DEBUG nova.compute.manager [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] [instance: 394b1955-c448-42c2-a718-28df7bd366e9] Start building networks asynchronously for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 652.111899] env[61957]: DEBUG oslo_concurrency.lockutils [None req-adf5e3c1-3919-48f3-8d1a-68a60974255e tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 20.326s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 652.530421] env[61957]: DEBUG nova.compute.utils [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] Using /dev/sd instead of None {{(pid=61957) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 652.539312] env[61957]: DEBUG nova.compute.manager [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] [instance: 394b1955-c448-42c2-a718-28df7bd366e9] Allocating IP information in the background. {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 652.539312] env[61957]: DEBUG nova.network.neutron [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] [instance: 394b1955-c448-42c2-a718-28df7bd366e9] allocate_for_instance() {{(pid=61957) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 652.671776] env[61957]: DEBUG nova.policy [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5283a75779d74aa7aee51e6b066769d4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '37ffd9c6cf05433e87a3739433cc7248', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61957) authorize /opt/stack/nova/nova/policy.py:203}} [ 653.037488] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3ae7018-d94c-42f8-8042-9645972436ef {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.040812] env[61957]: DEBUG nova.compute.manager [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] [instance: 394b1955-c448-42c2-a718-28df7bd366e9] Start building block device mappings for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 653.049650] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fffd142-1427-4c2a-b1d6-3aaeb4891c45 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.081481] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2390b19-d128-4d8c-8293-7f278fd8c260 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.089185] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bca10d6-4de1-4f6c-a8e0-dcf9eb80bb4f {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.105680] env[61957]: DEBUG nova.compute.provider_tree [None req-adf5e3c1-3919-48f3-8d1a-68a60974255e tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 653.198590] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d6c9937c-2526-4768-bade-4ca2ced8b569 tempest-ServersTestFqdnHostnames-1308211839 tempest-ServersTestFqdnHostnames-1308211839-project-member] Acquiring lock "5cb09b60-70b6-4d04-850e-049612ec4a89" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 653.198977] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d6c9937c-2526-4768-bade-4ca2ced8b569 tempest-ServersTestFqdnHostnames-1308211839 tempest-ServersTestFqdnHostnames-1308211839-project-member] Lock "5cb09b60-70b6-4d04-850e-049612ec4a89" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 653.505205] env[61957]: DEBUG nova.compute.manager [req-2ee6ddb6-4311-459a-9474-261c26208567 req-495c8ba7-1c66-417b-95d3-8c1b8f85e564 service nova] [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] Received event network-changed-ebb67595-2f07-4312-a3d9-11a91467c152 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 653.506139] env[61957]: DEBUG nova.compute.manager [req-2ee6ddb6-4311-459a-9474-261c26208567 req-495c8ba7-1c66-417b-95d3-8c1b8f85e564 service nova] [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] Refreshing instance network info cache due to event network-changed-ebb67595-2f07-4312-a3d9-11a91467c152. {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 653.506139] env[61957]: DEBUG oslo_concurrency.lockutils [req-2ee6ddb6-4311-459a-9474-261c26208567 req-495c8ba7-1c66-417b-95d3-8c1b8f85e564 service nova] Acquiring lock "refresh_cache-b541c437-93b7-4e4d-9ff8-bfc9cdf0e237" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 653.506139] env[61957]: DEBUG oslo_concurrency.lockutils [req-2ee6ddb6-4311-459a-9474-261c26208567 req-495c8ba7-1c66-417b-95d3-8c1b8f85e564 service nova] Acquired lock "refresh_cache-b541c437-93b7-4e4d-9ff8-bfc9cdf0e237" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 653.506139] env[61957]: DEBUG nova.network.neutron [req-2ee6ddb6-4311-459a-9474-261c26208567 req-495c8ba7-1c66-417b-95d3-8c1b8f85e564 service nova] [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] Refreshing network info cache for port ebb67595-2f07-4312-a3d9-11a91467c152 {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 653.522868] env[61957]: DEBUG nova.network.neutron [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] [instance: 394b1955-c448-42c2-a718-28df7bd366e9] Successfully created port: 10ac5929-6235-4c76-b38b-45cc8b90ca43 {{(pid=61957) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 653.608914] env[61957]: DEBUG nova.scheduler.client.report [None req-adf5e3c1-3919-48f3-8d1a-68a60974255e tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 654.050429] env[61957]: DEBUG nova.compute.manager [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] [instance: 394b1955-c448-42c2-a718-28df7bd366e9] Start spawning the instance on the hypervisor. {{(pid=61957) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 654.083096] env[61957]: DEBUG nova.network.neutron [req-2ee6ddb6-4311-459a-9474-261c26208567 req-495c8ba7-1c66-417b-95d3-8c1b8f85e564 service nova] [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 654.089171] env[61957]: DEBUG nova.virt.hardware [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T17:22:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T17:21:50Z,direct_url=,disk_format='vmdk',id=11c76a2c-f705-470a-ba9d-4657858bab38,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='81d3096d0f094373913fc3dc8f5c3c6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T17:21:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 654.089171] env[61957]: DEBUG nova.virt.hardware [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] Flavor limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 654.089351] env[61957]: DEBUG nova.virt.hardware [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] Image limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 654.089543] env[61957]: DEBUG nova.virt.hardware [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] Flavor pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 654.089695] env[61957]: DEBUG nova.virt.hardware [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] Image pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 654.089844] env[61957]: DEBUG nova.virt.hardware [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 654.090315] env[61957]: DEBUG nova.virt.hardware [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 654.090315] env[61957]: DEBUG nova.virt.hardware [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 654.090416] env[61957]: DEBUG nova.virt.hardware [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] Got 1 possible topologies {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 654.090760] env[61957]: DEBUG nova.virt.hardware [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 654.090955] env[61957]: DEBUG nova.virt.hardware [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 654.092226] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-533a48f8-6a2f-4361-bf76-c881ecc757d3 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.103678] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-199bf4ca-c527-4dae-a17f-3b60482a3d69 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.119648] env[61957]: DEBUG oslo_concurrency.lockutils [None req-adf5e3c1-3919-48f3-8d1a-68a60974255e tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.097s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 654.119648] env[61957]: ERROR nova.compute.manager [None req-adf5e3c1-3919-48f3-8d1a-68a60974255e tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] [instance: 3c099888-d5dc-4cee-9729-311808694625] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port cd92211b-fb72-4034-a229-b18203e54c1e, please check neutron logs for more information. [ 654.119648] env[61957]: ERROR nova.compute.manager [instance: 3c099888-d5dc-4cee-9729-311808694625] Traceback (most recent call last): [ 654.119648] env[61957]: ERROR nova.compute.manager [instance: 3c099888-d5dc-4cee-9729-311808694625] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 654.119648] env[61957]: ERROR nova.compute.manager [instance: 3c099888-d5dc-4cee-9729-311808694625] self.driver.spawn(context, instance, image_meta, [ 654.119648] env[61957]: ERROR nova.compute.manager [instance: 3c099888-d5dc-4cee-9729-311808694625] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 654.119648] env[61957]: ERROR nova.compute.manager [instance: 3c099888-d5dc-4cee-9729-311808694625] self._vmops.spawn(context, instance, image_meta, injected_files, [ 654.119648] env[61957]: ERROR nova.compute.manager [instance: 3c099888-d5dc-4cee-9729-311808694625] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 654.119648] env[61957]: ERROR nova.compute.manager [instance: 3c099888-d5dc-4cee-9729-311808694625] vm_ref = self.build_virtual_machine(instance, [ 654.119890] env[61957]: ERROR nova.compute.manager [instance: 3c099888-d5dc-4cee-9729-311808694625] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 654.119890] env[61957]: ERROR nova.compute.manager [instance: 3c099888-d5dc-4cee-9729-311808694625] vif_infos = vmwarevif.get_vif_info(self._session, [ 654.119890] env[61957]: ERROR nova.compute.manager [instance: 3c099888-d5dc-4cee-9729-311808694625] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 654.119890] env[61957]: ERROR nova.compute.manager [instance: 3c099888-d5dc-4cee-9729-311808694625] for vif in network_info: [ 654.119890] env[61957]: ERROR nova.compute.manager [instance: 3c099888-d5dc-4cee-9729-311808694625] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 654.119890] env[61957]: ERROR nova.compute.manager [instance: 3c099888-d5dc-4cee-9729-311808694625] return self._sync_wrapper(fn, *args, **kwargs) [ 654.119890] env[61957]: ERROR nova.compute.manager [instance: 3c099888-d5dc-4cee-9729-311808694625] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 654.119890] env[61957]: ERROR nova.compute.manager [instance: 3c099888-d5dc-4cee-9729-311808694625] self.wait() [ 654.119890] env[61957]: ERROR nova.compute.manager [instance: 3c099888-d5dc-4cee-9729-311808694625] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 654.119890] env[61957]: ERROR nova.compute.manager [instance: 3c099888-d5dc-4cee-9729-311808694625] self[:] = self._gt.wait() [ 654.119890] env[61957]: ERROR nova.compute.manager [instance: 3c099888-d5dc-4cee-9729-311808694625] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 654.119890] env[61957]: ERROR nova.compute.manager [instance: 3c099888-d5dc-4cee-9729-311808694625] return self._exit_event.wait() [ 654.119890] env[61957]: ERROR nova.compute.manager [instance: 3c099888-d5dc-4cee-9729-311808694625] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 654.120268] env[61957]: ERROR nova.compute.manager [instance: 3c099888-d5dc-4cee-9729-311808694625] result = hub.switch() [ 654.120268] env[61957]: ERROR nova.compute.manager [instance: 3c099888-d5dc-4cee-9729-311808694625] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 654.120268] env[61957]: ERROR nova.compute.manager [instance: 3c099888-d5dc-4cee-9729-311808694625] return self.greenlet.switch() [ 654.120268] env[61957]: ERROR nova.compute.manager [instance: 3c099888-d5dc-4cee-9729-311808694625] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 654.120268] env[61957]: ERROR nova.compute.manager [instance: 3c099888-d5dc-4cee-9729-311808694625] result = function(*args, **kwargs) [ 654.120268] env[61957]: ERROR nova.compute.manager [instance: 3c099888-d5dc-4cee-9729-311808694625] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 654.120268] env[61957]: ERROR nova.compute.manager [instance: 3c099888-d5dc-4cee-9729-311808694625] return func(*args, **kwargs) [ 654.120268] env[61957]: ERROR nova.compute.manager [instance: 3c099888-d5dc-4cee-9729-311808694625] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 654.120268] env[61957]: ERROR nova.compute.manager [instance: 3c099888-d5dc-4cee-9729-311808694625] raise e [ 654.120268] env[61957]: ERROR nova.compute.manager [instance: 3c099888-d5dc-4cee-9729-311808694625] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 654.120268] env[61957]: ERROR nova.compute.manager [instance: 3c099888-d5dc-4cee-9729-311808694625] nwinfo = self.network_api.allocate_for_instance( [ 654.120268] env[61957]: ERROR nova.compute.manager [instance: 3c099888-d5dc-4cee-9729-311808694625] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 654.120268] env[61957]: ERROR nova.compute.manager [instance: 3c099888-d5dc-4cee-9729-311808694625] created_port_ids = self._update_ports_for_instance( [ 654.120867] env[61957]: ERROR nova.compute.manager [instance: 3c099888-d5dc-4cee-9729-311808694625] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 654.120867] env[61957]: ERROR nova.compute.manager [instance: 3c099888-d5dc-4cee-9729-311808694625] with excutils.save_and_reraise_exception(): [ 654.120867] env[61957]: ERROR nova.compute.manager [instance: 3c099888-d5dc-4cee-9729-311808694625] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 654.120867] env[61957]: ERROR nova.compute.manager [instance: 3c099888-d5dc-4cee-9729-311808694625] self.force_reraise() [ 654.120867] env[61957]: ERROR nova.compute.manager [instance: 3c099888-d5dc-4cee-9729-311808694625] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 654.120867] env[61957]: ERROR nova.compute.manager [instance: 3c099888-d5dc-4cee-9729-311808694625] raise self.value [ 654.120867] env[61957]: ERROR nova.compute.manager [instance: 3c099888-d5dc-4cee-9729-311808694625] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 654.120867] env[61957]: ERROR nova.compute.manager [instance: 3c099888-d5dc-4cee-9729-311808694625] updated_port = self._update_port( [ 654.120867] env[61957]: ERROR nova.compute.manager [instance: 3c099888-d5dc-4cee-9729-311808694625] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 654.120867] env[61957]: ERROR nova.compute.manager [instance: 3c099888-d5dc-4cee-9729-311808694625] _ensure_no_port_binding_failure(port) [ 654.120867] env[61957]: ERROR nova.compute.manager [instance: 3c099888-d5dc-4cee-9729-311808694625] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 654.120867] env[61957]: ERROR nova.compute.manager [instance: 3c099888-d5dc-4cee-9729-311808694625] raise exception.PortBindingFailed(port_id=port['id']) [ 654.121187] env[61957]: ERROR nova.compute.manager [instance: 3c099888-d5dc-4cee-9729-311808694625] nova.exception.PortBindingFailed: Binding failed for port cd92211b-fb72-4034-a229-b18203e54c1e, please check neutron logs for more information. [ 654.121187] env[61957]: ERROR nova.compute.manager [instance: 3c099888-d5dc-4cee-9729-311808694625] [ 654.121187] env[61957]: DEBUG nova.compute.utils [None req-adf5e3c1-3919-48f3-8d1a-68a60974255e tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] [instance: 3c099888-d5dc-4cee-9729-311808694625] Binding failed for port cd92211b-fb72-4034-a229-b18203e54c1e, please check neutron logs for more information. {{(pid=61957) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 654.121876] env[61957]: DEBUG oslo_concurrency.lockutils [None req-4e7f7a28-6560-4b73-affe-356aa8d11b0b tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.080s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 654.124976] env[61957]: DEBUG nova.compute.manager [None req-adf5e3c1-3919-48f3-8d1a-68a60974255e tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] [instance: 3c099888-d5dc-4cee-9729-311808694625] Build of instance 3c099888-d5dc-4cee-9729-311808694625 was re-scheduled: Binding failed for port cd92211b-fb72-4034-a229-b18203e54c1e, please check neutron logs for more information. {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 654.125695] env[61957]: DEBUG nova.compute.manager [None req-adf5e3c1-3919-48f3-8d1a-68a60974255e tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] [instance: 3c099888-d5dc-4cee-9729-311808694625] Unplugging VIFs for instance {{(pid=61957) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 654.125695] env[61957]: DEBUG oslo_concurrency.lockutils [None req-adf5e3c1-3919-48f3-8d1a-68a60974255e tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] Acquiring lock "refresh_cache-3c099888-d5dc-4cee-9729-311808694625" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 654.125812] env[61957]: DEBUG oslo_concurrency.lockutils [None req-adf5e3c1-3919-48f3-8d1a-68a60974255e tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] Acquired lock "refresh_cache-3c099888-d5dc-4cee-9729-311808694625" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 654.126113] env[61957]: DEBUG nova.network.neutron [None req-adf5e3c1-3919-48f3-8d1a-68a60974255e tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] [instance: 3c099888-d5dc-4cee-9729-311808694625] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 654.160712] env[61957]: ERROR nova.compute.manager [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port ebb67595-2f07-4312-a3d9-11a91467c152, please check neutron logs for more information. [ 654.160712] env[61957]: ERROR nova.compute.manager Traceback (most recent call last): [ 654.160712] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 654.160712] env[61957]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 654.160712] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 654.160712] env[61957]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 654.160712] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 654.160712] env[61957]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 654.160712] env[61957]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 654.160712] env[61957]: ERROR nova.compute.manager self.force_reraise() [ 654.160712] env[61957]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 654.160712] env[61957]: ERROR nova.compute.manager raise self.value [ 654.160712] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 654.160712] env[61957]: ERROR nova.compute.manager updated_port = self._update_port( [ 654.160712] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 654.160712] env[61957]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 654.161300] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 654.161300] env[61957]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 654.161300] env[61957]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port ebb67595-2f07-4312-a3d9-11a91467c152, please check neutron logs for more information. [ 654.161300] env[61957]: ERROR nova.compute.manager [ 654.161300] env[61957]: Traceback (most recent call last): [ 654.161300] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 654.161300] env[61957]: listener.cb(fileno) [ 654.161300] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 654.161300] env[61957]: result = function(*args, **kwargs) [ 654.161300] env[61957]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 654.161300] env[61957]: return func(*args, **kwargs) [ 654.161300] env[61957]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 654.161300] env[61957]: raise e [ 654.161300] env[61957]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 654.161300] env[61957]: nwinfo = self.network_api.allocate_for_instance( [ 654.161300] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 654.161300] env[61957]: created_port_ids = self._update_ports_for_instance( [ 654.161300] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 654.161300] env[61957]: with excutils.save_and_reraise_exception(): [ 654.161300] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 654.161300] env[61957]: self.force_reraise() [ 654.161300] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 654.161300] env[61957]: raise self.value [ 654.161300] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 654.161300] env[61957]: updated_port = self._update_port( [ 654.161300] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 654.161300] env[61957]: _ensure_no_port_binding_failure(port) [ 654.161300] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 654.161300] env[61957]: raise exception.PortBindingFailed(port_id=port['id']) [ 654.161978] env[61957]: nova.exception.PortBindingFailed: Binding failed for port ebb67595-2f07-4312-a3d9-11a91467c152, please check neutron logs for more information. [ 654.161978] env[61957]: Removing descriptor: 15 [ 654.162345] env[61957]: ERROR nova.compute.manager [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port ebb67595-2f07-4312-a3d9-11a91467c152, please check neutron logs for more information. [ 654.162345] env[61957]: ERROR nova.compute.manager [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] Traceback (most recent call last): [ 654.162345] env[61957]: ERROR nova.compute.manager [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 654.162345] env[61957]: ERROR nova.compute.manager [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] yield resources [ 654.162345] env[61957]: ERROR nova.compute.manager [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 654.162345] env[61957]: ERROR nova.compute.manager [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] self.driver.spawn(context, instance, image_meta, [ 654.162345] env[61957]: ERROR nova.compute.manager [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 654.162345] env[61957]: ERROR nova.compute.manager [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] self._vmops.spawn(context, instance, image_meta, injected_files, [ 654.162345] env[61957]: ERROR nova.compute.manager [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 654.162345] env[61957]: ERROR nova.compute.manager [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] vm_ref = self.build_virtual_machine(instance, [ 654.162345] env[61957]: ERROR nova.compute.manager [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 654.162674] env[61957]: ERROR nova.compute.manager [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] vif_infos = vmwarevif.get_vif_info(self._session, [ 654.162674] env[61957]: ERROR nova.compute.manager [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 654.162674] env[61957]: ERROR nova.compute.manager [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] for vif in network_info: [ 654.162674] env[61957]: ERROR nova.compute.manager [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 654.162674] env[61957]: ERROR nova.compute.manager [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] return self._sync_wrapper(fn, *args, **kwargs) [ 654.162674] env[61957]: ERROR nova.compute.manager [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 654.162674] env[61957]: ERROR nova.compute.manager [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] self.wait() [ 654.162674] env[61957]: ERROR nova.compute.manager [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 654.162674] env[61957]: ERROR nova.compute.manager [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] self[:] = self._gt.wait() [ 654.162674] env[61957]: ERROR nova.compute.manager [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 654.162674] env[61957]: ERROR nova.compute.manager [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] return self._exit_event.wait() [ 654.162674] env[61957]: ERROR nova.compute.manager [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 654.162674] env[61957]: ERROR nova.compute.manager [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] result = hub.switch() [ 654.162985] env[61957]: ERROR nova.compute.manager [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 654.162985] env[61957]: ERROR nova.compute.manager [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] return self.greenlet.switch() [ 654.162985] env[61957]: ERROR nova.compute.manager [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 654.162985] env[61957]: ERROR nova.compute.manager [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] result = function(*args, **kwargs) [ 654.162985] env[61957]: ERROR nova.compute.manager [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 654.162985] env[61957]: ERROR nova.compute.manager [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] return func(*args, **kwargs) [ 654.162985] env[61957]: ERROR nova.compute.manager [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 654.162985] env[61957]: ERROR nova.compute.manager [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] raise e [ 654.162985] env[61957]: ERROR nova.compute.manager [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 654.162985] env[61957]: ERROR nova.compute.manager [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] nwinfo = self.network_api.allocate_for_instance( [ 654.162985] env[61957]: ERROR nova.compute.manager [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 654.162985] env[61957]: ERROR nova.compute.manager [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] created_port_ids = self._update_ports_for_instance( [ 654.162985] env[61957]: ERROR nova.compute.manager [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 654.164459] env[61957]: ERROR nova.compute.manager [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] with excutils.save_and_reraise_exception(): [ 654.164459] env[61957]: ERROR nova.compute.manager [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 654.164459] env[61957]: ERROR nova.compute.manager [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] self.force_reraise() [ 654.164459] env[61957]: ERROR nova.compute.manager [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 654.164459] env[61957]: ERROR nova.compute.manager [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] raise self.value [ 654.164459] env[61957]: ERROR nova.compute.manager [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 654.164459] env[61957]: ERROR nova.compute.manager [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] updated_port = self._update_port( [ 654.164459] env[61957]: ERROR nova.compute.manager [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 654.164459] env[61957]: ERROR nova.compute.manager [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] _ensure_no_port_binding_failure(port) [ 654.164459] env[61957]: ERROR nova.compute.manager [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 654.164459] env[61957]: ERROR nova.compute.manager [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] raise exception.PortBindingFailed(port_id=port['id']) [ 654.164459] env[61957]: ERROR nova.compute.manager [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] nova.exception.PortBindingFailed: Binding failed for port ebb67595-2f07-4312-a3d9-11a91467c152, please check neutron logs for more information. [ 654.164459] env[61957]: ERROR nova.compute.manager [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] [ 654.164807] env[61957]: INFO nova.compute.manager [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] Terminating instance [ 654.165360] env[61957]: DEBUG oslo_concurrency.lockutils [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] Acquiring lock "refresh_cache-b541c437-93b7-4e4d-9ff8-bfc9cdf0e237" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 654.297502] env[61957]: DEBUG nova.network.neutron [req-2ee6ddb6-4311-459a-9474-261c26208567 req-495c8ba7-1c66-417b-95d3-8c1b8f85e564 service nova] [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 654.625855] env[61957]: DEBUG oslo_concurrency.lockutils [None req-bb8cb37b-74c5-4607-9ba8-e5615134a1da tempest-InstanceActionsTestJSON-565478614 tempest-InstanceActionsTestJSON-565478614-project-member] Acquiring lock "2f84d2df-00f0-4860-acb4-4e10ecf8ce4e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 654.626162] env[61957]: DEBUG oslo_concurrency.lockutils [None req-bb8cb37b-74c5-4607-9ba8-e5615134a1da tempest-InstanceActionsTestJSON-565478614 tempest-InstanceActionsTestJSON-565478614-project-member] Lock "2f84d2df-00f0-4860-acb4-4e10ecf8ce4e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 654.668627] env[61957]: DEBUG nova.network.neutron [None req-adf5e3c1-3919-48f3-8d1a-68a60974255e tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] [instance: 3c099888-d5dc-4cee-9729-311808694625] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 654.799601] env[61957]: DEBUG oslo_concurrency.lockutils [req-2ee6ddb6-4311-459a-9474-261c26208567 req-495c8ba7-1c66-417b-95d3-8c1b8f85e564 service nova] Releasing lock "refresh_cache-b541c437-93b7-4e4d-9ff8-bfc9cdf0e237" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 654.800125] env[61957]: DEBUG oslo_concurrency.lockutils [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] Acquired lock "refresh_cache-b541c437-93b7-4e4d-9ff8-bfc9cdf0e237" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 654.800316] env[61957]: DEBUG nova.network.neutron [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 654.818293] env[61957]: DEBUG nova.network.neutron [None req-adf5e3c1-3919-48f3-8d1a-68a60974255e tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] [instance: 3c099888-d5dc-4cee-9729-311808694625] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 655.230074] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16662fdd-3d4c-45e3-b662-4307d5838dbb {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.240196] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8e54ddb-3003-4749-8a18-d801c363511a {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.273068] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97bf5e50-dfa7-4602-a54d-d7407d9c5dbb {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.280980] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c9655a5-268b-4af2-bdcf-c99af5a7e718 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.294354] env[61957]: DEBUG nova.compute.provider_tree [None req-4e7f7a28-6560-4b73-affe-356aa8d11b0b tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 655.325592] env[61957]: DEBUG oslo_concurrency.lockutils [None req-adf5e3c1-3919-48f3-8d1a-68a60974255e tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] Releasing lock "refresh_cache-3c099888-d5dc-4cee-9729-311808694625" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 655.326247] env[61957]: DEBUG nova.compute.manager [None req-adf5e3c1-3919-48f3-8d1a-68a60974255e tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61957) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 655.326446] env[61957]: DEBUG nova.compute.manager [None req-adf5e3c1-3919-48f3-8d1a-68a60974255e tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] [instance: 3c099888-d5dc-4cee-9729-311808694625] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 655.326608] env[61957]: DEBUG nova.network.neutron [None req-adf5e3c1-3919-48f3-8d1a-68a60974255e tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] [instance: 3c099888-d5dc-4cee-9729-311808694625] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 655.329433] env[61957]: DEBUG nova.network.neutron [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 655.352813] env[61957]: DEBUG nova.network.neutron [None req-adf5e3c1-3919-48f3-8d1a-68a60974255e tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] [instance: 3c099888-d5dc-4cee-9729-311808694625] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 655.465912] env[61957]: DEBUG nova.network.neutron [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 655.468898] env[61957]: ERROR nova.compute.manager [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 10ac5929-6235-4c76-b38b-45cc8b90ca43, please check neutron logs for more information. [ 655.468898] env[61957]: ERROR nova.compute.manager Traceback (most recent call last): [ 655.468898] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 655.468898] env[61957]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 655.468898] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 655.468898] env[61957]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 655.468898] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 655.468898] env[61957]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 655.468898] env[61957]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 655.468898] env[61957]: ERROR nova.compute.manager self.force_reraise() [ 655.468898] env[61957]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 655.468898] env[61957]: ERROR nova.compute.manager raise self.value [ 655.468898] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 655.468898] env[61957]: ERROR nova.compute.manager updated_port = self._update_port( [ 655.468898] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 655.468898] env[61957]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 655.469333] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 655.469333] env[61957]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 655.469333] env[61957]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 10ac5929-6235-4c76-b38b-45cc8b90ca43, please check neutron logs for more information. [ 655.469333] env[61957]: ERROR nova.compute.manager [ 655.469333] env[61957]: Traceback (most recent call last): [ 655.469333] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 655.469333] env[61957]: listener.cb(fileno) [ 655.469333] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 655.469333] env[61957]: result = function(*args, **kwargs) [ 655.469333] env[61957]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 655.469333] env[61957]: return func(*args, **kwargs) [ 655.469333] env[61957]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 655.469333] env[61957]: raise e [ 655.469333] env[61957]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 655.469333] env[61957]: nwinfo = self.network_api.allocate_for_instance( [ 655.469333] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 655.469333] env[61957]: created_port_ids = self._update_ports_for_instance( [ 655.469333] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 655.469333] env[61957]: with excutils.save_and_reraise_exception(): [ 655.469333] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 655.469333] env[61957]: self.force_reraise() [ 655.469333] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 655.469333] env[61957]: raise self.value [ 655.469333] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 655.469333] env[61957]: updated_port = self._update_port( [ 655.469333] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 655.469333] env[61957]: _ensure_no_port_binding_failure(port) [ 655.469333] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 655.469333] env[61957]: raise exception.PortBindingFailed(port_id=port['id']) [ 655.470015] env[61957]: nova.exception.PortBindingFailed: Binding failed for port 10ac5929-6235-4c76-b38b-45cc8b90ca43, please check neutron logs for more information. [ 655.470015] env[61957]: Removing descriptor: 17 [ 655.470015] env[61957]: ERROR nova.compute.manager [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] [instance: 394b1955-c448-42c2-a718-28df7bd366e9] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 10ac5929-6235-4c76-b38b-45cc8b90ca43, please check neutron logs for more information. [ 655.470015] env[61957]: ERROR nova.compute.manager [instance: 394b1955-c448-42c2-a718-28df7bd366e9] Traceback (most recent call last): [ 655.470015] env[61957]: ERROR nova.compute.manager [instance: 394b1955-c448-42c2-a718-28df7bd366e9] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 655.470015] env[61957]: ERROR nova.compute.manager [instance: 394b1955-c448-42c2-a718-28df7bd366e9] yield resources [ 655.470015] env[61957]: ERROR nova.compute.manager [instance: 394b1955-c448-42c2-a718-28df7bd366e9] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 655.470015] env[61957]: ERROR nova.compute.manager [instance: 394b1955-c448-42c2-a718-28df7bd366e9] self.driver.spawn(context, instance, image_meta, [ 655.470015] env[61957]: ERROR nova.compute.manager [instance: 394b1955-c448-42c2-a718-28df7bd366e9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 655.470015] env[61957]: ERROR nova.compute.manager [instance: 394b1955-c448-42c2-a718-28df7bd366e9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 655.470015] env[61957]: ERROR nova.compute.manager [instance: 394b1955-c448-42c2-a718-28df7bd366e9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 655.470015] env[61957]: ERROR nova.compute.manager [instance: 394b1955-c448-42c2-a718-28df7bd366e9] vm_ref = self.build_virtual_machine(instance, [ 655.470289] env[61957]: ERROR nova.compute.manager [instance: 394b1955-c448-42c2-a718-28df7bd366e9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 655.470289] env[61957]: ERROR nova.compute.manager [instance: 394b1955-c448-42c2-a718-28df7bd366e9] vif_infos = vmwarevif.get_vif_info(self._session, [ 655.470289] env[61957]: ERROR nova.compute.manager [instance: 394b1955-c448-42c2-a718-28df7bd366e9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 655.470289] env[61957]: ERROR nova.compute.manager [instance: 394b1955-c448-42c2-a718-28df7bd366e9] for vif in network_info: [ 655.470289] env[61957]: ERROR nova.compute.manager [instance: 394b1955-c448-42c2-a718-28df7bd366e9] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 655.470289] env[61957]: ERROR nova.compute.manager [instance: 394b1955-c448-42c2-a718-28df7bd366e9] return self._sync_wrapper(fn, *args, **kwargs) [ 655.470289] env[61957]: ERROR nova.compute.manager [instance: 394b1955-c448-42c2-a718-28df7bd366e9] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 655.470289] env[61957]: ERROR nova.compute.manager [instance: 394b1955-c448-42c2-a718-28df7bd366e9] self.wait() [ 655.470289] env[61957]: ERROR nova.compute.manager [instance: 394b1955-c448-42c2-a718-28df7bd366e9] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 655.470289] env[61957]: ERROR nova.compute.manager [instance: 394b1955-c448-42c2-a718-28df7bd366e9] self[:] = self._gt.wait() [ 655.470289] env[61957]: ERROR nova.compute.manager [instance: 394b1955-c448-42c2-a718-28df7bd366e9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 655.470289] env[61957]: ERROR nova.compute.manager [instance: 394b1955-c448-42c2-a718-28df7bd366e9] return self._exit_event.wait() [ 655.470289] env[61957]: ERROR nova.compute.manager [instance: 394b1955-c448-42c2-a718-28df7bd366e9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 655.470562] env[61957]: ERROR nova.compute.manager [instance: 394b1955-c448-42c2-a718-28df7bd366e9] result = hub.switch() [ 655.470562] env[61957]: ERROR nova.compute.manager [instance: 394b1955-c448-42c2-a718-28df7bd366e9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 655.470562] env[61957]: ERROR nova.compute.manager [instance: 394b1955-c448-42c2-a718-28df7bd366e9] return self.greenlet.switch() [ 655.470562] env[61957]: ERROR nova.compute.manager [instance: 394b1955-c448-42c2-a718-28df7bd366e9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 655.470562] env[61957]: ERROR nova.compute.manager [instance: 394b1955-c448-42c2-a718-28df7bd366e9] result = function(*args, **kwargs) [ 655.470562] env[61957]: ERROR nova.compute.manager [instance: 394b1955-c448-42c2-a718-28df7bd366e9] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 655.470562] env[61957]: ERROR nova.compute.manager [instance: 394b1955-c448-42c2-a718-28df7bd366e9] return func(*args, **kwargs) [ 655.470562] env[61957]: ERROR nova.compute.manager [instance: 394b1955-c448-42c2-a718-28df7bd366e9] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 655.470562] env[61957]: ERROR nova.compute.manager [instance: 394b1955-c448-42c2-a718-28df7bd366e9] raise e [ 655.470562] env[61957]: ERROR nova.compute.manager [instance: 394b1955-c448-42c2-a718-28df7bd366e9] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 655.470562] env[61957]: ERROR nova.compute.manager [instance: 394b1955-c448-42c2-a718-28df7bd366e9] nwinfo = self.network_api.allocate_for_instance( [ 655.470562] env[61957]: ERROR nova.compute.manager [instance: 394b1955-c448-42c2-a718-28df7bd366e9] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 655.470562] env[61957]: ERROR nova.compute.manager [instance: 394b1955-c448-42c2-a718-28df7bd366e9] created_port_ids = self._update_ports_for_instance( [ 655.470971] env[61957]: ERROR nova.compute.manager [instance: 394b1955-c448-42c2-a718-28df7bd366e9] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 655.470971] env[61957]: ERROR nova.compute.manager [instance: 394b1955-c448-42c2-a718-28df7bd366e9] with excutils.save_and_reraise_exception(): [ 655.470971] env[61957]: ERROR nova.compute.manager [instance: 394b1955-c448-42c2-a718-28df7bd366e9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 655.470971] env[61957]: ERROR nova.compute.manager [instance: 394b1955-c448-42c2-a718-28df7bd366e9] self.force_reraise() [ 655.470971] env[61957]: ERROR nova.compute.manager [instance: 394b1955-c448-42c2-a718-28df7bd366e9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 655.470971] env[61957]: ERROR nova.compute.manager [instance: 394b1955-c448-42c2-a718-28df7bd366e9] raise self.value [ 655.470971] env[61957]: ERROR nova.compute.manager [instance: 394b1955-c448-42c2-a718-28df7bd366e9] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 655.470971] env[61957]: ERROR nova.compute.manager [instance: 394b1955-c448-42c2-a718-28df7bd366e9] updated_port = self._update_port( [ 655.470971] env[61957]: ERROR nova.compute.manager [instance: 394b1955-c448-42c2-a718-28df7bd366e9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 655.470971] env[61957]: ERROR nova.compute.manager [instance: 394b1955-c448-42c2-a718-28df7bd366e9] _ensure_no_port_binding_failure(port) [ 655.470971] env[61957]: ERROR nova.compute.manager [instance: 394b1955-c448-42c2-a718-28df7bd366e9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 655.470971] env[61957]: ERROR nova.compute.manager [instance: 394b1955-c448-42c2-a718-28df7bd366e9] raise exception.PortBindingFailed(port_id=port['id']) [ 655.471251] env[61957]: ERROR nova.compute.manager [instance: 394b1955-c448-42c2-a718-28df7bd366e9] nova.exception.PortBindingFailed: Binding failed for port 10ac5929-6235-4c76-b38b-45cc8b90ca43, please check neutron logs for more information. [ 655.471251] env[61957]: ERROR nova.compute.manager [instance: 394b1955-c448-42c2-a718-28df7bd366e9] [ 655.471251] env[61957]: INFO nova.compute.manager [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] [instance: 394b1955-c448-42c2-a718-28df7bd366e9] Terminating instance [ 655.472264] env[61957]: DEBUG oslo_concurrency.lockutils [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] Acquiring lock "refresh_cache-394b1955-c448-42c2-a718-28df7bd366e9" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 655.472422] env[61957]: DEBUG oslo_concurrency.lockutils [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] Acquired lock "refresh_cache-394b1955-c448-42c2-a718-28df7bd366e9" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 655.472612] env[61957]: DEBUG nova.network.neutron [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] [instance: 394b1955-c448-42c2-a718-28df7bd366e9] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 655.797543] env[61957]: DEBUG nova.scheduler.client.report [None req-4e7f7a28-6560-4b73-affe-356aa8d11b0b tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 655.840459] env[61957]: DEBUG nova.compute.manager [req-80789b21-fb9d-4736-9088-c66739fe521e req-f864c0e8-73d8-40ee-9134-813e2db26fb6 service nova] [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] Received event network-vif-deleted-ebb67595-2f07-4312-a3d9-11a91467c152 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 655.840459] env[61957]: DEBUG nova.compute.manager [req-80789b21-fb9d-4736-9088-c66739fe521e req-f864c0e8-73d8-40ee-9134-813e2db26fb6 service nova] [instance: 394b1955-c448-42c2-a718-28df7bd366e9] Received event network-changed-10ac5929-6235-4c76-b38b-45cc8b90ca43 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 655.840459] env[61957]: DEBUG nova.compute.manager [req-80789b21-fb9d-4736-9088-c66739fe521e req-f864c0e8-73d8-40ee-9134-813e2db26fb6 service nova] [instance: 394b1955-c448-42c2-a718-28df7bd366e9] Refreshing instance network info cache due to event network-changed-10ac5929-6235-4c76-b38b-45cc8b90ca43. {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 655.840562] env[61957]: DEBUG oslo_concurrency.lockutils [req-80789b21-fb9d-4736-9088-c66739fe521e req-f864c0e8-73d8-40ee-9134-813e2db26fb6 service nova] Acquiring lock "refresh_cache-394b1955-c448-42c2-a718-28df7bd366e9" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 655.856877] env[61957]: DEBUG nova.network.neutron [None req-adf5e3c1-3919-48f3-8d1a-68a60974255e tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] [instance: 3c099888-d5dc-4cee-9729-311808694625] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 655.967785] env[61957]: DEBUG oslo_concurrency.lockutils [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] Releasing lock "refresh_cache-b541c437-93b7-4e4d-9ff8-bfc9cdf0e237" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 655.968269] env[61957]: DEBUG nova.compute.manager [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] Start destroying the instance on the hypervisor. {{(pid=61957) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 655.968461] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] Destroying instance {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 655.969159] env[61957]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5d9adc4b-a139-478d-a05f-09e5e8016e0c {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.979881] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dca4e631-38cf-47f9-84f2-7cbee50f9950 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.993322] env[61957]: DEBUG nova.network.neutron [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] [instance: 394b1955-c448-42c2-a718-28df7bd366e9] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 656.002435] env[61957]: WARNING nova.virt.vmwareapi.vmops [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance b541c437-93b7-4e4d-9ff8-bfc9cdf0e237 could not be found. [ 656.002521] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] Instance destroyed {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 656.002635] env[61957]: INFO nova.compute.manager [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] Took 0.03 seconds to destroy the instance on the hypervisor. [ 656.003857] env[61957]: DEBUG oslo.service.loopingcall [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 656.003857] env[61957]: DEBUG nova.compute.manager [-] [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 656.003857] env[61957]: DEBUG nova.network.neutron [-] [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 656.025068] env[61957]: DEBUG nova.network.neutron [-] [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 656.085976] env[61957]: DEBUG nova.network.neutron [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] [instance: 394b1955-c448-42c2-a718-28df7bd366e9] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 656.244677] env[61957]: DEBUG oslo_concurrency.lockutils [None req-7860e034-eff8-489c-addf-33cf2776246c tempest-ServerActionsTestOtherA-1610268163 tempest-ServerActionsTestOtherA-1610268163-project-member] Acquiring lock "6372a384-61a8-43e8-8f6c-376e2a735045" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 656.244935] env[61957]: DEBUG oslo_concurrency.lockutils [None req-7860e034-eff8-489c-addf-33cf2776246c tempest-ServerActionsTestOtherA-1610268163 tempest-ServerActionsTestOtherA-1610268163-project-member] Lock "6372a384-61a8-43e8-8f6c-376e2a735045" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 656.306467] env[61957]: DEBUG oslo_concurrency.lockutils [None req-4e7f7a28-6560-4b73-affe-356aa8d11b0b tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.184s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 656.307135] env[61957]: ERROR nova.compute.manager [None req-4e7f7a28-6560-4b73-affe-356aa8d11b0b tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 09ed1ca2-59f3-48fa-b1a0-4d61495c3c4c, please check neutron logs for more information. [ 656.307135] env[61957]: ERROR nova.compute.manager [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] Traceback (most recent call last): [ 656.307135] env[61957]: ERROR nova.compute.manager [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 656.307135] env[61957]: ERROR nova.compute.manager [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] self.driver.spawn(context, instance, image_meta, [ 656.307135] env[61957]: ERROR nova.compute.manager [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 656.307135] env[61957]: ERROR nova.compute.manager [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] self._vmops.spawn(context, instance, image_meta, injected_files, [ 656.307135] env[61957]: ERROR nova.compute.manager [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 656.307135] env[61957]: ERROR nova.compute.manager [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] vm_ref = self.build_virtual_machine(instance, [ 656.307135] env[61957]: ERROR nova.compute.manager [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 656.307135] env[61957]: ERROR nova.compute.manager [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] vif_infos = vmwarevif.get_vif_info(self._session, [ 656.307135] env[61957]: ERROR nova.compute.manager [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 656.307400] env[61957]: ERROR nova.compute.manager [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] for vif in network_info: [ 656.307400] env[61957]: ERROR nova.compute.manager [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 656.307400] env[61957]: ERROR nova.compute.manager [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] return self._sync_wrapper(fn, *args, **kwargs) [ 656.307400] env[61957]: ERROR nova.compute.manager [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 656.307400] env[61957]: ERROR nova.compute.manager [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] self.wait() [ 656.307400] env[61957]: ERROR nova.compute.manager [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 656.307400] env[61957]: ERROR nova.compute.manager [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] self[:] = self._gt.wait() [ 656.307400] env[61957]: ERROR nova.compute.manager [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 656.307400] env[61957]: ERROR nova.compute.manager [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] return self._exit_event.wait() [ 656.307400] env[61957]: ERROR nova.compute.manager [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 656.307400] env[61957]: ERROR nova.compute.manager [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] result = hub.switch() [ 656.307400] env[61957]: ERROR nova.compute.manager [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 656.307400] env[61957]: ERROR nova.compute.manager [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] return self.greenlet.switch() [ 656.307730] env[61957]: ERROR nova.compute.manager [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 656.307730] env[61957]: ERROR nova.compute.manager [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] result = function(*args, **kwargs) [ 656.307730] env[61957]: ERROR nova.compute.manager [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 656.307730] env[61957]: ERROR nova.compute.manager [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] return func(*args, **kwargs) [ 656.307730] env[61957]: ERROR nova.compute.manager [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 656.307730] env[61957]: ERROR nova.compute.manager [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] raise e [ 656.307730] env[61957]: ERROR nova.compute.manager [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 656.307730] env[61957]: ERROR nova.compute.manager [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] nwinfo = self.network_api.allocate_for_instance( [ 656.307730] env[61957]: ERROR nova.compute.manager [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 656.307730] env[61957]: ERROR nova.compute.manager [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] created_port_ids = self._update_ports_for_instance( [ 656.307730] env[61957]: ERROR nova.compute.manager [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 656.307730] env[61957]: ERROR nova.compute.manager [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] with excutils.save_and_reraise_exception(): [ 656.307730] env[61957]: ERROR nova.compute.manager [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 656.308018] env[61957]: ERROR nova.compute.manager [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] self.force_reraise() [ 656.308018] env[61957]: ERROR nova.compute.manager [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 656.308018] env[61957]: ERROR nova.compute.manager [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] raise self.value [ 656.308018] env[61957]: ERROR nova.compute.manager [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 656.308018] env[61957]: ERROR nova.compute.manager [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] updated_port = self._update_port( [ 656.308018] env[61957]: ERROR nova.compute.manager [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 656.308018] env[61957]: ERROR nova.compute.manager [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] _ensure_no_port_binding_failure(port) [ 656.308018] env[61957]: ERROR nova.compute.manager [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 656.308018] env[61957]: ERROR nova.compute.manager [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] raise exception.PortBindingFailed(port_id=port['id']) [ 656.308018] env[61957]: ERROR nova.compute.manager [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] nova.exception.PortBindingFailed: Binding failed for port 09ed1ca2-59f3-48fa-b1a0-4d61495c3c4c, please check neutron logs for more information. [ 656.308018] env[61957]: ERROR nova.compute.manager [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] [ 656.308266] env[61957]: DEBUG nova.compute.utils [None req-4e7f7a28-6560-4b73-affe-356aa8d11b0b tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] Binding failed for port 09ed1ca2-59f3-48fa-b1a0-4d61495c3c4c, please check neutron logs for more information. {{(pid=61957) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 656.309283] env[61957]: DEBUG oslo_concurrency.lockutils [None req-b1f62379-e574-4175-a099-f938af5ddaa8 tempest-ServerPasswordTestJSON-1543821420 tempest-ServerPasswordTestJSON-1543821420-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.128s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 656.310808] env[61957]: INFO nova.compute.claims [None req-b1f62379-e574-4175-a099-f938af5ddaa8 tempest-ServerPasswordTestJSON-1543821420 tempest-ServerPasswordTestJSON-1543821420-project-member] [instance: 49e99297-ffb9-4104-bd06-911243908828] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 656.317950] env[61957]: DEBUG nova.compute.manager [None req-4e7f7a28-6560-4b73-affe-356aa8d11b0b tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] Build of instance 763d87bb-652c-43e0-ba39-135bae2cc368 was re-scheduled: Binding failed for port 09ed1ca2-59f3-48fa-b1a0-4d61495c3c4c, please check neutron logs for more information. {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 656.318431] env[61957]: DEBUG nova.compute.manager [None req-4e7f7a28-6560-4b73-affe-356aa8d11b0b tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] Unplugging VIFs for instance {{(pid=61957) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 656.318653] env[61957]: DEBUG oslo_concurrency.lockutils [None req-4e7f7a28-6560-4b73-affe-356aa8d11b0b tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] Acquiring lock "refresh_cache-763d87bb-652c-43e0-ba39-135bae2cc368" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 656.318799] env[61957]: DEBUG oslo_concurrency.lockutils [None req-4e7f7a28-6560-4b73-affe-356aa8d11b0b tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] Acquired lock "refresh_cache-763d87bb-652c-43e0-ba39-135bae2cc368" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 656.318951] env[61957]: DEBUG nova.network.neutron [None req-4e7f7a28-6560-4b73-affe-356aa8d11b0b tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 656.358850] env[61957]: INFO nova.compute.manager [None req-adf5e3c1-3919-48f3-8d1a-68a60974255e tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] [instance: 3c099888-d5dc-4cee-9729-311808694625] Took 1.03 seconds to deallocate network for instance. [ 656.527842] env[61957]: DEBUG nova.network.neutron [-] [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 656.591943] env[61957]: DEBUG oslo_concurrency.lockutils [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] Releasing lock "refresh_cache-394b1955-c448-42c2-a718-28df7bd366e9" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 656.592453] env[61957]: DEBUG nova.compute.manager [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] [instance: 394b1955-c448-42c2-a718-28df7bd366e9] Start destroying the instance on the hypervisor. {{(pid=61957) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 656.593655] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] [instance: 394b1955-c448-42c2-a718-28df7bd366e9] Destroying instance {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 656.593655] env[61957]: DEBUG oslo_concurrency.lockutils [req-80789b21-fb9d-4736-9088-c66739fe521e req-f864c0e8-73d8-40ee-9134-813e2db26fb6 service nova] Acquired lock "refresh_cache-394b1955-c448-42c2-a718-28df7bd366e9" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 656.593655] env[61957]: DEBUG nova.network.neutron [req-80789b21-fb9d-4736-9088-c66739fe521e req-f864c0e8-73d8-40ee-9134-813e2db26fb6 service nova] [instance: 394b1955-c448-42c2-a718-28df7bd366e9] Refreshing network info cache for port 10ac5929-6235-4c76-b38b-45cc8b90ca43 {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 656.594299] env[61957]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-480210b5-8164-456d-a0db-61db5e4b8512 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.609193] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5f2f9ee-4a13-4d8d-8ced-5bea5b4a6274 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.632423] env[61957]: WARNING nova.virt.vmwareapi.vmops [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] [instance: 394b1955-c448-42c2-a718-28df7bd366e9] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 394b1955-c448-42c2-a718-28df7bd366e9 could not be found. [ 656.632423] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] [instance: 394b1955-c448-42c2-a718-28df7bd366e9] Instance destroyed {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 656.632575] env[61957]: INFO nova.compute.manager [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] [instance: 394b1955-c448-42c2-a718-28df7bd366e9] Took 0.04 seconds to destroy the instance on the hypervisor. [ 656.632738] env[61957]: DEBUG oslo.service.loopingcall [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 656.632952] env[61957]: DEBUG nova.compute.manager [-] [instance: 394b1955-c448-42c2-a718-28df7bd366e9] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 656.633051] env[61957]: DEBUG nova.network.neutron [-] [instance: 394b1955-c448-42c2-a718-28df7bd366e9] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 656.651653] env[61957]: DEBUG nova.network.neutron [-] [instance: 394b1955-c448-42c2-a718-28df7bd366e9] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 656.841851] env[61957]: DEBUG nova.network.neutron [None req-4e7f7a28-6560-4b73-affe-356aa8d11b0b tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 656.898067] env[61957]: DEBUG nova.network.neutron [None req-4e7f7a28-6560-4b73-affe-356aa8d11b0b tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 657.032291] env[61957]: INFO nova.compute.manager [-] [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] Took 1.03 seconds to deallocate network for instance. [ 657.034949] env[61957]: DEBUG nova.compute.claims [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] Aborting claim: {{(pid=61957) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 657.034949] env[61957]: DEBUG oslo_concurrency.lockutils [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 657.114215] env[61957]: DEBUG nova.network.neutron [req-80789b21-fb9d-4736-9088-c66739fe521e req-f864c0e8-73d8-40ee-9134-813e2db26fb6 service nova] [instance: 394b1955-c448-42c2-a718-28df7bd366e9] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 657.159871] env[61957]: DEBUG nova.network.neutron [-] [instance: 394b1955-c448-42c2-a718-28df7bd366e9] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 657.191706] env[61957]: DEBUG nova.network.neutron [req-80789b21-fb9d-4736-9088-c66739fe521e req-f864c0e8-73d8-40ee-9134-813e2db26fb6 service nova] [instance: 394b1955-c448-42c2-a718-28df7bd366e9] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 657.227024] env[61957]: DEBUG oslo_concurrency.lockutils [None req-79f56a2c-6af4-4a0f-b13f-0dc131681bf7 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Acquiring lock "b0af6294-f30c-4266-ae46-6fb03dc0cbeb" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 657.227256] env[61957]: DEBUG oslo_concurrency.lockutils [None req-79f56a2c-6af4-4a0f-b13f-0dc131681bf7 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Lock "b0af6294-f30c-4266-ae46-6fb03dc0cbeb" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 657.388244] env[61957]: INFO nova.scheduler.client.report [None req-adf5e3c1-3919-48f3-8d1a-68a60974255e tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] Deleted allocations for instance 3c099888-d5dc-4cee-9729-311808694625 [ 657.401231] env[61957]: DEBUG oslo_concurrency.lockutils [None req-4e7f7a28-6560-4b73-affe-356aa8d11b0b tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] Releasing lock "refresh_cache-763d87bb-652c-43e0-ba39-135bae2cc368" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 657.401323] env[61957]: DEBUG nova.compute.manager [None req-4e7f7a28-6560-4b73-affe-356aa8d11b0b tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61957) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 657.401420] env[61957]: DEBUG nova.compute.manager [None req-4e7f7a28-6560-4b73-affe-356aa8d11b0b tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 657.401939] env[61957]: DEBUG nova.network.neutron [None req-4e7f7a28-6560-4b73-affe-356aa8d11b0b tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 657.418989] env[61957]: DEBUG nova.network.neutron [None req-4e7f7a28-6560-4b73-affe-356aa8d11b0b tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 657.664670] env[61957]: INFO nova.compute.manager [-] [instance: 394b1955-c448-42c2-a718-28df7bd366e9] Took 1.03 seconds to deallocate network for instance. [ 657.667386] env[61957]: DEBUG nova.compute.claims [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] [instance: 394b1955-c448-42c2-a718-28df7bd366e9] Aborting claim: {{(pid=61957) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 657.667865] env[61957]: DEBUG oslo_concurrency.lockutils [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 657.694910] env[61957]: DEBUG oslo_concurrency.lockutils [req-80789b21-fb9d-4736-9088-c66739fe521e req-f864c0e8-73d8-40ee-9134-813e2db26fb6 service nova] Releasing lock "refresh_cache-394b1955-c448-42c2-a718-28df7bd366e9" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 657.694993] env[61957]: DEBUG nova.compute.manager [req-80789b21-fb9d-4736-9088-c66739fe521e req-f864c0e8-73d8-40ee-9134-813e2db26fb6 service nova] [instance: 394b1955-c448-42c2-a718-28df7bd366e9] Received event network-vif-deleted-10ac5929-6235-4c76-b38b-45cc8b90ca43 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 657.731530] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aeccf64f-3721-4da7-bd24-71d884b3d7dc {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.739217] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1cb8f425-60f3-4754-9505-e9ed7210f946 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.769414] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2fb9bd1-0cff-4e70-8c29-2cd4e6a219f5 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.776512] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bd95e8b-08a5-48b9-8b11-ff7c6a73baf1 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.790732] env[61957]: DEBUG nova.compute.provider_tree [None req-b1f62379-e574-4175-a099-f938af5ddaa8 tempest-ServerPasswordTestJSON-1543821420 tempest-ServerPasswordTestJSON-1543821420-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 657.896959] env[61957]: DEBUG oslo_concurrency.lockutils [None req-adf5e3c1-3919-48f3-8d1a-68a60974255e tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] Lock "3c099888-d5dc-4cee-9729-311808694625" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 69.863s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 657.923525] env[61957]: DEBUG nova.network.neutron [None req-4e7f7a28-6560-4b73-affe-356aa8d11b0b tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 658.293634] env[61957]: DEBUG nova.scheduler.client.report [None req-b1f62379-e574-4175-a099-f938af5ddaa8 tempest-ServerPasswordTestJSON-1543821420 tempest-ServerPasswordTestJSON-1543821420-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 658.399587] env[61957]: DEBUG nova.compute.manager [None req-af238887-1f40-4ff3-bfc6-fdb11b6cd8bf tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] [instance: 188b0fd3-5d71-4feb-aca5-75a2bd28895a] Starting instance... {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 658.425765] env[61957]: INFO nova.compute.manager [None req-4e7f7a28-6560-4b73-affe-356aa8d11b0b tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] [instance: 763d87bb-652c-43e0-ba39-135bae2cc368] Took 1.02 seconds to deallocate network for instance. [ 658.801202] env[61957]: DEBUG oslo_concurrency.lockutils [None req-b1f62379-e574-4175-a099-f938af5ddaa8 tempest-ServerPasswordTestJSON-1543821420 tempest-ServerPasswordTestJSON-1543821420-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.491s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 658.801202] env[61957]: DEBUG nova.compute.manager [None req-b1f62379-e574-4175-a099-f938af5ddaa8 tempest-ServerPasswordTestJSON-1543821420 tempest-ServerPasswordTestJSON-1543821420-project-member] [instance: 49e99297-ffb9-4104-bd06-911243908828] Start building networks asynchronously for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 658.803652] env[61957]: DEBUG oslo_concurrency.lockutils [None req-9c3c6101-e50c-439a-91f6-f293e059d0d2 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 20.490s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 658.920141] env[61957]: DEBUG oslo_concurrency.lockutils [None req-af238887-1f40-4ff3-bfc6-fdb11b6cd8bf tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 659.308504] env[61957]: DEBUG nova.compute.utils [None req-b1f62379-e574-4175-a099-f938af5ddaa8 tempest-ServerPasswordTestJSON-1543821420 tempest-ServerPasswordTestJSON-1543821420-project-member] Using /dev/sd instead of None {{(pid=61957) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 659.310395] env[61957]: DEBUG nova.compute.manager [None req-b1f62379-e574-4175-a099-f938af5ddaa8 tempest-ServerPasswordTestJSON-1543821420 tempest-ServerPasswordTestJSON-1543821420-project-member] [instance: 49e99297-ffb9-4104-bd06-911243908828] Allocating IP information in the background. {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 659.311140] env[61957]: DEBUG nova.network.neutron [None req-b1f62379-e574-4175-a099-f938af5ddaa8 tempest-ServerPasswordTestJSON-1543821420 tempest-ServerPasswordTestJSON-1543821420-project-member] [instance: 49e99297-ffb9-4104-bd06-911243908828] allocate_for_instance() {{(pid=61957) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 659.379255] env[61957]: DEBUG nova.policy [None req-b1f62379-e574-4175-a099-f938af5ddaa8 tempest-ServerPasswordTestJSON-1543821420 tempest-ServerPasswordTestJSON-1543821420-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '347c757973794214ad340f6f041a868d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b9beb7c3b5394c75ad0520bf8680872d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61957) authorize /opt/stack/nova/nova/policy.py:203}} [ 659.457579] env[61957]: INFO nova.scheduler.client.report [None req-4e7f7a28-6560-4b73-affe-356aa8d11b0b tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] Deleted allocations for instance 763d87bb-652c-43e0-ba39-135bae2cc368 [ 659.727141] env[61957]: DEBUG nova.network.neutron [None req-b1f62379-e574-4175-a099-f938af5ddaa8 tempest-ServerPasswordTestJSON-1543821420 tempest-ServerPasswordTestJSON-1543821420-project-member] [instance: 49e99297-ffb9-4104-bd06-911243908828] Successfully created port: 4372cc6c-40b2-4067-93d2-12710b09600f {{(pid=61957) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 659.779814] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56f6245a-df9d-4b0f-9d9b-693f58226897 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.788698] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6724b3d-6721-4052-9382-b3caba3cc0e9 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.820013] env[61957]: DEBUG nova.compute.manager [None req-b1f62379-e574-4175-a099-f938af5ddaa8 tempest-ServerPasswordTestJSON-1543821420 tempest-ServerPasswordTestJSON-1543821420-project-member] [instance: 49e99297-ffb9-4104-bd06-911243908828] Start building block device mappings for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 659.825273] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38c312a6-044c-4b8c-bd91-027a9042de22 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.834705] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2e39e57-25d1-493e-8c01-4c46f77a2ec3 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.857793] env[61957]: DEBUG nova.compute.provider_tree [None req-9c3c6101-e50c-439a-91f6-f293e059d0d2 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 659.964782] env[61957]: DEBUG oslo_concurrency.lockutils [None req-4e7f7a28-6560-4b73-affe-356aa8d11b0b tempest-ListImageFiltersTestJSON-1025456348 tempest-ListImageFiltersTestJSON-1025456348-project-member] Lock "763d87bb-652c-43e0-ba39-135bae2cc368" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 70.341s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 660.362600] env[61957]: DEBUG nova.scheduler.client.report [None req-9c3c6101-e50c-439a-91f6-f293e059d0d2 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 660.467383] env[61957]: DEBUG nova.compute.manager [None req-74fa7fa3-fb2c-4064-80f1-1405089fca10 tempest-ServerMetadataNegativeTestJSON-1373515003 tempest-ServerMetadataNegativeTestJSON-1373515003-project-member] [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] Starting instance... {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 660.627229] env[61957]: DEBUG nova.compute.manager [req-553eab8b-315c-4016-b21c-c28e58fab840 req-5202400d-ffec-445f-811a-719c01f7f73d service nova] [instance: 49e99297-ffb9-4104-bd06-911243908828] Received event network-changed-4372cc6c-40b2-4067-93d2-12710b09600f {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 660.627229] env[61957]: DEBUG nova.compute.manager [req-553eab8b-315c-4016-b21c-c28e58fab840 req-5202400d-ffec-445f-811a-719c01f7f73d service nova] [instance: 49e99297-ffb9-4104-bd06-911243908828] Refreshing instance network info cache due to event network-changed-4372cc6c-40b2-4067-93d2-12710b09600f. {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 660.627229] env[61957]: DEBUG oslo_concurrency.lockutils [req-553eab8b-315c-4016-b21c-c28e58fab840 req-5202400d-ffec-445f-811a-719c01f7f73d service nova] Acquiring lock "refresh_cache-49e99297-ffb9-4104-bd06-911243908828" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 660.627229] env[61957]: DEBUG oslo_concurrency.lockutils [req-553eab8b-315c-4016-b21c-c28e58fab840 req-5202400d-ffec-445f-811a-719c01f7f73d service nova] Acquired lock "refresh_cache-49e99297-ffb9-4104-bd06-911243908828" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 660.627229] env[61957]: DEBUG nova.network.neutron [req-553eab8b-315c-4016-b21c-c28e58fab840 req-5202400d-ffec-445f-811a-719c01f7f73d service nova] [instance: 49e99297-ffb9-4104-bd06-911243908828] Refreshing network info cache for port 4372cc6c-40b2-4067-93d2-12710b09600f {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 660.714470] env[61957]: ERROR nova.compute.manager [None req-b1f62379-e574-4175-a099-f938af5ddaa8 tempest-ServerPasswordTestJSON-1543821420 tempest-ServerPasswordTestJSON-1543821420-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 4372cc6c-40b2-4067-93d2-12710b09600f, please check neutron logs for more information. [ 660.714470] env[61957]: ERROR nova.compute.manager Traceback (most recent call last): [ 660.714470] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 660.714470] env[61957]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 660.714470] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 660.714470] env[61957]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 660.714470] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 660.714470] env[61957]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 660.714470] env[61957]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 660.714470] env[61957]: ERROR nova.compute.manager self.force_reraise() [ 660.714470] env[61957]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 660.714470] env[61957]: ERROR nova.compute.manager raise self.value [ 660.714470] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 660.714470] env[61957]: ERROR nova.compute.manager updated_port = self._update_port( [ 660.714470] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 660.714470] env[61957]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 660.714951] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 660.714951] env[61957]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 660.714951] env[61957]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 4372cc6c-40b2-4067-93d2-12710b09600f, please check neutron logs for more information. [ 660.714951] env[61957]: ERROR nova.compute.manager [ 660.714951] env[61957]: Traceback (most recent call last): [ 660.714951] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 660.714951] env[61957]: listener.cb(fileno) [ 660.714951] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 660.714951] env[61957]: result = function(*args, **kwargs) [ 660.714951] env[61957]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 660.714951] env[61957]: return func(*args, **kwargs) [ 660.714951] env[61957]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 660.714951] env[61957]: raise e [ 660.714951] env[61957]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 660.714951] env[61957]: nwinfo = self.network_api.allocate_for_instance( [ 660.714951] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 660.714951] env[61957]: created_port_ids = self._update_ports_for_instance( [ 660.714951] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 660.714951] env[61957]: with excutils.save_and_reraise_exception(): [ 660.714951] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 660.714951] env[61957]: self.force_reraise() [ 660.714951] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 660.714951] env[61957]: raise self.value [ 660.714951] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 660.714951] env[61957]: updated_port = self._update_port( [ 660.714951] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 660.714951] env[61957]: _ensure_no_port_binding_failure(port) [ 660.714951] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 660.714951] env[61957]: raise exception.PortBindingFailed(port_id=port['id']) [ 660.715718] env[61957]: nova.exception.PortBindingFailed: Binding failed for port 4372cc6c-40b2-4067-93d2-12710b09600f, please check neutron logs for more information. [ 660.715718] env[61957]: Removing descriptor: 17 [ 660.834831] env[61957]: DEBUG nova.compute.manager [None req-b1f62379-e574-4175-a099-f938af5ddaa8 tempest-ServerPasswordTestJSON-1543821420 tempest-ServerPasswordTestJSON-1543821420-project-member] [instance: 49e99297-ffb9-4104-bd06-911243908828] Start spawning the instance on the hypervisor. {{(pid=61957) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 660.860681] env[61957]: DEBUG nova.virt.hardware [None req-b1f62379-e574-4175-a099-f938af5ddaa8 tempest-ServerPasswordTestJSON-1543821420 tempest-ServerPasswordTestJSON-1543821420-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T17:22:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T17:21:50Z,direct_url=,disk_format='vmdk',id=11c76a2c-f705-470a-ba9d-4657858bab38,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='81d3096d0f094373913fc3dc8f5c3c6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T17:21:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 660.861019] env[61957]: DEBUG nova.virt.hardware [None req-b1f62379-e574-4175-a099-f938af5ddaa8 tempest-ServerPasswordTestJSON-1543821420 tempest-ServerPasswordTestJSON-1543821420-project-member] Flavor limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 660.861158] env[61957]: DEBUG nova.virt.hardware [None req-b1f62379-e574-4175-a099-f938af5ddaa8 tempest-ServerPasswordTestJSON-1543821420 tempest-ServerPasswordTestJSON-1543821420-project-member] Image limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 660.861348] env[61957]: DEBUG nova.virt.hardware [None req-b1f62379-e574-4175-a099-f938af5ddaa8 tempest-ServerPasswordTestJSON-1543821420 tempest-ServerPasswordTestJSON-1543821420-project-member] Flavor pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 660.861495] env[61957]: DEBUG nova.virt.hardware [None req-b1f62379-e574-4175-a099-f938af5ddaa8 tempest-ServerPasswordTestJSON-1543821420 tempest-ServerPasswordTestJSON-1543821420-project-member] Image pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 660.861642] env[61957]: DEBUG nova.virt.hardware [None req-b1f62379-e574-4175-a099-f938af5ddaa8 tempest-ServerPasswordTestJSON-1543821420 tempest-ServerPasswordTestJSON-1543821420-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 660.861852] env[61957]: DEBUG nova.virt.hardware [None req-b1f62379-e574-4175-a099-f938af5ddaa8 tempest-ServerPasswordTestJSON-1543821420 tempest-ServerPasswordTestJSON-1543821420-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 660.862016] env[61957]: DEBUG nova.virt.hardware [None req-b1f62379-e574-4175-a099-f938af5ddaa8 tempest-ServerPasswordTestJSON-1543821420 tempest-ServerPasswordTestJSON-1543821420-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 660.862183] env[61957]: DEBUG nova.virt.hardware [None req-b1f62379-e574-4175-a099-f938af5ddaa8 tempest-ServerPasswordTestJSON-1543821420 tempest-ServerPasswordTestJSON-1543821420-project-member] Got 1 possible topologies {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 660.862344] env[61957]: DEBUG nova.virt.hardware [None req-b1f62379-e574-4175-a099-f938af5ddaa8 tempest-ServerPasswordTestJSON-1543821420 tempest-ServerPasswordTestJSON-1543821420-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 660.862517] env[61957]: DEBUG nova.virt.hardware [None req-b1f62379-e574-4175-a099-f938af5ddaa8 tempest-ServerPasswordTestJSON-1543821420 tempest-ServerPasswordTestJSON-1543821420-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 660.863383] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-450d7def-a164-40d8-9ada-dcdb900ceb6c {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.870951] env[61957]: DEBUG oslo_concurrency.lockutils [None req-9c3c6101-e50c-439a-91f6-f293e059d0d2 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.067s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 660.871546] env[61957]: ERROR nova.compute.manager [None req-9c3c6101-e50c-439a-91f6-f293e059d0d2 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port ea3b3b9c-21bf-4374-a7c5-d040c2ca5f7a, please check neutron logs for more information. [ 660.871546] env[61957]: ERROR nova.compute.manager [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] Traceback (most recent call last): [ 660.871546] env[61957]: ERROR nova.compute.manager [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 660.871546] env[61957]: ERROR nova.compute.manager [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] self.driver.spawn(context, instance, image_meta, [ 660.871546] env[61957]: ERROR nova.compute.manager [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 660.871546] env[61957]: ERROR nova.compute.manager [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 660.871546] env[61957]: ERROR nova.compute.manager [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 660.871546] env[61957]: ERROR nova.compute.manager [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] vm_ref = self.build_virtual_machine(instance, [ 660.871546] env[61957]: ERROR nova.compute.manager [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 660.871546] env[61957]: ERROR nova.compute.manager [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] vif_infos = vmwarevif.get_vif_info(self._session, [ 660.871546] env[61957]: ERROR nova.compute.manager [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 660.871871] env[61957]: ERROR nova.compute.manager [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] for vif in network_info: [ 660.871871] env[61957]: ERROR nova.compute.manager [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 660.871871] env[61957]: ERROR nova.compute.manager [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] return self._sync_wrapper(fn, *args, **kwargs) [ 660.871871] env[61957]: ERROR nova.compute.manager [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 660.871871] env[61957]: ERROR nova.compute.manager [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] self.wait() [ 660.871871] env[61957]: ERROR nova.compute.manager [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 660.871871] env[61957]: ERROR nova.compute.manager [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] self[:] = self._gt.wait() [ 660.871871] env[61957]: ERROR nova.compute.manager [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 660.871871] env[61957]: ERROR nova.compute.manager [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] return self._exit_event.wait() [ 660.871871] env[61957]: ERROR nova.compute.manager [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 660.871871] env[61957]: ERROR nova.compute.manager [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] result = hub.switch() [ 660.871871] env[61957]: ERROR nova.compute.manager [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 660.871871] env[61957]: ERROR nova.compute.manager [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] return self.greenlet.switch() [ 660.872211] env[61957]: ERROR nova.compute.manager [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 660.872211] env[61957]: ERROR nova.compute.manager [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] result = function(*args, **kwargs) [ 660.872211] env[61957]: ERROR nova.compute.manager [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 660.872211] env[61957]: ERROR nova.compute.manager [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] return func(*args, **kwargs) [ 660.872211] env[61957]: ERROR nova.compute.manager [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 660.872211] env[61957]: ERROR nova.compute.manager [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] raise e [ 660.872211] env[61957]: ERROR nova.compute.manager [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 660.872211] env[61957]: ERROR nova.compute.manager [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] nwinfo = self.network_api.allocate_for_instance( [ 660.872211] env[61957]: ERROR nova.compute.manager [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 660.872211] env[61957]: ERROR nova.compute.manager [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] created_port_ids = self._update_ports_for_instance( [ 660.872211] env[61957]: ERROR nova.compute.manager [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 660.872211] env[61957]: ERROR nova.compute.manager [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] with excutils.save_and_reraise_exception(): [ 660.872211] env[61957]: ERROR nova.compute.manager [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 660.872529] env[61957]: ERROR nova.compute.manager [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] self.force_reraise() [ 660.872529] env[61957]: ERROR nova.compute.manager [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 660.872529] env[61957]: ERROR nova.compute.manager [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] raise self.value [ 660.872529] env[61957]: ERROR nova.compute.manager [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 660.872529] env[61957]: ERROR nova.compute.manager [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] updated_port = self._update_port( [ 660.872529] env[61957]: ERROR nova.compute.manager [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 660.872529] env[61957]: ERROR nova.compute.manager [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] _ensure_no_port_binding_failure(port) [ 660.872529] env[61957]: ERROR nova.compute.manager [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 660.872529] env[61957]: ERROR nova.compute.manager [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] raise exception.PortBindingFailed(port_id=port['id']) [ 660.872529] env[61957]: ERROR nova.compute.manager [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] nova.exception.PortBindingFailed: Binding failed for port ea3b3b9c-21bf-4374-a7c5-d040c2ca5f7a, please check neutron logs for more information. [ 660.872529] env[61957]: ERROR nova.compute.manager [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] [ 660.872824] env[61957]: DEBUG nova.compute.utils [None req-9c3c6101-e50c-439a-91f6-f293e059d0d2 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] Binding failed for port ea3b3b9c-21bf-4374-a7c5-d040c2ca5f7a, please check neutron logs for more information. {{(pid=61957) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 660.873362] env[61957]: DEBUG oslo_concurrency.lockutils [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 20.563s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 660.873536] env[61957]: DEBUG oslo_concurrency.lockutils [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 660.873683] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61957) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 660.873928] env[61957]: DEBUG oslo_concurrency.lockutils [None req-fc6b77c8-b5cd-4608-beab-d2cb5c9f7440 tempest-ImagesNegativeTestJSON-1548152881 tempest-ImagesNegativeTestJSON-1548152881-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.025s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 660.875919] env[61957]: INFO nova.compute.claims [None req-fc6b77c8-b5cd-4608-beab-d2cb5c9f7440 tempest-ImagesNegativeTestJSON-1548152881 tempest-ImagesNegativeTestJSON-1548152881-project-member] [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 660.879148] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fd64c22-4ac5-4e55-b3bb-c4c5a0a6b9b2 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.884031] env[61957]: DEBUG nova.compute.manager [None req-9c3c6101-e50c-439a-91f6-f293e059d0d2 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] Build of instance 4be3eabd-4e58-48b8-b14a-798ba419655e was re-scheduled: Binding failed for port ea3b3b9c-21bf-4374-a7c5-d040c2ca5f7a, please check neutron logs for more information. {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 660.884031] env[61957]: DEBUG nova.compute.manager [None req-9c3c6101-e50c-439a-91f6-f293e059d0d2 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] Unplugging VIFs for instance {{(pid=61957) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 660.884031] env[61957]: DEBUG oslo_concurrency.lockutils [None req-9c3c6101-e50c-439a-91f6-f293e059d0d2 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] Acquiring lock "refresh_cache-4be3eabd-4e58-48b8-b14a-798ba419655e" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 660.884211] env[61957]: DEBUG oslo_concurrency.lockutils [None req-9c3c6101-e50c-439a-91f6-f293e059d0d2 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] Acquired lock "refresh_cache-4be3eabd-4e58-48b8-b14a-798ba419655e" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 660.884239] env[61957]: DEBUG nova.network.neutron [None req-9c3c6101-e50c-439a-91f6-f293e059d0d2 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 660.885565] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc212ff2-3f12-44c5-81bb-d61f415b8a42 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.901918] env[61957]: ERROR nova.compute.manager [None req-b1f62379-e574-4175-a099-f938af5ddaa8 tempest-ServerPasswordTestJSON-1543821420 tempest-ServerPasswordTestJSON-1543821420-project-member] [instance: 49e99297-ffb9-4104-bd06-911243908828] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 4372cc6c-40b2-4067-93d2-12710b09600f, please check neutron logs for more information. [ 660.901918] env[61957]: ERROR nova.compute.manager [instance: 49e99297-ffb9-4104-bd06-911243908828] Traceback (most recent call last): [ 660.901918] env[61957]: ERROR nova.compute.manager [instance: 49e99297-ffb9-4104-bd06-911243908828] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 660.901918] env[61957]: ERROR nova.compute.manager [instance: 49e99297-ffb9-4104-bd06-911243908828] yield resources [ 660.901918] env[61957]: ERROR nova.compute.manager [instance: 49e99297-ffb9-4104-bd06-911243908828] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 660.901918] env[61957]: ERROR nova.compute.manager [instance: 49e99297-ffb9-4104-bd06-911243908828] self.driver.spawn(context, instance, image_meta, [ 660.901918] env[61957]: ERROR nova.compute.manager [instance: 49e99297-ffb9-4104-bd06-911243908828] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 660.901918] env[61957]: ERROR nova.compute.manager [instance: 49e99297-ffb9-4104-bd06-911243908828] self._vmops.spawn(context, instance, image_meta, injected_files, [ 660.901918] env[61957]: ERROR nova.compute.manager [instance: 49e99297-ffb9-4104-bd06-911243908828] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 660.901918] env[61957]: ERROR nova.compute.manager [instance: 49e99297-ffb9-4104-bd06-911243908828] vm_ref = self.build_virtual_machine(instance, [ 660.901918] env[61957]: ERROR nova.compute.manager [instance: 49e99297-ffb9-4104-bd06-911243908828] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 660.902240] env[61957]: ERROR nova.compute.manager [instance: 49e99297-ffb9-4104-bd06-911243908828] vif_infos = vmwarevif.get_vif_info(self._session, [ 660.902240] env[61957]: ERROR nova.compute.manager [instance: 49e99297-ffb9-4104-bd06-911243908828] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 660.902240] env[61957]: ERROR nova.compute.manager [instance: 49e99297-ffb9-4104-bd06-911243908828] for vif in network_info: [ 660.902240] env[61957]: ERROR nova.compute.manager [instance: 49e99297-ffb9-4104-bd06-911243908828] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 660.902240] env[61957]: ERROR nova.compute.manager [instance: 49e99297-ffb9-4104-bd06-911243908828] return self._sync_wrapper(fn, *args, **kwargs) [ 660.902240] env[61957]: ERROR nova.compute.manager [instance: 49e99297-ffb9-4104-bd06-911243908828] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 660.902240] env[61957]: ERROR nova.compute.manager [instance: 49e99297-ffb9-4104-bd06-911243908828] self.wait() [ 660.902240] env[61957]: ERROR nova.compute.manager [instance: 49e99297-ffb9-4104-bd06-911243908828] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 660.902240] env[61957]: ERROR nova.compute.manager [instance: 49e99297-ffb9-4104-bd06-911243908828] self[:] = self._gt.wait() [ 660.902240] env[61957]: ERROR nova.compute.manager [instance: 49e99297-ffb9-4104-bd06-911243908828] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 660.902240] env[61957]: ERROR nova.compute.manager [instance: 49e99297-ffb9-4104-bd06-911243908828] return self._exit_event.wait() [ 660.902240] env[61957]: ERROR nova.compute.manager [instance: 49e99297-ffb9-4104-bd06-911243908828] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 660.902240] env[61957]: ERROR nova.compute.manager [instance: 49e99297-ffb9-4104-bd06-911243908828] current.throw(*self._exc) [ 660.902563] env[61957]: ERROR nova.compute.manager [instance: 49e99297-ffb9-4104-bd06-911243908828] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 660.902563] env[61957]: ERROR nova.compute.manager [instance: 49e99297-ffb9-4104-bd06-911243908828] result = function(*args, **kwargs) [ 660.902563] env[61957]: ERROR nova.compute.manager [instance: 49e99297-ffb9-4104-bd06-911243908828] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 660.902563] env[61957]: ERROR nova.compute.manager [instance: 49e99297-ffb9-4104-bd06-911243908828] return func(*args, **kwargs) [ 660.902563] env[61957]: ERROR nova.compute.manager [instance: 49e99297-ffb9-4104-bd06-911243908828] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 660.902563] env[61957]: ERROR nova.compute.manager [instance: 49e99297-ffb9-4104-bd06-911243908828] raise e [ 660.902563] env[61957]: ERROR nova.compute.manager [instance: 49e99297-ffb9-4104-bd06-911243908828] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 660.902563] env[61957]: ERROR nova.compute.manager [instance: 49e99297-ffb9-4104-bd06-911243908828] nwinfo = self.network_api.allocate_for_instance( [ 660.902563] env[61957]: ERROR nova.compute.manager [instance: 49e99297-ffb9-4104-bd06-911243908828] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 660.902563] env[61957]: ERROR nova.compute.manager [instance: 49e99297-ffb9-4104-bd06-911243908828] created_port_ids = self._update_ports_for_instance( [ 660.902563] env[61957]: ERROR nova.compute.manager [instance: 49e99297-ffb9-4104-bd06-911243908828] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 660.902563] env[61957]: ERROR nova.compute.manager [instance: 49e99297-ffb9-4104-bd06-911243908828] with excutils.save_and_reraise_exception(): [ 660.902563] env[61957]: ERROR nova.compute.manager [instance: 49e99297-ffb9-4104-bd06-911243908828] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 660.902840] env[61957]: ERROR nova.compute.manager [instance: 49e99297-ffb9-4104-bd06-911243908828] self.force_reraise() [ 660.902840] env[61957]: ERROR nova.compute.manager [instance: 49e99297-ffb9-4104-bd06-911243908828] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 660.902840] env[61957]: ERROR nova.compute.manager [instance: 49e99297-ffb9-4104-bd06-911243908828] raise self.value [ 660.902840] env[61957]: ERROR nova.compute.manager [instance: 49e99297-ffb9-4104-bd06-911243908828] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 660.902840] env[61957]: ERROR nova.compute.manager [instance: 49e99297-ffb9-4104-bd06-911243908828] updated_port = self._update_port( [ 660.902840] env[61957]: ERROR nova.compute.manager [instance: 49e99297-ffb9-4104-bd06-911243908828] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 660.902840] env[61957]: ERROR nova.compute.manager [instance: 49e99297-ffb9-4104-bd06-911243908828] _ensure_no_port_binding_failure(port) [ 660.902840] env[61957]: ERROR nova.compute.manager [instance: 49e99297-ffb9-4104-bd06-911243908828] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 660.902840] env[61957]: ERROR nova.compute.manager [instance: 49e99297-ffb9-4104-bd06-911243908828] raise exception.PortBindingFailed(port_id=port['id']) [ 660.902840] env[61957]: ERROR nova.compute.manager [instance: 49e99297-ffb9-4104-bd06-911243908828] nova.exception.PortBindingFailed: Binding failed for port 4372cc6c-40b2-4067-93d2-12710b09600f, please check neutron logs for more information. [ 660.902840] env[61957]: ERROR nova.compute.manager [instance: 49e99297-ffb9-4104-bd06-911243908828] [ 660.902840] env[61957]: INFO nova.compute.manager [None req-b1f62379-e574-4175-a099-f938af5ddaa8 tempest-ServerPasswordTestJSON-1543821420 tempest-ServerPasswordTestJSON-1543821420-project-member] [instance: 49e99297-ffb9-4104-bd06-911243908828] Terminating instance [ 660.904788] env[61957]: DEBUG oslo_concurrency.lockutils [None req-b1f62379-e574-4175-a099-f938af5ddaa8 tempest-ServerPasswordTestJSON-1543821420 tempest-ServerPasswordTestJSON-1543821420-project-member] Acquiring lock "refresh_cache-49e99297-ffb9-4104-bd06-911243908828" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 660.905966] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef2074d5-257d-469c-b0ce-de708f06e457 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.920854] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0dc41b6e-fc1b-4d6f-ba79-afd4cae33d0a {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.928089] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3864fb7-625c-4ef0-9242-405e06bf5e65 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.960450] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181411MB free_disk=142GB free_vcpus=48 pci_devices=None {{(pid=61957) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 660.960650] env[61957]: DEBUG oslo_concurrency.lockutils [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 660.992931] env[61957]: DEBUG oslo_concurrency.lockutils [None req-74fa7fa3-fb2c-4064-80f1-1405089fca10 tempest-ServerMetadataNegativeTestJSON-1373515003 tempest-ServerMetadataNegativeTestJSON-1373515003-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 661.151164] env[61957]: DEBUG nova.network.neutron [req-553eab8b-315c-4016-b21c-c28e58fab840 req-5202400d-ffec-445f-811a-719c01f7f73d service nova] [instance: 49e99297-ffb9-4104-bd06-911243908828] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 661.242621] env[61957]: DEBUG nova.network.neutron [req-553eab8b-315c-4016-b21c-c28e58fab840 req-5202400d-ffec-445f-811a-719c01f7f73d service nova] [instance: 49e99297-ffb9-4104-bd06-911243908828] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 661.405321] env[61957]: DEBUG nova.network.neutron [None req-9c3c6101-e50c-439a-91f6-f293e059d0d2 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 661.463015] env[61957]: DEBUG nova.network.neutron [None req-9c3c6101-e50c-439a-91f6-f293e059d0d2 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 661.745146] env[61957]: DEBUG oslo_concurrency.lockutils [req-553eab8b-315c-4016-b21c-c28e58fab840 req-5202400d-ffec-445f-811a-719c01f7f73d service nova] Releasing lock "refresh_cache-49e99297-ffb9-4104-bd06-911243908828" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 661.745528] env[61957]: DEBUG oslo_concurrency.lockutils [None req-b1f62379-e574-4175-a099-f938af5ddaa8 tempest-ServerPasswordTestJSON-1543821420 tempest-ServerPasswordTestJSON-1543821420-project-member] Acquired lock "refresh_cache-49e99297-ffb9-4104-bd06-911243908828" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 661.745725] env[61957]: DEBUG nova.network.neutron [None req-b1f62379-e574-4175-a099-f938af5ddaa8 tempest-ServerPasswordTestJSON-1543821420 tempest-ServerPasswordTestJSON-1543821420-project-member] [instance: 49e99297-ffb9-4104-bd06-911243908828] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 661.969144] env[61957]: DEBUG oslo_concurrency.lockutils [None req-9c3c6101-e50c-439a-91f6-f293e059d0d2 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] Releasing lock "refresh_cache-4be3eabd-4e58-48b8-b14a-798ba419655e" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 661.969393] env[61957]: DEBUG nova.compute.manager [None req-9c3c6101-e50c-439a-91f6-f293e059d0d2 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61957) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 661.970042] env[61957]: DEBUG nova.compute.manager [None req-9c3c6101-e50c-439a-91f6-f293e059d0d2 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 661.970042] env[61957]: DEBUG nova.network.neutron [None req-9c3c6101-e50c-439a-91f6-f293e059d0d2 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 661.989847] env[61957]: DEBUG nova.network.neutron [None req-9c3c6101-e50c-439a-91f6-f293e059d0d2 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 662.278951] env[61957]: DEBUG nova.network.neutron [None req-b1f62379-e574-4175-a099-f938af5ddaa8 tempest-ServerPasswordTestJSON-1543821420 tempest-ServerPasswordTestJSON-1543821420-project-member] [instance: 49e99297-ffb9-4104-bd06-911243908828] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 662.309093] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-614d6427-b530-41df-aa4d-2e28f7954b07 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.314215] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e659ebcc-c53d-4ce5-bb27-34bd75628951 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.354499] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-466a3ff8-f404-4b9c-89fa-f6b259de8200 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.362636] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-048c9ac5-a0cd-4e6b-a63c-5a7de9b6b948 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.380617] env[61957]: DEBUG nova.compute.provider_tree [None req-fc6b77c8-b5cd-4608-beab-d2cb5c9f7440 tempest-ImagesNegativeTestJSON-1548152881 tempest-ImagesNegativeTestJSON-1548152881-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 662.407119] env[61957]: DEBUG nova.network.neutron [None req-b1f62379-e574-4175-a099-f938af5ddaa8 tempest-ServerPasswordTestJSON-1543821420 tempest-ServerPasswordTestJSON-1543821420-project-member] [instance: 49e99297-ffb9-4104-bd06-911243908828] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 662.492212] env[61957]: DEBUG nova.network.neutron [None req-9c3c6101-e50c-439a-91f6-f293e059d0d2 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 662.670306] env[61957]: DEBUG nova.compute.manager [req-fbf951cf-244a-4e5c-b6f4-ca8effe13597 req-d825606b-741d-4f28-8399-b8d97004cf19 service nova] [instance: 49e99297-ffb9-4104-bd06-911243908828] Received event network-vif-deleted-4372cc6c-40b2-4067-93d2-12710b09600f {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 662.885044] env[61957]: DEBUG nova.scheduler.client.report [None req-fc6b77c8-b5cd-4608-beab-d2cb5c9f7440 tempest-ImagesNegativeTestJSON-1548152881 tempest-ImagesNegativeTestJSON-1548152881-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 662.910644] env[61957]: DEBUG oslo_concurrency.lockutils [None req-b1f62379-e574-4175-a099-f938af5ddaa8 tempest-ServerPasswordTestJSON-1543821420 tempest-ServerPasswordTestJSON-1543821420-project-member] Releasing lock "refresh_cache-49e99297-ffb9-4104-bd06-911243908828" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 662.910644] env[61957]: DEBUG nova.compute.manager [None req-b1f62379-e574-4175-a099-f938af5ddaa8 tempest-ServerPasswordTestJSON-1543821420 tempest-ServerPasswordTestJSON-1543821420-project-member] [instance: 49e99297-ffb9-4104-bd06-911243908828] Start destroying the instance on the hypervisor. {{(pid=61957) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 662.910644] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-b1f62379-e574-4175-a099-f938af5ddaa8 tempest-ServerPasswordTestJSON-1543821420 tempest-ServerPasswordTestJSON-1543821420-project-member] [instance: 49e99297-ffb9-4104-bd06-911243908828] Destroying instance {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 662.911729] env[61957]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c10cac8d-ddd7-4cb8-abf5-e0a33fe8c8bc {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.920894] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8de750a4-0008-46db-bce1-9fd1e8784427 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.945063] env[61957]: WARNING nova.virt.vmwareapi.vmops [None req-b1f62379-e574-4175-a099-f938af5ddaa8 tempest-ServerPasswordTestJSON-1543821420 tempest-ServerPasswordTestJSON-1543821420-project-member] [instance: 49e99297-ffb9-4104-bd06-911243908828] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 49e99297-ffb9-4104-bd06-911243908828 could not be found. [ 662.945482] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-b1f62379-e574-4175-a099-f938af5ddaa8 tempest-ServerPasswordTestJSON-1543821420 tempest-ServerPasswordTestJSON-1543821420-project-member] [instance: 49e99297-ffb9-4104-bd06-911243908828] Instance destroyed {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 662.946174] env[61957]: INFO nova.compute.manager [None req-b1f62379-e574-4175-a099-f938af5ddaa8 tempest-ServerPasswordTestJSON-1543821420 tempest-ServerPasswordTestJSON-1543821420-project-member] [instance: 49e99297-ffb9-4104-bd06-911243908828] Took 0.04 seconds to destroy the instance on the hypervisor. [ 662.946174] env[61957]: DEBUG oslo.service.loopingcall [None req-b1f62379-e574-4175-a099-f938af5ddaa8 tempest-ServerPasswordTestJSON-1543821420 tempest-ServerPasswordTestJSON-1543821420-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 662.947326] env[61957]: DEBUG nova.compute.manager [-] [instance: 49e99297-ffb9-4104-bd06-911243908828] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 662.947326] env[61957]: DEBUG nova.network.neutron [-] [instance: 49e99297-ffb9-4104-bd06-911243908828] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 662.963695] env[61957]: DEBUG nova.network.neutron [-] [instance: 49e99297-ffb9-4104-bd06-911243908828] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 662.995645] env[61957]: INFO nova.compute.manager [None req-9c3c6101-e50c-439a-91f6-f293e059d0d2 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] [instance: 4be3eabd-4e58-48b8-b14a-798ba419655e] Took 1.03 seconds to deallocate network for instance. [ 663.393767] env[61957]: DEBUG oslo_concurrency.lockutils [None req-fc6b77c8-b5cd-4608-beab-d2cb5c9f7440 tempest-ImagesNegativeTestJSON-1548152881 tempest-ImagesNegativeTestJSON-1548152881-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.520s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 663.394378] env[61957]: DEBUG nova.compute.manager [None req-fc6b77c8-b5cd-4608-beab-d2cb5c9f7440 tempest-ImagesNegativeTestJSON-1548152881 tempest-ImagesNegativeTestJSON-1548152881-project-member] [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] Start building networks asynchronously for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 663.397467] env[61957]: DEBUG oslo_concurrency.lockutils [None req-af971b81-6348-4338-b0a9-b06675d34db5 tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.349s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 663.467238] env[61957]: DEBUG nova.network.neutron [-] [instance: 49e99297-ffb9-4104-bd06-911243908828] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 663.903213] env[61957]: DEBUG nova.compute.utils [None req-fc6b77c8-b5cd-4608-beab-d2cb5c9f7440 tempest-ImagesNegativeTestJSON-1548152881 tempest-ImagesNegativeTestJSON-1548152881-project-member] Using /dev/sd instead of None {{(pid=61957) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 663.908335] env[61957]: DEBUG nova.compute.manager [None req-fc6b77c8-b5cd-4608-beab-d2cb5c9f7440 tempest-ImagesNegativeTestJSON-1548152881 tempest-ImagesNegativeTestJSON-1548152881-project-member] [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] Allocating IP information in the background. {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 663.908431] env[61957]: DEBUG nova.network.neutron [None req-fc6b77c8-b5cd-4608-beab-d2cb5c9f7440 tempest-ImagesNegativeTestJSON-1548152881 tempest-ImagesNegativeTestJSON-1548152881-project-member] [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] allocate_for_instance() {{(pid=61957) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 663.954617] env[61957]: DEBUG nova.policy [None req-fc6b77c8-b5cd-4608-beab-d2cb5c9f7440 tempest-ImagesNegativeTestJSON-1548152881 tempest-ImagesNegativeTestJSON-1548152881-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '04b21d4362734dafa2277c1016fdb32a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'cd06961ca33249bcaa9a62dbd3f6f587', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61957) authorize /opt/stack/nova/nova/policy.py:203}} [ 663.972891] env[61957]: INFO nova.compute.manager [-] [instance: 49e99297-ffb9-4104-bd06-911243908828] Took 1.03 seconds to deallocate network for instance. [ 663.978467] env[61957]: DEBUG nova.compute.claims [None req-b1f62379-e574-4175-a099-f938af5ddaa8 tempest-ServerPasswordTestJSON-1543821420 tempest-ServerPasswordTestJSON-1543821420-project-member] [instance: 49e99297-ffb9-4104-bd06-911243908828] Aborting claim: {{(pid=61957) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 663.978467] env[61957]: DEBUG oslo_concurrency.lockutils [None req-b1f62379-e574-4175-a099-f938af5ddaa8 tempest-ServerPasswordTestJSON-1543821420 tempest-ServerPasswordTestJSON-1543821420-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 664.023825] env[61957]: INFO nova.scheduler.client.report [None req-9c3c6101-e50c-439a-91f6-f293e059d0d2 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] Deleted allocations for instance 4be3eabd-4e58-48b8-b14a-798ba419655e [ 664.278822] env[61957]: DEBUG nova.network.neutron [None req-fc6b77c8-b5cd-4608-beab-d2cb5c9f7440 tempest-ImagesNegativeTestJSON-1548152881 tempest-ImagesNegativeTestJSON-1548152881-project-member] [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] Successfully created port: 7ebf5e95-b147-40b2-9271-c193cd89e949 {{(pid=61957) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 664.409133] env[61957]: DEBUG nova.compute.manager [None req-fc6b77c8-b5cd-4608-beab-d2cb5c9f7440 tempest-ImagesNegativeTestJSON-1548152881 tempest-ImagesNegativeTestJSON-1548152881-project-member] [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] Start building block device mappings for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 664.420924] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-545cf139-b7c4-40ed-8c15-ee51127b08a7 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.430997] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-218711b6-861b-49ac-a057-e4124961a25b {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.475655] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0ef2927-a670-4db6-be32-049b531857bc {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.485021] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b427933-056b-4c49-90a3-2f6dfc7f3d76 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.119828] env[61957]: DEBUG oslo_concurrency.lockutils [None req-9c3c6101-e50c-439a-91f6-f293e059d0d2 tempest-DeleteServersAdminTestJSON-1869369113 tempest-DeleteServersAdminTestJSON-1869369113-project-member] Lock "4be3eabd-4e58-48b8-b14a-798ba419655e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 73.684s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 665.136244] env[61957]: DEBUG nova.compute.provider_tree [None req-af971b81-6348-4338-b0a9-b06675d34db5 tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 665.139938] env[61957]: DEBUG oslo_concurrency.lockutils [None req-da68b092-4602-4484-b1b0-adcac4aae9f8 tempest-InstanceActionsV221TestJSON-1768084888 tempest-InstanceActionsV221TestJSON-1768084888-project-member] Acquiring lock "6e7bd89c-2c2a-450a-9858-3526d96c28ab" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 665.139938] env[61957]: DEBUG oslo_concurrency.lockutils [None req-da68b092-4602-4484-b1b0-adcac4aae9f8 tempest-InstanceActionsV221TestJSON-1768084888 tempest-InstanceActionsV221TestJSON-1768084888-project-member] Lock "6e7bd89c-2c2a-450a-9858-3526d96c28ab" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 665.254146] env[61957]: DEBUG nova.compute.manager [req-f1ec6200-ba8c-4ddd-a5d7-43cc055a799a req-5156cb75-bd32-42b9-96db-d5db3d75f335 service nova] [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] Received event network-changed-7ebf5e95-b147-40b2-9271-c193cd89e949 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 665.254846] env[61957]: DEBUG nova.compute.manager [req-f1ec6200-ba8c-4ddd-a5d7-43cc055a799a req-5156cb75-bd32-42b9-96db-d5db3d75f335 service nova] [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] Refreshing instance network info cache due to event network-changed-7ebf5e95-b147-40b2-9271-c193cd89e949. {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 665.254846] env[61957]: DEBUG oslo_concurrency.lockutils [req-f1ec6200-ba8c-4ddd-a5d7-43cc055a799a req-5156cb75-bd32-42b9-96db-d5db3d75f335 service nova] Acquiring lock "refresh_cache-9df88f23-0219-43e0-b28a-e78f30a473a7" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 665.254846] env[61957]: DEBUG oslo_concurrency.lockutils [req-f1ec6200-ba8c-4ddd-a5d7-43cc055a799a req-5156cb75-bd32-42b9-96db-d5db3d75f335 service nova] Acquired lock "refresh_cache-9df88f23-0219-43e0-b28a-e78f30a473a7" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 665.255197] env[61957]: DEBUG nova.network.neutron [req-f1ec6200-ba8c-4ddd-a5d7-43cc055a799a req-5156cb75-bd32-42b9-96db-d5db3d75f335 service nova] [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] Refreshing network info cache for port 7ebf5e95-b147-40b2-9271-c193cd89e949 {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 665.399921] env[61957]: ERROR nova.compute.manager [None req-fc6b77c8-b5cd-4608-beab-d2cb5c9f7440 tempest-ImagesNegativeTestJSON-1548152881 tempest-ImagesNegativeTestJSON-1548152881-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 7ebf5e95-b147-40b2-9271-c193cd89e949, please check neutron logs for more information. [ 665.399921] env[61957]: ERROR nova.compute.manager Traceback (most recent call last): [ 665.399921] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 665.399921] env[61957]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 665.399921] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 665.399921] env[61957]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 665.399921] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 665.399921] env[61957]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 665.399921] env[61957]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 665.399921] env[61957]: ERROR nova.compute.manager self.force_reraise() [ 665.399921] env[61957]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 665.399921] env[61957]: ERROR nova.compute.manager raise self.value [ 665.399921] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 665.399921] env[61957]: ERROR nova.compute.manager updated_port = self._update_port( [ 665.399921] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 665.399921] env[61957]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 665.400396] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 665.400396] env[61957]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 665.400396] env[61957]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 7ebf5e95-b147-40b2-9271-c193cd89e949, please check neutron logs for more information. [ 665.400396] env[61957]: ERROR nova.compute.manager [ 665.400396] env[61957]: Traceback (most recent call last): [ 665.400396] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 665.400396] env[61957]: listener.cb(fileno) [ 665.400396] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 665.400396] env[61957]: result = function(*args, **kwargs) [ 665.400396] env[61957]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 665.400396] env[61957]: return func(*args, **kwargs) [ 665.400396] env[61957]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 665.400396] env[61957]: raise e [ 665.400396] env[61957]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 665.400396] env[61957]: nwinfo = self.network_api.allocate_for_instance( [ 665.400396] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 665.400396] env[61957]: created_port_ids = self._update_ports_for_instance( [ 665.400396] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 665.400396] env[61957]: with excutils.save_and_reraise_exception(): [ 665.400396] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 665.400396] env[61957]: self.force_reraise() [ 665.400396] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 665.400396] env[61957]: raise self.value [ 665.400396] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 665.400396] env[61957]: updated_port = self._update_port( [ 665.400396] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 665.400396] env[61957]: _ensure_no_port_binding_failure(port) [ 665.400396] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 665.400396] env[61957]: raise exception.PortBindingFailed(port_id=port['id']) [ 665.401085] env[61957]: nova.exception.PortBindingFailed: Binding failed for port 7ebf5e95-b147-40b2-9271-c193cd89e949, please check neutron logs for more information. [ 665.401085] env[61957]: Removing descriptor: 17 [ 665.629595] env[61957]: DEBUG nova.compute.manager [None req-fc6b77c8-b5cd-4608-beab-d2cb5c9f7440 tempest-ImagesNegativeTestJSON-1548152881 tempest-ImagesNegativeTestJSON-1548152881-project-member] [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] Start spawning the instance on the hypervisor. {{(pid=61957) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 665.630444] env[61957]: DEBUG nova.compute.manager [None req-dbfaab8a-0784-4441-ad61-79e56d9c3c60 tempest-FloatingIPsAssociationTestJSON-1786153562 tempest-FloatingIPsAssociationTestJSON-1786153562-project-member] [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] Starting instance... {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 665.641088] env[61957]: DEBUG nova.scheduler.client.report [None req-af971b81-6348-4338-b0a9-b06675d34db5 tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 665.669254] env[61957]: DEBUG nova.virt.hardware [None req-fc6b77c8-b5cd-4608-beab-d2cb5c9f7440 tempest-ImagesNegativeTestJSON-1548152881 tempest-ImagesNegativeTestJSON-1548152881-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T17:22:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T17:21:50Z,direct_url=,disk_format='vmdk',id=11c76a2c-f705-470a-ba9d-4657858bab38,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='81d3096d0f094373913fc3dc8f5c3c6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T17:21:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 665.669545] env[61957]: DEBUG nova.virt.hardware [None req-fc6b77c8-b5cd-4608-beab-d2cb5c9f7440 tempest-ImagesNegativeTestJSON-1548152881 tempest-ImagesNegativeTestJSON-1548152881-project-member] Flavor limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 665.669716] env[61957]: DEBUG nova.virt.hardware [None req-fc6b77c8-b5cd-4608-beab-d2cb5c9f7440 tempest-ImagesNegativeTestJSON-1548152881 tempest-ImagesNegativeTestJSON-1548152881-project-member] Image limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 665.669912] env[61957]: DEBUG nova.virt.hardware [None req-fc6b77c8-b5cd-4608-beab-d2cb5c9f7440 tempest-ImagesNegativeTestJSON-1548152881 tempest-ImagesNegativeTestJSON-1548152881-project-member] Flavor pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 665.670946] env[61957]: DEBUG nova.virt.hardware [None req-fc6b77c8-b5cd-4608-beab-d2cb5c9f7440 tempest-ImagesNegativeTestJSON-1548152881 tempest-ImagesNegativeTestJSON-1548152881-project-member] Image pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 665.670946] env[61957]: DEBUG nova.virt.hardware [None req-fc6b77c8-b5cd-4608-beab-d2cb5c9f7440 tempest-ImagesNegativeTestJSON-1548152881 tempest-ImagesNegativeTestJSON-1548152881-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 665.670946] env[61957]: DEBUG nova.virt.hardware [None req-fc6b77c8-b5cd-4608-beab-d2cb5c9f7440 tempest-ImagesNegativeTestJSON-1548152881 tempest-ImagesNegativeTestJSON-1548152881-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 665.670946] env[61957]: DEBUG nova.virt.hardware [None req-fc6b77c8-b5cd-4608-beab-d2cb5c9f7440 tempest-ImagesNegativeTestJSON-1548152881 tempest-ImagesNegativeTestJSON-1548152881-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 665.671143] env[61957]: DEBUG nova.virt.hardware [None req-fc6b77c8-b5cd-4608-beab-d2cb5c9f7440 tempest-ImagesNegativeTestJSON-1548152881 tempest-ImagesNegativeTestJSON-1548152881-project-member] Got 1 possible topologies {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 665.671261] env[61957]: DEBUG nova.virt.hardware [None req-fc6b77c8-b5cd-4608-beab-d2cb5c9f7440 tempest-ImagesNegativeTestJSON-1548152881 tempest-ImagesNegativeTestJSON-1548152881-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 665.671432] env[61957]: DEBUG nova.virt.hardware [None req-fc6b77c8-b5cd-4608-beab-d2cb5c9f7440 tempest-ImagesNegativeTestJSON-1548152881 tempest-ImagesNegativeTestJSON-1548152881-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 665.672310] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7aee9b03-cbed-4a48-9634-c98320a571a2 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.684720] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-759f7e78-5f8c-4c52-93b9-45552028f6f0 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.699577] env[61957]: ERROR nova.compute.manager [None req-fc6b77c8-b5cd-4608-beab-d2cb5c9f7440 tempest-ImagesNegativeTestJSON-1548152881 tempest-ImagesNegativeTestJSON-1548152881-project-member] [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 7ebf5e95-b147-40b2-9271-c193cd89e949, please check neutron logs for more information. [ 665.699577] env[61957]: ERROR nova.compute.manager [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] Traceback (most recent call last): [ 665.699577] env[61957]: ERROR nova.compute.manager [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 665.699577] env[61957]: ERROR nova.compute.manager [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] yield resources [ 665.699577] env[61957]: ERROR nova.compute.manager [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 665.699577] env[61957]: ERROR nova.compute.manager [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] self.driver.spawn(context, instance, image_meta, [ 665.699577] env[61957]: ERROR nova.compute.manager [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 665.699577] env[61957]: ERROR nova.compute.manager [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] self._vmops.spawn(context, instance, image_meta, injected_files, [ 665.699577] env[61957]: ERROR nova.compute.manager [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 665.699577] env[61957]: ERROR nova.compute.manager [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] vm_ref = self.build_virtual_machine(instance, [ 665.699577] env[61957]: ERROR nova.compute.manager [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 665.699967] env[61957]: ERROR nova.compute.manager [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] vif_infos = vmwarevif.get_vif_info(self._session, [ 665.699967] env[61957]: ERROR nova.compute.manager [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 665.699967] env[61957]: ERROR nova.compute.manager [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] for vif in network_info: [ 665.699967] env[61957]: ERROR nova.compute.manager [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 665.699967] env[61957]: ERROR nova.compute.manager [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] return self._sync_wrapper(fn, *args, **kwargs) [ 665.699967] env[61957]: ERROR nova.compute.manager [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 665.699967] env[61957]: ERROR nova.compute.manager [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] self.wait() [ 665.699967] env[61957]: ERROR nova.compute.manager [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 665.699967] env[61957]: ERROR nova.compute.manager [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] self[:] = self._gt.wait() [ 665.699967] env[61957]: ERROR nova.compute.manager [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 665.699967] env[61957]: ERROR nova.compute.manager [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] return self._exit_event.wait() [ 665.699967] env[61957]: ERROR nova.compute.manager [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 665.699967] env[61957]: ERROR nova.compute.manager [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] current.throw(*self._exc) [ 665.700382] env[61957]: ERROR nova.compute.manager [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 665.700382] env[61957]: ERROR nova.compute.manager [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] result = function(*args, **kwargs) [ 665.700382] env[61957]: ERROR nova.compute.manager [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 665.700382] env[61957]: ERROR nova.compute.manager [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] return func(*args, **kwargs) [ 665.700382] env[61957]: ERROR nova.compute.manager [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 665.700382] env[61957]: ERROR nova.compute.manager [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] raise e [ 665.700382] env[61957]: ERROR nova.compute.manager [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 665.700382] env[61957]: ERROR nova.compute.manager [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] nwinfo = self.network_api.allocate_for_instance( [ 665.700382] env[61957]: ERROR nova.compute.manager [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 665.700382] env[61957]: ERROR nova.compute.manager [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] created_port_ids = self._update_ports_for_instance( [ 665.700382] env[61957]: ERROR nova.compute.manager [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 665.700382] env[61957]: ERROR nova.compute.manager [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] with excutils.save_and_reraise_exception(): [ 665.700382] env[61957]: ERROR nova.compute.manager [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 665.700737] env[61957]: ERROR nova.compute.manager [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] self.force_reraise() [ 665.700737] env[61957]: ERROR nova.compute.manager [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 665.700737] env[61957]: ERROR nova.compute.manager [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] raise self.value [ 665.700737] env[61957]: ERROR nova.compute.manager [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 665.700737] env[61957]: ERROR nova.compute.manager [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] updated_port = self._update_port( [ 665.700737] env[61957]: ERROR nova.compute.manager [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 665.700737] env[61957]: ERROR nova.compute.manager [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] _ensure_no_port_binding_failure(port) [ 665.700737] env[61957]: ERROR nova.compute.manager [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 665.700737] env[61957]: ERROR nova.compute.manager [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] raise exception.PortBindingFailed(port_id=port['id']) [ 665.700737] env[61957]: ERROR nova.compute.manager [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] nova.exception.PortBindingFailed: Binding failed for port 7ebf5e95-b147-40b2-9271-c193cd89e949, please check neutron logs for more information. [ 665.700737] env[61957]: ERROR nova.compute.manager [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] [ 665.700737] env[61957]: INFO nova.compute.manager [None req-fc6b77c8-b5cd-4608-beab-d2cb5c9f7440 tempest-ImagesNegativeTestJSON-1548152881 tempest-ImagesNegativeTestJSON-1548152881-project-member] [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] Terminating instance [ 665.702149] env[61957]: DEBUG oslo_concurrency.lockutils [None req-fc6b77c8-b5cd-4608-beab-d2cb5c9f7440 tempest-ImagesNegativeTestJSON-1548152881 tempest-ImagesNegativeTestJSON-1548152881-project-member] Acquiring lock "refresh_cache-9df88f23-0219-43e0-b28a-e78f30a473a7" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 665.773196] env[61957]: DEBUG nova.network.neutron [req-f1ec6200-ba8c-4ddd-a5d7-43cc055a799a req-5156cb75-bd32-42b9-96db-d5db3d75f335 service nova] [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 665.856210] env[61957]: DEBUG nova.network.neutron [req-f1ec6200-ba8c-4ddd-a5d7-43cc055a799a req-5156cb75-bd32-42b9-96db-d5db3d75f335 service nova] [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 666.146386] env[61957]: DEBUG oslo_concurrency.lockutils [None req-af971b81-6348-4338-b0a9-b06675d34db5 tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.749s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 666.148486] env[61957]: ERROR nova.compute.manager [None req-af971b81-6348-4338-b0a9-b06675d34db5 tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port b29b2b7c-e839-4ae7-a55b-fc34bee248b2, please check neutron logs for more information. [ 666.148486] env[61957]: ERROR nova.compute.manager [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] Traceback (most recent call last): [ 666.148486] env[61957]: ERROR nova.compute.manager [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 666.148486] env[61957]: ERROR nova.compute.manager [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] self.driver.spawn(context, instance, image_meta, [ 666.148486] env[61957]: ERROR nova.compute.manager [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 666.148486] env[61957]: ERROR nova.compute.manager [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] self._vmops.spawn(context, instance, image_meta, injected_files, [ 666.148486] env[61957]: ERROR nova.compute.manager [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 666.148486] env[61957]: ERROR nova.compute.manager [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] vm_ref = self.build_virtual_machine(instance, [ 666.148486] env[61957]: ERROR nova.compute.manager [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 666.148486] env[61957]: ERROR nova.compute.manager [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] vif_infos = vmwarevif.get_vif_info(self._session, [ 666.148486] env[61957]: ERROR nova.compute.manager [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 666.149031] env[61957]: ERROR nova.compute.manager [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] for vif in network_info: [ 666.149031] env[61957]: ERROR nova.compute.manager [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 666.149031] env[61957]: ERROR nova.compute.manager [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] return self._sync_wrapper(fn, *args, **kwargs) [ 666.149031] env[61957]: ERROR nova.compute.manager [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 666.149031] env[61957]: ERROR nova.compute.manager [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] self.wait() [ 666.149031] env[61957]: ERROR nova.compute.manager [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 666.149031] env[61957]: ERROR nova.compute.manager [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] self[:] = self._gt.wait() [ 666.149031] env[61957]: ERROR nova.compute.manager [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 666.149031] env[61957]: ERROR nova.compute.manager [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] return self._exit_event.wait() [ 666.149031] env[61957]: ERROR nova.compute.manager [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 666.149031] env[61957]: ERROR nova.compute.manager [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] result = hub.switch() [ 666.149031] env[61957]: ERROR nova.compute.manager [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 666.149031] env[61957]: ERROR nova.compute.manager [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] return self.greenlet.switch() [ 666.149381] env[61957]: ERROR nova.compute.manager [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 666.149381] env[61957]: ERROR nova.compute.manager [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] result = function(*args, **kwargs) [ 666.149381] env[61957]: ERROR nova.compute.manager [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 666.149381] env[61957]: ERROR nova.compute.manager [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] return func(*args, **kwargs) [ 666.149381] env[61957]: ERROR nova.compute.manager [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 666.149381] env[61957]: ERROR nova.compute.manager [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] raise e [ 666.149381] env[61957]: ERROR nova.compute.manager [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 666.149381] env[61957]: ERROR nova.compute.manager [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] nwinfo = self.network_api.allocate_for_instance( [ 666.149381] env[61957]: ERROR nova.compute.manager [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 666.149381] env[61957]: ERROR nova.compute.manager [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] created_port_ids = self._update_ports_for_instance( [ 666.149381] env[61957]: ERROR nova.compute.manager [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 666.149381] env[61957]: ERROR nova.compute.manager [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] with excutils.save_and_reraise_exception(): [ 666.149381] env[61957]: ERROR nova.compute.manager [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 666.149760] env[61957]: ERROR nova.compute.manager [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] self.force_reraise() [ 666.149760] env[61957]: ERROR nova.compute.manager [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 666.149760] env[61957]: ERROR nova.compute.manager [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] raise self.value [ 666.149760] env[61957]: ERROR nova.compute.manager [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 666.149760] env[61957]: ERROR nova.compute.manager [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] updated_port = self._update_port( [ 666.149760] env[61957]: ERROR nova.compute.manager [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 666.149760] env[61957]: ERROR nova.compute.manager [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] _ensure_no_port_binding_failure(port) [ 666.149760] env[61957]: ERROR nova.compute.manager [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 666.149760] env[61957]: ERROR nova.compute.manager [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] raise exception.PortBindingFailed(port_id=port['id']) [ 666.149760] env[61957]: ERROR nova.compute.manager [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] nova.exception.PortBindingFailed: Binding failed for port b29b2b7c-e839-4ae7-a55b-fc34bee248b2, please check neutron logs for more information. [ 666.149760] env[61957]: ERROR nova.compute.manager [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] [ 666.150073] env[61957]: DEBUG nova.compute.utils [None req-af971b81-6348-4338-b0a9-b06675d34db5 tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] Binding failed for port b29b2b7c-e839-4ae7-a55b-fc34bee248b2, please check neutron logs for more information. {{(pid=61957) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 666.153779] env[61957]: DEBUG oslo_concurrency.lockutils [None req-3d56c405-76b3-4bdb-9787-ecf239aeeb0d tempest-ImagesOneServerNegativeTestJSON-1581220397 tempest-ImagesOneServerNegativeTestJSON-1581220397-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.637s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 666.153779] env[61957]: INFO nova.compute.claims [None req-3d56c405-76b3-4bdb-9787-ecf239aeeb0d tempest-ImagesOneServerNegativeTestJSON-1581220397 tempest-ImagesOneServerNegativeTestJSON-1581220397-project-member] [instance: 2b151571-348f-4543-a0c5-afe6458b1973] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 666.159676] env[61957]: DEBUG nova.compute.manager [None req-af971b81-6348-4338-b0a9-b06675d34db5 tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] Build of instance 967056dd-1679-421c-b4fd-9d663b1411e2 was re-scheduled: Binding failed for port b29b2b7c-e839-4ae7-a55b-fc34bee248b2, please check neutron logs for more information. {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 666.159676] env[61957]: DEBUG nova.compute.manager [None req-af971b81-6348-4338-b0a9-b06675d34db5 tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] Unplugging VIFs for instance {{(pid=61957) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 666.159676] env[61957]: DEBUG oslo_concurrency.lockutils [None req-af971b81-6348-4338-b0a9-b06675d34db5 tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] Acquiring lock "refresh_cache-967056dd-1679-421c-b4fd-9d663b1411e2" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 666.159676] env[61957]: DEBUG oslo_concurrency.lockutils [None req-af971b81-6348-4338-b0a9-b06675d34db5 tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] Acquired lock "refresh_cache-967056dd-1679-421c-b4fd-9d663b1411e2" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 666.160016] env[61957]: DEBUG nova.network.neutron [None req-af971b81-6348-4338-b0a9-b06675d34db5 tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 666.160016] env[61957]: DEBUG oslo_concurrency.lockutils [None req-dbfaab8a-0784-4441-ad61-79e56d9c3c60 tempest-FloatingIPsAssociationTestJSON-1786153562 tempest-FloatingIPsAssociationTestJSON-1786153562-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 666.359550] env[61957]: DEBUG oslo_concurrency.lockutils [req-f1ec6200-ba8c-4ddd-a5d7-43cc055a799a req-5156cb75-bd32-42b9-96db-d5db3d75f335 service nova] Releasing lock "refresh_cache-9df88f23-0219-43e0-b28a-e78f30a473a7" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 666.359926] env[61957]: DEBUG oslo_concurrency.lockutils [None req-fc6b77c8-b5cd-4608-beab-d2cb5c9f7440 tempest-ImagesNegativeTestJSON-1548152881 tempest-ImagesNegativeTestJSON-1548152881-project-member] Acquired lock "refresh_cache-9df88f23-0219-43e0-b28a-e78f30a473a7" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 666.360174] env[61957]: DEBUG nova.network.neutron [None req-fc6b77c8-b5cd-4608-beab-d2cb5c9f7440 tempest-ImagesNegativeTestJSON-1548152881 tempest-ImagesNegativeTestJSON-1548152881-project-member] [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 666.684942] env[61957]: DEBUG nova.network.neutron [None req-af971b81-6348-4338-b0a9-b06675d34db5 tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 666.776287] env[61957]: DEBUG nova.network.neutron [None req-af971b81-6348-4338-b0a9-b06675d34db5 tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 666.880288] env[61957]: DEBUG nova.network.neutron [None req-fc6b77c8-b5cd-4608-beab-d2cb5c9f7440 tempest-ImagesNegativeTestJSON-1548152881 tempest-ImagesNegativeTestJSON-1548152881-project-member] [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 666.954742] env[61957]: DEBUG nova.network.neutron [None req-fc6b77c8-b5cd-4608-beab-d2cb5c9f7440 tempest-ImagesNegativeTestJSON-1548152881 tempest-ImagesNegativeTestJSON-1548152881-project-member] [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 667.281799] env[61957]: DEBUG oslo_concurrency.lockutils [None req-af971b81-6348-4338-b0a9-b06675d34db5 tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] Releasing lock "refresh_cache-967056dd-1679-421c-b4fd-9d663b1411e2" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 667.281909] env[61957]: DEBUG nova.compute.manager [None req-af971b81-6348-4338-b0a9-b06675d34db5 tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61957) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 667.282101] env[61957]: DEBUG nova.compute.manager [None req-af971b81-6348-4338-b0a9-b06675d34db5 tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 667.282266] env[61957]: DEBUG nova.network.neutron [None req-af971b81-6348-4338-b0a9-b06675d34db5 tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 667.313759] env[61957]: DEBUG nova.compute.manager [req-178144bd-d57d-4c2e-b0b3-7f2ad3ac695c req-e84d4453-a2fc-4d04-9d72-537c95848475 service nova] [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] Received event network-vif-deleted-7ebf5e95-b147-40b2-9271-c193cd89e949 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 667.443083] env[61957]: DEBUG nova.network.neutron [None req-af971b81-6348-4338-b0a9-b06675d34db5 tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 667.456969] env[61957]: DEBUG oslo_concurrency.lockutils [None req-fc6b77c8-b5cd-4608-beab-d2cb5c9f7440 tempest-ImagesNegativeTestJSON-1548152881 tempest-ImagesNegativeTestJSON-1548152881-project-member] Releasing lock "refresh_cache-9df88f23-0219-43e0-b28a-e78f30a473a7" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 667.457429] env[61957]: DEBUG nova.compute.manager [None req-fc6b77c8-b5cd-4608-beab-d2cb5c9f7440 tempest-ImagesNegativeTestJSON-1548152881 tempest-ImagesNegativeTestJSON-1548152881-project-member] [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] Start destroying the instance on the hypervisor. {{(pid=61957) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 667.457616] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-fc6b77c8-b5cd-4608-beab-d2cb5c9f7440 tempest-ImagesNegativeTestJSON-1548152881 tempest-ImagesNegativeTestJSON-1548152881-project-member] [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] Destroying instance {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 667.457922] env[61957]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-bf773591-bee9-4a13-a95c-1218f8c08f85 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.467456] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1954666-7825-4954-9904-08bf0dc8ea77 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.492743] env[61957]: WARNING nova.virt.vmwareapi.vmops [None req-fc6b77c8-b5cd-4608-beab-d2cb5c9f7440 tempest-ImagesNegativeTestJSON-1548152881 tempest-ImagesNegativeTestJSON-1548152881-project-member] [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 9df88f23-0219-43e0-b28a-e78f30a473a7 could not be found. [ 667.492980] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-fc6b77c8-b5cd-4608-beab-d2cb5c9f7440 tempest-ImagesNegativeTestJSON-1548152881 tempest-ImagesNegativeTestJSON-1548152881-project-member] [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] Instance destroyed {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 667.493185] env[61957]: INFO nova.compute.manager [None req-fc6b77c8-b5cd-4608-beab-d2cb5c9f7440 tempest-ImagesNegativeTestJSON-1548152881 tempest-ImagesNegativeTestJSON-1548152881-project-member] [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] Took 0.04 seconds to destroy the instance on the hypervisor. [ 667.494250] env[61957]: DEBUG oslo.service.loopingcall [None req-fc6b77c8-b5cd-4608-beab-d2cb5c9f7440 tempest-ImagesNegativeTestJSON-1548152881 tempest-ImagesNegativeTestJSON-1548152881-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 667.496065] env[61957]: DEBUG nova.compute.manager [-] [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 667.496065] env[61957]: DEBUG nova.network.neutron [-] [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 667.514662] env[61957]: DEBUG nova.network.neutron [-] [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 667.681109] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5028f859-c098-4ad3-a3e9-8d5a5f5d7c37 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.688745] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21297102-ddf4-400e-947a-60bdf9cb70a8 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.719201] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6553e858-6f79-448d-a51c-fe1f24ae7302 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.726743] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e270a63e-8a23-4996-a33f-5b2a54328134 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.740902] env[61957]: DEBUG nova.compute.provider_tree [None req-3d56c405-76b3-4bdb-9787-ecf239aeeb0d tempest-ImagesOneServerNegativeTestJSON-1581220397 tempest-ImagesOneServerNegativeTestJSON-1581220397-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 667.945182] env[61957]: DEBUG nova.network.neutron [None req-af971b81-6348-4338-b0a9-b06675d34db5 tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 668.017267] env[61957]: DEBUG nova.network.neutron [-] [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 668.244364] env[61957]: DEBUG nova.scheduler.client.report [None req-3d56c405-76b3-4bdb-9787-ecf239aeeb0d tempest-ImagesOneServerNegativeTestJSON-1581220397 tempest-ImagesOneServerNegativeTestJSON-1581220397-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 668.448043] env[61957]: INFO nova.compute.manager [None req-af971b81-6348-4338-b0a9-b06675d34db5 tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] [instance: 967056dd-1679-421c-b4fd-9d663b1411e2] Took 1.17 seconds to deallocate network for instance. [ 668.520880] env[61957]: INFO nova.compute.manager [-] [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] Took 1.02 seconds to deallocate network for instance. [ 668.524234] env[61957]: DEBUG nova.compute.claims [None req-fc6b77c8-b5cd-4608-beab-d2cb5c9f7440 tempest-ImagesNegativeTestJSON-1548152881 tempest-ImagesNegativeTestJSON-1548152881-project-member] [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] Aborting claim: {{(pid=61957) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 668.524234] env[61957]: DEBUG oslo_concurrency.lockutils [None req-fc6b77c8-b5cd-4608-beab-d2cb5c9f7440 tempest-ImagesNegativeTestJSON-1548152881 tempest-ImagesNegativeTestJSON-1548152881-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 668.749328] env[61957]: DEBUG oslo_concurrency.lockutils [None req-3d56c405-76b3-4bdb-9787-ecf239aeeb0d tempest-ImagesOneServerNegativeTestJSON-1581220397 tempest-ImagesOneServerNegativeTestJSON-1581220397-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.597s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 668.750464] env[61957]: DEBUG nova.compute.manager [None req-3d56c405-76b3-4bdb-9787-ecf239aeeb0d tempest-ImagesOneServerNegativeTestJSON-1581220397 tempest-ImagesOneServerNegativeTestJSON-1581220397-project-member] [instance: 2b151571-348f-4543-a0c5-afe6458b1973] Start building networks asynchronously for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 668.754951] env[61957]: DEBUG oslo_concurrency.lockutils [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.468s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 669.262558] env[61957]: DEBUG nova.compute.utils [None req-3d56c405-76b3-4bdb-9787-ecf239aeeb0d tempest-ImagesOneServerNegativeTestJSON-1581220397 tempest-ImagesOneServerNegativeTestJSON-1581220397-project-member] Using /dev/sd instead of None {{(pid=61957) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 669.264927] env[61957]: DEBUG nova.compute.manager [None req-3d56c405-76b3-4bdb-9787-ecf239aeeb0d tempest-ImagesOneServerNegativeTestJSON-1581220397 tempest-ImagesOneServerNegativeTestJSON-1581220397-project-member] [instance: 2b151571-348f-4543-a0c5-afe6458b1973] Allocating IP information in the background. {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 669.265131] env[61957]: DEBUG nova.network.neutron [None req-3d56c405-76b3-4bdb-9787-ecf239aeeb0d tempest-ImagesOneServerNegativeTestJSON-1581220397 tempest-ImagesOneServerNegativeTestJSON-1581220397-project-member] [instance: 2b151571-348f-4543-a0c5-afe6458b1973] allocate_for_instance() {{(pid=61957) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 669.315498] env[61957]: DEBUG nova.policy [None req-3d56c405-76b3-4bdb-9787-ecf239aeeb0d tempest-ImagesOneServerNegativeTestJSON-1581220397 tempest-ImagesOneServerNegativeTestJSON-1581220397-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '17df66f0ffeb43afb366a927e10341d6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8bc28221cdea4821bd1671a37569b4fd', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61957) authorize /opt/stack/nova/nova/policy.py:203}} [ 669.481222] env[61957]: INFO nova.scheduler.client.report [None req-af971b81-6348-4338-b0a9-b06675d34db5 tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] Deleted allocations for instance 967056dd-1679-421c-b4fd-9d663b1411e2 [ 669.603755] env[61957]: DEBUG nova.network.neutron [None req-3d56c405-76b3-4bdb-9787-ecf239aeeb0d tempest-ImagesOneServerNegativeTestJSON-1581220397 tempest-ImagesOneServerNegativeTestJSON-1581220397-project-member] [instance: 2b151571-348f-4543-a0c5-afe6458b1973] Successfully created port: 892c3e98-da9b-4830-8c81-221309a32b88 {{(pid=61957) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 669.768668] env[61957]: DEBUG nova.compute.manager [None req-3d56c405-76b3-4bdb-9787-ecf239aeeb0d tempest-ImagesOneServerNegativeTestJSON-1581220397 tempest-ImagesOneServerNegativeTestJSON-1581220397-project-member] [instance: 2b151571-348f-4543-a0c5-afe6458b1973] Start building block device mappings for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 669.800285] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a8c591d-61e4-4e35-9889-d901b9b637cc {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.813153] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8e83060-0ae2-4cba-b70b-d9146d988936 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.847486] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4f6ac49-5a40-407f-9a45-cb7d90015a7e {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.855865] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5f784d5-c239-4801-9222-6128002eea3d {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.872411] env[61957]: DEBUG nova.compute.provider_tree [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 669.997197] env[61957]: DEBUG oslo_concurrency.lockutils [None req-af971b81-6348-4338-b0a9-b06675d34db5 tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] Lock "967056dd-1679-421c-b4fd-9d663b1411e2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 76.742s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 670.041957] env[61957]: DEBUG oslo_concurrency.lockutils [None req-182a8e01-a1e1-4ecf-8692-2abcfc9944ba tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Acquiring lock "e0249936-d616-4ffb-8f77-d8107633c42a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 670.041957] env[61957]: DEBUG oslo_concurrency.lockutils [None req-182a8e01-a1e1-4ecf-8692-2abcfc9944ba tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Lock "e0249936-d616-4ffb-8f77-d8107633c42a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 670.376761] env[61957]: DEBUG nova.scheduler.client.report [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 670.464345] env[61957]: DEBUG nova.compute.manager [req-dac0a6e4-934d-4871-b1aa-beb0a7af8ef7 req-46289399-5dfe-43df-a920-3763b9586b78 service nova] [instance: 2b151571-348f-4543-a0c5-afe6458b1973] Received event network-changed-892c3e98-da9b-4830-8c81-221309a32b88 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 670.464548] env[61957]: DEBUG nova.compute.manager [req-dac0a6e4-934d-4871-b1aa-beb0a7af8ef7 req-46289399-5dfe-43df-a920-3763b9586b78 service nova] [instance: 2b151571-348f-4543-a0c5-afe6458b1973] Refreshing instance network info cache due to event network-changed-892c3e98-da9b-4830-8c81-221309a32b88. {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 670.464757] env[61957]: DEBUG oslo_concurrency.lockutils [req-dac0a6e4-934d-4871-b1aa-beb0a7af8ef7 req-46289399-5dfe-43df-a920-3763b9586b78 service nova] Acquiring lock "refresh_cache-2b151571-348f-4543-a0c5-afe6458b1973" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 670.464895] env[61957]: DEBUG oslo_concurrency.lockutils [req-dac0a6e4-934d-4871-b1aa-beb0a7af8ef7 req-46289399-5dfe-43df-a920-3763b9586b78 service nova] Acquired lock "refresh_cache-2b151571-348f-4543-a0c5-afe6458b1973" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 670.465306] env[61957]: DEBUG nova.network.neutron [req-dac0a6e4-934d-4871-b1aa-beb0a7af8ef7 req-46289399-5dfe-43df-a920-3763b9586b78 service nova] [instance: 2b151571-348f-4543-a0c5-afe6458b1973] Refreshing network info cache for port 892c3e98-da9b-4830-8c81-221309a32b88 {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 670.499013] env[61957]: DEBUG nova.compute.manager [None req-c2f4fc75-b8bd-417b-8703-3b9048472a92 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] Starting instance... {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 670.697256] env[61957]: ERROR nova.compute.manager [None req-3d56c405-76b3-4bdb-9787-ecf239aeeb0d tempest-ImagesOneServerNegativeTestJSON-1581220397 tempest-ImagesOneServerNegativeTestJSON-1581220397-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 892c3e98-da9b-4830-8c81-221309a32b88, please check neutron logs for more information. [ 670.697256] env[61957]: ERROR nova.compute.manager Traceback (most recent call last): [ 670.697256] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 670.697256] env[61957]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 670.697256] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 670.697256] env[61957]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 670.697256] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 670.697256] env[61957]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 670.697256] env[61957]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 670.697256] env[61957]: ERROR nova.compute.manager self.force_reraise() [ 670.697256] env[61957]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 670.697256] env[61957]: ERROR nova.compute.manager raise self.value [ 670.697256] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 670.697256] env[61957]: ERROR nova.compute.manager updated_port = self._update_port( [ 670.697256] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 670.697256] env[61957]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 670.697737] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 670.697737] env[61957]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 670.697737] env[61957]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 892c3e98-da9b-4830-8c81-221309a32b88, please check neutron logs for more information. [ 670.697737] env[61957]: ERROR nova.compute.manager [ 670.697737] env[61957]: Traceback (most recent call last): [ 670.697737] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 670.697737] env[61957]: listener.cb(fileno) [ 670.697737] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 670.697737] env[61957]: result = function(*args, **kwargs) [ 670.697737] env[61957]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 670.697737] env[61957]: return func(*args, **kwargs) [ 670.697737] env[61957]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 670.697737] env[61957]: raise e [ 670.697737] env[61957]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 670.697737] env[61957]: nwinfo = self.network_api.allocate_for_instance( [ 670.697737] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 670.697737] env[61957]: created_port_ids = self._update_ports_for_instance( [ 670.697737] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 670.697737] env[61957]: with excutils.save_and_reraise_exception(): [ 670.697737] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 670.697737] env[61957]: self.force_reraise() [ 670.697737] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 670.697737] env[61957]: raise self.value [ 670.697737] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 670.697737] env[61957]: updated_port = self._update_port( [ 670.697737] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 670.697737] env[61957]: _ensure_no_port_binding_failure(port) [ 670.697737] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 670.697737] env[61957]: raise exception.PortBindingFailed(port_id=port['id']) [ 670.698514] env[61957]: nova.exception.PortBindingFailed: Binding failed for port 892c3e98-da9b-4830-8c81-221309a32b88, please check neutron logs for more information. [ 670.698514] env[61957]: Removing descriptor: 17 [ 670.790155] env[61957]: DEBUG nova.compute.manager [None req-3d56c405-76b3-4bdb-9787-ecf239aeeb0d tempest-ImagesOneServerNegativeTestJSON-1581220397 tempest-ImagesOneServerNegativeTestJSON-1581220397-project-member] [instance: 2b151571-348f-4543-a0c5-afe6458b1973] Start spawning the instance on the hypervisor. {{(pid=61957) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 670.817711] env[61957]: DEBUG nova.virt.hardware [None req-3d56c405-76b3-4bdb-9787-ecf239aeeb0d tempest-ImagesOneServerNegativeTestJSON-1581220397 tempest-ImagesOneServerNegativeTestJSON-1581220397-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T17:22:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T17:21:50Z,direct_url=,disk_format='vmdk',id=11c76a2c-f705-470a-ba9d-4657858bab38,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='81d3096d0f094373913fc3dc8f5c3c6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T17:21:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 670.819104] env[61957]: DEBUG nova.virt.hardware [None req-3d56c405-76b3-4bdb-9787-ecf239aeeb0d tempest-ImagesOneServerNegativeTestJSON-1581220397 tempest-ImagesOneServerNegativeTestJSON-1581220397-project-member] Flavor limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 670.819104] env[61957]: DEBUG nova.virt.hardware [None req-3d56c405-76b3-4bdb-9787-ecf239aeeb0d tempest-ImagesOneServerNegativeTestJSON-1581220397 tempest-ImagesOneServerNegativeTestJSON-1581220397-project-member] Image limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 670.819104] env[61957]: DEBUG nova.virt.hardware [None req-3d56c405-76b3-4bdb-9787-ecf239aeeb0d tempest-ImagesOneServerNegativeTestJSON-1581220397 tempest-ImagesOneServerNegativeTestJSON-1581220397-project-member] Flavor pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 670.819104] env[61957]: DEBUG nova.virt.hardware [None req-3d56c405-76b3-4bdb-9787-ecf239aeeb0d tempest-ImagesOneServerNegativeTestJSON-1581220397 tempest-ImagesOneServerNegativeTestJSON-1581220397-project-member] Image pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 670.819104] env[61957]: DEBUG nova.virt.hardware [None req-3d56c405-76b3-4bdb-9787-ecf239aeeb0d tempest-ImagesOneServerNegativeTestJSON-1581220397 tempest-ImagesOneServerNegativeTestJSON-1581220397-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 670.819394] env[61957]: DEBUG nova.virt.hardware [None req-3d56c405-76b3-4bdb-9787-ecf239aeeb0d tempest-ImagesOneServerNegativeTestJSON-1581220397 tempest-ImagesOneServerNegativeTestJSON-1581220397-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 670.819738] env[61957]: DEBUG nova.virt.hardware [None req-3d56c405-76b3-4bdb-9787-ecf239aeeb0d tempest-ImagesOneServerNegativeTestJSON-1581220397 tempest-ImagesOneServerNegativeTestJSON-1581220397-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 670.820262] env[61957]: DEBUG nova.virt.hardware [None req-3d56c405-76b3-4bdb-9787-ecf239aeeb0d tempest-ImagesOneServerNegativeTestJSON-1581220397 tempest-ImagesOneServerNegativeTestJSON-1581220397-project-member] Got 1 possible topologies {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 670.820563] env[61957]: DEBUG nova.virt.hardware [None req-3d56c405-76b3-4bdb-9787-ecf239aeeb0d tempest-ImagesOneServerNegativeTestJSON-1581220397 tempest-ImagesOneServerNegativeTestJSON-1581220397-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 670.820902] env[61957]: DEBUG nova.virt.hardware [None req-3d56c405-76b3-4bdb-9787-ecf239aeeb0d tempest-ImagesOneServerNegativeTestJSON-1581220397 tempest-ImagesOneServerNegativeTestJSON-1581220397-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 670.822048] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0cb70852-c43f-4d8d-9343-bb7388dc23c1 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.830518] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-134704b0-3cb5-45fa-9c39-fe7ee1be0056 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.844946] env[61957]: ERROR nova.compute.manager [None req-3d56c405-76b3-4bdb-9787-ecf239aeeb0d tempest-ImagesOneServerNegativeTestJSON-1581220397 tempest-ImagesOneServerNegativeTestJSON-1581220397-project-member] [instance: 2b151571-348f-4543-a0c5-afe6458b1973] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 892c3e98-da9b-4830-8c81-221309a32b88, please check neutron logs for more information. [ 670.844946] env[61957]: ERROR nova.compute.manager [instance: 2b151571-348f-4543-a0c5-afe6458b1973] Traceback (most recent call last): [ 670.844946] env[61957]: ERROR nova.compute.manager [instance: 2b151571-348f-4543-a0c5-afe6458b1973] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 670.844946] env[61957]: ERROR nova.compute.manager [instance: 2b151571-348f-4543-a0c5-afe6458b1973] yield resources [ 670.844946] env[61957]: ERROR nova.compute.manager [instance: 2b151571-348f-4543-a0c5-afe6458b1973] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 670.844946] env[61957]: ERROR nova.compute.manager [instance: 2b151571-348f-4543-a0c5-afe6458b1973] self.driver.spawn(context, instance, image_meta, [ 670.844946] env[61957]: ERROR nova.compute.manager [instance: 2b151571-348f-4543-a0c5-afe6458b1973] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 670.844946] env[61957]: ERROR nova.compute.manager [instance: 2b151571-348f-4543-a0c5-afe6458b1973] self._vmops.spawn(context, instance, image_meta, injected_files, [ 670.844946] env[61957]: ERROR nova.compute.manager [instance: 2b151571-348f-4543-a0c5-afe6458b1973] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 670.844946] env[61957]: ERROR nova.compute.manager [instance: 2b151571-348f-4543-a0c5-afe6458b1973] vm_ref = self.build_virtual_machine(instance, [ 670.844946] env[61957]: ERROR nova.compute.manager [instance: 2b151571-348f-4543-a0c5-afe6458b1973] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 670.845360] env[61957]: ERROR nova.compute.manager [instance: 2b151571-348f-4543-a0c5-afe6458b1973] vif_infos = vmwarevif.get_vif_info(self._session, [ 670.845360] env[61957]: ERROR nova.compute.manager [instance: 2b151571-348f-4543-a0c5-afe6458b1973] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 670.845360] env[61957]: ERROR nova.compute.manager [instance: 2b151571-348f-4543-a0c5-afe6458b1973] for vif in network_info: [ 670.845360] env[61957]: ERROR nova.compute.manager [instance: 2b151571-348f-4543-a0c5-afe6458b1973] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 670.845360] env[61957]: ERROR nova.compute.manager [instance: 2b151571-348f-4543-a0c5-afe6458b1973] return self._sync_wrapper(fn, *args, **kwargs) [ 670.845360] env[61957]: ERROR nova.compute.manager [instance: 2b151571-348f-4543-a0c5-afe6458b1973] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 670.845360] env[61957]: ERROR nova.compute.manager [instance: 2b151571-348f-4543-a0c5-afe6458b1973] self.wait() [ 670.845360] env[61957]: ERROR nova.compute.manager [instance: 2b151571-348f-4543-a0c5-afe6458b1973] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 670.845360] env[61957]: ERROR nova.compute.manager [instance: 2b151571-348f-4543-a0c5-afe6458b1973] self[:] = self._gt.wait() [ 670.845360] env[61957]: ERROR nova.compute.manager [instance: 2b151571-348f-4543-a0c5-afe6458b1973] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 670.845360] env[61957]: ERROR nova.compute.manager [instance: 2b151571-348f-4543-a0c5-afe6458b1973] return self._exit_event.wait() [ 670.845360] env[61957]: ERROR nova.compute.manager [instance: 2b151571-348f-4543-a0c5-afe6458b1973] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 670.845360] env[61957]: ERROR nova.compute.manager [instance: 2b151571-348f-4543-a0c5-afe6458b1973] current.throw(*self._exc) [ 670.845939] env[61957]: ERROR nova.compute.manager [instance: 2b151571-348f-4543-a0c5-afe6458b1973] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 670.845939] env[61957]: ERROR nova.compute.manager [instance: 2b151571-348f-4543-a0c5-afe6458b1973] result = function(*args, **kwargs) [ 670.845939] env[61957]: ERROR nova.compute.manager [instance: 2b151571-348f-4543-a0c5-afe6458b1973] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 670.845939] env[61957]: ERROR nova.compute.manager [instance: 2b151571-348f-4543-a0c5-afe6458b1973] return func(*args, **kwargs) [ 670.845939] env[61957]: ERROR nova.compute.manager [instance: 2b151571-348f-4543-a0c5-afe6458b1973] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 670.845939] env[61957]: ERROR nova.compute.manager [instance: 2b151571-348f-4543-a0c5-afe6458b1973] raise e [ 670.845939] env[61957]: ERROR nova.compute.manager [instance: 2b151571-348f-4543-a0c5-afe6458b1973] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 670.845939] env[61957]: ERROR nova.compute.manager [instance: 2b151571-348f-4543-a0c5-afe6458b1973] nwinfo = self.network_api.allocate_for_instance( [ 670.845939] env[61957]: ERROR nova.compute.manager [instance: 2b151571-348f-4543-a0c5-afe6458b1973] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 670.845939] env[61957]: ERROR nova.compute.manager [instance: 2b151571-348f-4543-a0c5-afe6458b1973] created_port_ids = self._update_ports_for_instance( [ 670.845939] env[61957]: ERROR nova.compute.manager [instance: 2b151571-348f-4543-a0c5-afe6458b1973] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 670.845939] env[61957]: ERROR nova.compute.manager [instance: 2b151571-348f-4543-a0c5-afe6458b1973] with excutils.save_and_reraise_exception(): [ 670.845939] env[61957]: ERROR nova.compute.manager [instance: 2b151571-348f-4543-a0c5-afe6458b1973] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 670.846267] env[61957]: ERROR nova.compute.manager [instance: 2b151571-348f-4543-a0c5-afe6458b1973] self.force_reraise() [ 670.846267] env[61957]: ERROR nova.compute.manager [instance: 2b151571-348f-4543-a0c5-afe6458b1973] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 670.846267] env[61957]: ERROR nova.compute.manager [instance: 2b151571-348f-4543-a0c5-afe6458b1973] raise self.value [ 670.846267] env[61957]: ERROR nova.compute.manager [instance: 2b151571-348f-4543-a0c5-afe6458b1973] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 670.846267] env[61957]: ERROR nova.compute.manager [instance: 2b151571-348f-4543-a0c5-afe6458b1973] updated_port = self._update_port( [ 670.846267] env[61957]: ERROR nova.compute.manager [instance: 2b151571-348f-4543-a0c5-afe6458b1973] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 670.846267] env[61957]: ERROR nova.compute.manager [instance: 2b151571-348f-4543-a0c5-afe6458b1973] _ensure_no_port_binding_failure(port) [ 670.846267] env[61957]: ERROR nova.compute.manager [instance: 2b151571-348f-4543-a0c5-afe6458b1973] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 670.846267] env[61957]: ERROR nova.compute.manager [instance: 2b151571-348f-4543-a0c5-afe6458b1973] raise exception.PortBindingFailed(port_id=port['id']) [ 670.846267] env[61957]: ERROR nova.compute.manager [instance: 2b151571-348f-4543-a0c5-afe6458b1973] nova.exception.PortBindingFailed: Binding failed for port 892c3e98-da9b-4830-8c81-221309a32b88, please check neutron logs for more information. [ 670.846267] env[61957]: ERROR nova.compute.manager [instance: 2b151571-348f-4543-a0c5-afe6458b1973] [ 670.846267] env[61957]: INFO nova.compute.manager [None req-3d56c405-76b3-4bdb-9787-ecf239aeeb0d tempest-ImagesOneServerNegativeTestJSON-1581220397 tempest-ImagesOneServerNegativeTestJSON-1581220397-project-member] [instance: 2b151571-348f-4543-a0c5-afe6458b1973] Terminating instance [ 670.847223] env[61957]: DEBUG oslo_concurrency.lockutils [None req-3d56c405-76b3-4bdb-9787-ecf239aeeb0d tempest-ImagesOneServerNegativeTestJSON-1581220397 tempest-ImagesOneServerNegativeTestJSON-1581220397-project-member] Acquiring lock "refresh_cache-2b151571-348f-4543-a0c5-afe6458b1973" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 670.883197] env[61957]: DEBUG oslo_concurrency.lockutils [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.129s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 670.883835] env[61957]: ERROR nova.compute.manager [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 9758f59c-2ebd-4314-8d87-c51d41bbbcbe, please check neutron logs for more information. [ 670.883835] env[61957]: ERROR nova.compute.manager [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] Traceback (most recent call last): [ 670.883835] env[61957]: ERROR nova.compute.manager [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 670.883835] env[61957]: ERROR nova.compute.manager [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] self.driver.spawn(context, instance, image_meta, [ 670.883835] env[61957]: ERROR nova.compute.manager [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 670.883835] env[61957]: ERROR nova.compute.manager [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 670.883835] env[61957]: ERROR nova.compute.manager [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 670.883835] env[61957]: ERROR nova.compute.manager [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] vm_ref = self.build_virtual_machine(instance, [ 670.883835] env[61957]: ERROR nova.compute.manager [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 670.883835] env[61957]: ERROR nova.compute.manager [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] vif_infos = vmwarevif.get_vif_info(self._session, [ 670.883835] env[61957]: ERROR nova.compute.manager [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 670.884168] env[61957]: ERROR nova.compute.manager [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] for vif in network_info: [ 670.884168] env[61957]: ERROR nova.compute.manager [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 670.884168] env[61957]: ERROR nova.compute.manager [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] return self._sync_wrapper(fn, *args, **kwargs) [ 670.884168] env[61957]: ERROR nova.compute.manager [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 670.884168] env[61957]: ERROR nova.compute.manager [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] self.wait() [ 670.884168] env[61957]: ERROR nova.compute.manager [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 670.884168] env[61957]: ERROR nova.compute.manager [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] self[:] = self._gt.wait() [ 670.884168] env[61957]: ERROR nova.compute.manager [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 670.884168] env[61957]: ERROR nova.compute.manager [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] return self._exit_event.wait() [ 670.884168] env[61957]: ERROR nova.compute.manager [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 670.884168] env[61957]: ERROR nova.compute.manager [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] result = hub.switch() [ 670.884168] env[61957]: ERROR nova.compute.manager [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 670.884168] env[61957]: ERROR nova.compute.manager [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] return self.greenlet.switch() [ 670.884585] env[61957]: ERROR nova.compute.manager [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 670.884585] env[61957]: ERROR nova.compute.manager [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] result = function(*args, **kwargs) [ 670.884585] env[61957]: ERROR nova.compute.manager [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 670.884585] env[61957]: ERROR nova.compute.manager [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] return func(*args, **kwargs) [ 670.884585] env[61957]: ERROR nova.compute.manager [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 670.884585] env[61957]: ERROR nova.compute.manager [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] raise e [ 670.884585] env[61957]: ERROR nova.compute.manager [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 670.884585] env[61957]: ERROR nova.compute.manager [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] nwinfo = self.network_api.allocate_for_instance( [ 670.884585] env[61957]: ERROR nova.compute.manager [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 670.884585] env[61957]: ERROR nova.compute.manager [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] created_port_ids = self._update_ports_for_instance( [ 670.884585] env[61957]: ERROR nova.compute.manager [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 670.884585] env[61957]: ERROR nova.compute.manager [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] with excutils.save_and_reraise_exception(): [ 670.884585] env[61957]: ERROR nova.compute.manager [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 670.885937] env[61957]: ERROR nova.compute.manager [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] self.force_reraise() [ 670.885937] env[61957]: ERROR nova.compute.manager [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 670.885937] env[61957]: ERROR nova.compute.manager [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] raise self.value [ 670.885937] env[61957]: ERROR nova.compute.manager [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 670.885937] env[61957]: ERROR nova.compute.manager [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] updated_port = self._update_port( [ 670.885937] env[61957]: ERROR nova.compute.manager [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 670.885937] env[61957]: ERROR nova.compute.manager [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] _ensure_no_port_binding_failure(port) [ 670.885937] env[61957]: ERROR nova.compute.manager [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 670.885937] env[61957]: ERROR nova.compute.manager [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] raise exception.PortBindingFailed(port_id=port['id']) [ 670.885937] env[61957]: ERROR nova.compute.manager [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] nova.exception.PortBindingFailed: Binding failed for port 9758f59c-2ebd-4314-8d87-c51d41bbbcbe, please check neutron logs for more information. [ 670.885937] env[61957]: ERROR nova.compute.manager [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] [ 670.886246] env[61957]: DEBUG nova.compute.utils [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] Binding failed for port 9758f59c-2ebd-4314-8d87-c51d41bbbcbe, please check neutron logs for more information. {{(pid=61957) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 670.886448] env[61957]: DEBUG nova.compute.manager [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] Build of instance 74d3de77-1272-449d-8b64-75e21fff8d7f was re-scheduled: Binding failed for port 9758f59c-2ebd-4314-8d87-c51d41bbbcbe, please check neutron logs for more information. {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 670.886991] env[61957]: DEBUG nova.compute.manager [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] Unplugging VIFs for instance {{(pid=61957) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 670.887292] env[61957]: DEBUG oslo_concurrency.lockutils [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] Acquiring lock "refresh_cache-74d3de77-1272-449d-8b64-75e21fff8d7f" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 670.887474] env[61957]: DEBUG oslo_concurrency.lockutils [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] Acquired lock "refresh_cache-74d3de77-1272-449d-8b64-75e21fff8d7f" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 670.887662] env[61957]: DEBUG nova.network.neutron [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 670.892061] env[61957]: DEBUG oslo_concurrency.lockutils [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.854s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 670.984745] env[61957]: DEBUG nova.network.neutron [req-dac0a6e4-934d-4871-b1aa-beb0a7af8ef7 req-46289399-5dfe-43df-a920-3763b9586b78 service nova] [instance: 2b151571-348f-4543-a0c5-afe6458b1973] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 671.027240] env[61957]: DEBUG oslo_concurrency.lockutils [None req-c2f4fc75-b8bd-417b-8703-3b9048472a92 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 671.119392] env[61957]: DEBUG nova.network.neutron [req-dac0a6e4-934d-4871-b1aa-beb0a7af8ef7 req-46289399-5dfe-43df-a920-3763b9586b78 service nova] [instance: 2b151571-348f-4543-a0c5-afe6458b1973] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 671.415600] env[61957]: DEBUG nova.network.neutron [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 671.489751] env[61957]: DEBUG nova.network.neutron [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 671.621941] env[61957]: DEBUG oslo_concurrency.lockutils [req-dac0a6e4-934d-4871-b1aa-beb0a7af8ef7 req-46289399-5dfe-43df-a920-3763b9586b78 service nova] Releasing lock "refresh_cache-2b151571-348f-4543-a0c5-afe6458b1973" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 671.622408] env[61957]: DEBUG oslo_concurrency.lockutils [None req-3d56c405-76b3-4bdb-9787-ecf239aeeb0d tempest-ImagesOneServerNegativeTestJSON-1581220397 tempest-ImagesOneServerNegativeTestJSON-1581220397-project-member] Acquired lock "refresh_cache-2b151571-348f-4543-a0c5-afe6458b1973" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 671.622630] env[61957]: DEBUG nova.network.neutron [None req-3d56c405-76b3-4bdb-9787-ecf239aeeb0d tempest-ImagesOneServerNegativeTestJSON-1581220397 tempest-ImagesOneServerNegativeTestJSON-1581220397-project-member] [instance: 2b151571-348f-4543-a0c5-afe6458b1973] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 671.791747] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-379ea34e-5104-49ec-b8cc-3737c8e8c5d2 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.799660] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f78d0e71-3724-4023-8fd5-76b4e09cd3a0 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.834082] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1d2a948-5c42-46ad-a3ce-3ec50ca2c831 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.840675] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40cac235-fc4b-4541-96af-dff68b89f54e {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.857740] env[61957]: DEBUG nova.compute.provider_tree [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 671.994887] env[61957]: DEBUG oslo_concurrency.lockutils [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] Releasing lock "refresh_cache-74d3de77-1272-449d-8b64-75e21fff8d7f" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 671.995364] env[61957]: DEBUG nova.compute.manager [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61957) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 671.997497] env[61957]: DEBUG nova.compute.manager [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 671.997497] env[61957]: DEBUG nova.network.neutron [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 672.018980] env[61957]: DEBUG nova.network.neutron [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 672.141303] env[61957]: DEBUG nova.network.neutron [None req-3d56c405-76b3-4bdb-9787-ecf239aeeb0d tempest-ImagesOneServerNegativeTestJSON-1581220397 tempest-ImagesOneServerNegativeTestJSON-1581220397-project-member] [instance: 2b151571-348f-4543-a0c5-afe6458b1973] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 672.222811] env[61957]: DEBUG nova.network.neutron [None req-3d56c405-76b3-4bdb-9787-ecf239aeeb0d tempest-ImagesOneServerNegativeTestJSON-1581220397 tempest-ImagesOneServerNegativeTestJSON-1581220397-project-member] [instance: 2b151571-348f-4543-a0c5-afe6458b1973] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 672.362119] env[61957]: DEBUG nova.scheduler.client.report [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 672.520099] env[61957]: DEBUG nova.network.neutron [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 672.532638] env[61957]: DEBUG nova.compute.manager [req-1bda9fc2-cf13-4b30-b006-c88b401cee31 req-0ee86c5a-3f1e-4e7a-97bb-84f743792f92 service nova] [instance: 2b151571-348f-4543-a0c5-afe6458b1973] Received event network-vif-deleted-892c3e98-da9b-4830-8c81-221309a32b88 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 672.725635] env[61957]: DEBUG oslo_concurrency.lockutils [None req-3d56c405-76b3-4bdb-9787-ecf239aeeb0d tempest-ImagesOneServerNegativeTestJSON-1581220397 tempest-ImagesOneServerNegativeTestJSON-1581220397-project-member] Releasing lock "refresh_cache-2b151571-348f-4543-a0c5-afe6458b1973" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 672.726148] env[61957]: DEBUG nova.compute.manager [None req-3d56c405-76b3-4bdb-9787-ecf239aeeb0d tempest-ImagesOneServerNegativeTestJSON-1581220397 tempest-ImagesOneServerNegativeTestJSON-1581220397-project-member] [instance: 2b151571-348f-4543-a0c5-afe6458b1973] Start destroying the instance on the hypervisor. {{(pid=61957) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 672.726350] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-3d56c405-76b3-4bdb-9787-ecf239aeeb0d tempest-ImagesOneServerNegativeTestJSON-1581220397 tempest-ImagesOneServerNegativeTestJSON-1581220397-project-member] [instance: 2b151571-348f-4543-a0c5-afe6458b1973] Destroying instance {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 672.726647] env[61957]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ea00aade-725f-4b3b-9165-37f033f1b540 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.735620] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02fd0588-7372-4991-9827-2df4299e9962 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.757806] env[61957]: WARNING nova.virt.vmwareapi.vmops [None req-3d56c405-76b3-4bdb-9787-ecf239aeeb0d tempest-ImagesOneServerNegativeTestJSON-1581220397 tempest-ImagesOneServerNegativeTestJSON-1581220397-project-member] [instance: 2b151571-348f-4543-a0c5-afe6458b1973] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 2b151571-348f-4543-a0c5-afe6458b1973 could not be found. [ 672.758075] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-3d56c405-76b3-4bdb-9787-ecf239aeeb0d tempest-ImagesOneServerNegativeTestJSON-1581220397 tempest-ImagesOneServerNegativeTestJSON-1581220397-project-member] [instance: 2b151571-348f-4543-a0c5-afe6458b1973] Instance destroyed {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 672.758272] env[61957]: INFO nova.compute.manager [None req-3d56c405-76b3-4bdb-9787-ecf239aeeb0d tempest-ImagesOneServerNegativeTestJSON-1581220397 tempest-ImagesOneServerNegativeTestJSON-1581220397-project-member] [instance: 2b151571-348f-4543-a0c5-afe6458b1973] Took 0.03 seconds to destroy the instance on the hypervisor. [ 672.758532] env[61957]: DEBUG oslo.service.loopingcall [None req-3d56c405-76b3-4bdb-9787-ecf239aeeb0d tempest-ImagesOneServerNegativeTestJSON-1581220397 tempest-ImagesOneServerNegativeTestJSON-1581220397-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 672.758742] env[61957]: DEBUG nova.compute.manager [-] [instance: 2b151571-348f-4543-a0c5-afe6458b1973] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 672.758839] env[61957]: DEBUG nova.network.neutron [-] [instance: 2b151571-348f-4543-a0c5-afe6458b1973] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 672.774731] env[61957]: DEBUG nova.network.neutron [-] [instance: 2b151571-348f-4543-a0c5-afe6458b1973] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 672.867787] env[61957]: DEBUG oslo_concurrency.lockutils [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.979s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 672.869093] env[61957]: ERROR nova.compute.manager [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port ebb67595-2f07-4312-a3d9-11a91467c152, please check neutron logs for more information. [ 672.869093] env[61957]: ERROR nova.compute.manager [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] Traceback (most recent call last): [ 672.869093] env[61957]: ERROR nova.compute.manager [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 672.869093] env[61957]: ERROR nova.compute.manager [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] self.driver.spawn(context, instance, image_meta, [ 672.869093] env[61957]: ERROR nova.compute.manager [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 672.869093] env[61957]: ERROR nova.compute.manager [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] self._vmops.spawn(context, instance, image_meta, injected_files, [ 672.869093] env[61957]: ERROR nova.compute.manager [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 672.869093] env[61957]: ERROR nova.compute.manager [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] vm_ref = self.build_virtual_machine(instance, [ 672.869093] env[61957]: ERROR nova.compute.manager [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 672.869093] env[61957]: ERROR nova.compute.manager [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] vif_infos = vmwarevif.get_vif_info(self._session, [ 672.869093] env[61957]: ERROR nova.compute.manager [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 672.869415] env[61957]: ERROR nova.compute.manager [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] for vif in network_info: [ 672.869415] env[61957]: ERROR nova.compute.manager [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 672.869415] env[61957]: ERROR nova.compute.manager [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] return self._sync_wrapper(fn, *args, **kwargs) [ 672.869415] env[61957]: ERROR nova.compute.manager [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 672.869415] env[61957]: ERROR nova.compute.manager [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] self.wait() [ 672.869415] env[61957]: ERROR nova.compute.manager [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 672.869415] env[61957]: ERROR nova.compute.manager [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] self[:] = self._gt.wait() [ 672.869415] env[61957]: ERROR nova.compute.manager [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 672.869415] env[61957]: ERROR nova.compute.manager [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] return self._exit_event.wait() [ 672.869415] env[61957]: ERROR nova.compute.manager [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 672.869415] env[61957]: ERROR nova.compute.manager [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] result = hub.switch() [ 672.869415] env[61957]: ERROR nova.compute.manager [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 672.869415] env[61957]: ERROR nova.compute.manager [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] return self.greenlet.switch() [ 672.869713] env[61957]: ERROR nova.compute.manager [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 672.869713] env[61957]: ERROR nova.compute.manager [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] result = function(*args, **kwargs) [ 672.869713] env[61957]: ERROR nova.compute.manager [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 672.869713] env[61957]: ERROR nova.compute.manager [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] return func(*args, **kwargs) [ 672.869713] env[61957]: ERROR nova.compute.manager [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 672.869713] env[61957]: ERROR nova.compute.manager [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] raise e [ 672.869713] env[61957]: ERROR nova.compute.manager [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 672.869713] env[61957]: ERROR nova.compute.manager [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] nwinfo = self.network_api.allocate_for_instance( [ 672.869713] env[61957]: ERROR nova.compute.manager [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 672.869713] env[61957]: ERROR nova.compute.manager [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] created_port_ids = self._update_ports_for_instance( [ 672.869713] env[61957]: ERROR nova.compute.manager [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 672.869713] env[61957]: ERROR nova.compute.manager [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] with excutils.save_and_reraise_exception(): [ 672.869713] env[61957]: ERROR nova.compute.manager [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 672.870036] env[61957]: ERROR nova.compute.manager [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] self.force_reraise() [ 672.870036] env[61957]: ERROR nova.compute.manager [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 672.870036] env[61957]: ERROR nova.compute.manager [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] raise self.value [ 672.870036] env[61957]: ERROR nova.compute.manager [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 672.870036] env[61957]: ERROR nova.compute.manager [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] updated_port = self._update_port( [ 672.870036] env[61957]: ERROR nova.compute.manager [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 672.870036] env[61957]: ERROR nova.compute.manager [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] _ensure_no_port_binding_failure(port) [ 672.870036] env[61957]: ERROR nova.compute.manager [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 672.870036] env[61957]: ERROR nova.compute.manager [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] raise exception.PortBindingFailed(port_id=port['id']) [ 672.870036] env[61957]: ERROR nova.compute.manager [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] nova.exception.PortBindingFailed: Binding failed for port ebb67595-2f07-4312-a3d9-11a91467c152, please check neutron logs for more information. [ 672.870036] env[61957]: ERROR nova.compute.manager [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] [ 672.870321] env[61957]: DEBUG nova.compute.utils [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] Binding failed for port ebb67595-2f07-4312-a3d9-11a91467c152, please check neutron logs for more information. {{(pid=61957) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 672.871226] env[61957]: DEBUG oslo_concurrency.lockutils [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.204s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 672.874429] env[61957]: DEBUG nova.compute.manager [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] Build of instance b541c437-93b7-4e4d-9ff8-bfc9cdf0e237 was re-scheduled: Binding failed for port ebb67595-2f07-4312-a3d9-11a91467c152, please check neutron logs for more information. {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 672.874995] env[61957]: DEBUG nova.compute.manager [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] Unplugging VIFs for instance {{(pid=61957) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 672.875283] env[61957]: DEBUG oslo_concurrency.lockutils [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] Acquiring lock "refresh_cache-b541c437-93b7-4e4d-9ff8-bfc9cdf0e237" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 672.875473] env[61957]: DEBUG oslo_concurrency.lockutils [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] Acquired lock "refresh_cache-b541c437-93b7-4e4d-9ff8-bfc9cdf0e237" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 672.875664] env[61957]: DEBUG nova.network.neutron [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 673.022696] env[61957]: INFO nova.compute.manager [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] [instance: 74d3de77-1272-449d-8b64-75e21fff8d7f] Took 1.03 seconds to deallocate network for instance. [ 673.159479] env[61957]: DEBUG oslo_concurrency.lockutils [None req-bc6cc00e-6b7d-4787-8382-d5cf71a4537c tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] Acquiring lock "615dbbed-2b02-4351-9e03-8c13f424a133" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 673.159777] env[61957]: DEBUG oslo_concurrency.lockutils [None req-bc6cc00e-6b7d-4787-8382-d5cf71a4537c tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] Lock "615dbbed-2b02-4351-9e03-8c13f424a133" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 673.277594] env[61957]: DEBUG nova.network.neutron [-] [instance: 2b151571-348f-4543-a0c5-afe6458b1973] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 673.396304] env[61957]: DEBUG nova.network.neutron [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 673.484369] env[61957]: DEBUG nova.network.neutron [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 673.715423] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef91b5d6-4f41-405f-ba91-3dbecc9330c7 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.723208] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-533ce7a3-a7b0-42d8-b767-72a7ab668d5e {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.753636] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f4205c6-fd85-4ef0-b86e-05cdbeb47a65 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.762019] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-047db92a-c5f5-44a7-9350-8aae28357271 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.778316] env[61957]: DEBUG nova.compute.provider_tree [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 673.781518] env[61957]: INFO nova.compute.manager [-] [instance: 2b151571-348f-4543-a0c5-afe6458b1973] Took 1.02 seconds to deallocate network for instance. [ 673.783764] env[61957]: DEBUG nova.compute.claims [None req-3d56c405-76b3-4bdb-9787-ecf239aeeb0d tempest-ImagesOneServerNegativeTestJSON-1581220397 tempest-ImagesOneServerNegativeTestJSON-1581220397-project-member] [instance: 2b151571-348f-4543-a0c5-afe6458b1973] Aborting claim: {{(pid=61957) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 673.784070] env[61957]: DEBUG oslo_concurrency.lockutils [None req-3d56c405-76b3-4bdb-9787-ecf239aeeb0d tempest-ImagesOneServerNegativeTestJSON-1581220397 tempest-ImagesOneServerNegativeTestJSON-1581220397-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 673.988612] env[61957]: DEBUG oslo_concurrency.lockutils [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] Releasing lock "refresh_cache-b541c437-93b7-4e4d-9ff8-bfc9cdf0e237" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 673.988915] env[61957]: DEBUG nova.compute.manager [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61957) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 673.989081] env[61957]: DEBUG nova.compute.manager [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 673.989287] env[61957]: DEBUG nova.network.neutron [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 674.004582] env[61957]: DEBUG nova.network.neutron [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 674.055812] env[61957]: INFO nova.scheduler.client.report [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] Deleted allocations for instance 74d3de77-1272-449d-8b64-75e21fff8d7f [ 674.282025] env[61957]: DEBUG nova.scheduler.client.report [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 674.508027] env[61957]: DEBUG nova.network.neutron [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 674.564128] env[61957]: DEBUG oslo_concurrency.lockutils [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] Lock "74d3de77-1272-449d-8b64-75e21fff8d7f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 80.742s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 674.787145] env[61957]: DEBUG oslo_concurrency.lockutils [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.916s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 674.787861] env[61957]: ERROR nova.compute.manager [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] [instance: 394b1955-c448-42c2-a718-28df7bd366e9] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 10ac5929-6235-4c76-b38b-45cc8b90ca43, please check neutron logs for more information. [ 674.787861] env[61957]: ERROR nova.compute.manager [instance: 394b1955-c448-42c2-a718-28df7bd366e9] Traceback (most recent call last): [ 674.787861] env[61957]: ERROR nova.compute.manager [instance: 394b1955-c448-42c2-a718-28df7bd366e9] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 674.787861] env[61957]: ERROR nova.compute.manager [instance: 394b1955-c448-42c2-a718-28df7bd366e9] self.driver.spawn(context, instance, image_meta, [ 674.787861] env[61957]: ERROR nova.compute.manager [instance: 394b1955-c448-42c2-a718-28df7bd366e9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 674.787861] env[61957]: ERROR nova.compute.manager [instance: 394b1955-c448-42c2-a718-28df7bd366e9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 674.787861] env[61957]: ERROR nova.compute.manager [instance: 394b1955-c448-42c2-a718-28df7bd366e9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 674.787861] env[61957]: ERROR nova.compute.manager [instance: 394b1955-c448-42c2-a718-28df7bd366e9] vm_ref = self.build_virtual_machine(instance, [ 674.787861] env[61957]: ERROR nova.compute.manager [instance: 394b1955-c448-42c2-a718-28df7bd366e9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 674.787861] env[61957]: ERROR nova.compute.manager [instance: 394b1955-c448-42c2-a718-28df7bd366e9] vif_infos = vmwarevif.get_vif_info(self._session, [ 674.787861] env[61957]: ERROR nova.compute.manager [instance: 394b1955-c448-42c2-a718-28df7bd366e9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 674.788514] env[61957]: ERROR nova.compute.manager [instance: 394b1955-c448-42c2-a718-28df7bd366e9] for vif in network_info: [ 674.788514] env[61957]: ERROR nova.compute.manager [instance: 394b1955-c448-42c2-a718-28df7bd366e9] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 674.788514] env[61957]: ERROR nova.compute.manager [instance: 394b1955-c448-42c2-a718-28df7bd366e9] return self._sync_wrapper(fn, *args, **kwargs) [ 674.788514] env[61957]: ERROR nova.compute.manager [instance: 394b1955-c448-42c2-a718-28df7bd366e9] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 674.788514] env[61957]: ERROR nova.compute.manager [instance: 394b1955-c448-42c2-a718-28df7bd366e9] self.wait() [ 674.788514] env[61957]: ERROR nova.compute.manager [instance: 394b1955-c448-42c2-a718-28df7bd366e9] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 674.788514] env[61957]: ERROR nova.compute.manager [instance: 394b1955-c448-42c2-a718-28df7bd366e9] self[:] = self._gt.wait() [ 674.788514] env[61957]: ERROR nova.compute.manager [instance: 394b1955-c448-42c2-a718-28df7bd366e9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 674.788514] env[61957]: ERROR nova.compute.manager [instance: 394b1955-c448-42c2-a718-28df7bd366e9] return self._exit_event.wait() [ 674.788514] env[61957]: ERROR nova.compute.manager [instance: 394b1955-c448-42c2-a718-28df7bd366e9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 674.788514] env[61957]: ERROR nova.compute.manager [instance: 394b1955-c448-42c2-a718-28df7bd366e9] result = hub.switch() [ 674.788514] env[61957]: ERROR nova.compute.manager [instance: 394b1955-c448-42c2-a718-28df7bd366e9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 674.788514] env[61957]: ERROR nova.compute.manager [instance: 394b1955-c448-42c2-a718-28df7bd366e9] return self.greenlet.switch() [ 674.789102] env[61957]: ERROR nova.compute.manager [instance: 394b1955-c448-42c2-a718-28df7bd366e9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 674.789102] env[61957]: ERROR nova.compute.manager [instance: 394b1955-c448-42c2-a718-28df7bd366e9] result = function(*args, **kwargs) [ 674.789102] env[61957]: ERROR nova.compute.manager [instance: 394b1955-c448-42c2-a718-28df7bd366e9] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 674.789102] env[61957]: ERROR nova.compute.manager [instance: 394b1955-c448-42c2-a718-28df7bd366e9] return func(*args, **kwargs) [ 674.789102] env[61957]: ERROR nova.compute.manager [instance: 394b1955-c448-42c2-a718-28df7bd366e9] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 674.789102] env[61957]: ERROR nova.compute.manager [instance: 394b1955-c448-42c2-a718-28df7bd366e9] raise e [ 674.789102] env[61957]: ERROR nova.compute.manager [instance: 394b1955-c448-42c2-a718-28df7bd366e9] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 674.789102] env[61957]: ERROR nova.compute.manager [instance: 394b1955-c448-42c2-a718-28df7bd366e9] nwinfo = self.network_api.allocate_for_instance( [ 674.789102] env[61957]: ERROR nova.compute.manager [instance: 394b1955-c448-42c2-a718-28df7bd366e9] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 674.789102] env[61957]: ERROR nova.compute.manager [instance: 394b1955-c448-42c2-a718-28df7bd366e9] created_port_ids = self._update_ports_for_instance( [ 674.789102] env[61957]: ERROR nova.compute.manager [instance: 394b1955-c448-42c2-a718-28df7bd366e9] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 674.789102] env[61957]: ERROR nova.compute.manager [instance: 394b1955-c448-42c2-a718-28df7bd366e9] with excutils.save_and_reraise_exception(): [ 674.789102] env[61957]: ERROR nova.compute.manager [instance: 394b1955-c448-42c2-a718-28df7bd366e9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 674.789627] env[61957]: ERROR nova.compute.manager [instance: 394b1955-c448-42c2-a718-28df7bd366e9] self.force_reraise() [ 674.789627] env[61957]: ERROR nova.compute.manager [instance: 394b1955-c448-42c2-a718-28df7bd366e9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 674.789627] env[61957]: ERROR nova.compute.manager [instance: 394b1955-c448-42c2-a718-28df7bd366e9] raise self.value [ 674.789627] env[61957]: ERROR nova.compute.manager [instance: 394b1955-c448-42c2-a718-28df7bd366e9] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 674.789627] env[61957]: ERROR nova.compute.manager [instance: 394b1955-c448-42c2-a718-28df7bd366e9] updated_port = self._update_port( [ 674.789627] env[61957]: ERROR nova.compute.manager [instance: 394b1955-c448-42c2-a718-28df7bd366e9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 674.789627] env[61957]: ERROR nova.compute.manager [instance: 394b1955-c448-42c2-a718-28df7bd366e9] _ensure_no_port_binding_failure(port) [ 674.789627] env[61957]: ERROR nova.compute.manager [instance: 394b1955-c448-42c2-a718-28df7bd366e9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 674.789627] env[61957]: ERROR nova.compute.manager [instance: 394b1955-c448-42c2-a718-28df7bd366e9] raise exception.PortBindingFailed(port_id=port['id']) [ 674.789627] env[61957]: ERROR nova.compute.manager [instance: 394b1955-c448-42c2-a718-28df7bd366e9] nova.exception.PortBindingFailed: Binding failed for port 10ac5929-6235-4c76-b38b-45cc8b90ca43, please check neutron logs for more information. [ 674.789627] env[61957]: ERROR nova.compute.manager [instance: 394b1955-c448-42c2-a718-28df7bd366e9] [ 674.790090] env[61957]: DEBUG nova.compute.utils [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] [instance: 394b1955-c448-42c2-a718-28df7bd366e9] Binding failed for port 10ac5929-6235-4c76-b38b-45cc8b90ca43, please check neutron logs for more information. {{(pid=61957) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 674.790342] env[61957]: DEBUG nova.compute.manager [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] [instance: 394b1955-c448-42c2-a718-28df7bd366e9] Build of instance 394b1955-c448-42c2-a718-28df7bd366e9 was re-scheduled: Binding failed for port 10ac5929-6235-4c76-b38b-45cc8b90ca43, please check neutron logs for more information. {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 674.790757] env[61957]: DEBUG nova.compute.manager [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] [instance: 394b1955-c448-42c2-a718-28df7bd366e9] Unplugging VIFs for instance {{(pid=61957) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 674.790984] env[61957]: DEBUG oslo_concurrency.lockutils [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] Acquiring lock "refresh_cache-394b1955-c448-42c2-a718-28df7bd366e9" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 674.791147] env[61957]: DEBUG oslo_concurrency.lockutils [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] Acquired lock "refresh_cache-394b1955-c448-42c2-a718-28df7bd366e9" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 674.791302] env[61957]: DEBUG nova.network.neutron [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] [instance: 394b1955-c448-42c2-a718-28df7bd366e9] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 674.792335] env[61957]: DEBUG oslo_concurrency.lockutils [None req-af238887-1f40-4ff3-bfc6-fdb11b6cd8bf tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.872s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 674.794399] env[61957]: INFO nova.compute.claims [None req-af238887-1f40-4ff3-bfc6-fdb11b6cd8bf tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] [instance: 188b0fd3-5d71-4feb-aca5-75a2bd28895a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 675.011925] env[61957]: INFO nova.compute.manager [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] [instance: b541c437-93b7-4e4d-9ff8-bfc9cdf0e237] Took 1.02 seconds to deallocate network for instance. [ 675.067298] env[61957]: DEBUG nova.compute.manager [None req-691c932a-6913-4bb6-98f2-4142c7d131a8 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] Starting instance... {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 675.316137] env[61957]: DEBUG nova.network.neutron [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] [instance: 394b1955-c448-42c2-a718-28df7bd366e9] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 675.397228] env[61957]: DEBUG nova.network.neutron [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] [instance: 394b1955-c448-42c2-a718-28df7bd366e9] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 675.590830] env[61957]: DEBUG oslo_concurrency.lockutils [None req-691c932a-6913-4bb6-98f2-4142c7d131a8 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 675.900340] env[61957]: DEBUG oslo_concurrency.lockutils [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] Releasing lock "refresh_cache-394b1955-c448-42c2-a718-28df7bd366e9" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 675.900602] env[61957]: DEBUG nova.compute.manager [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61957) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 675.900789] env[61957]: DEBUG nova.compute.manager [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] [instance: 394b1955-c448-42c2-a718-28df7bd366e9] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 675.900958] env[61957]: DEBUG nova.network.neutron [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] [instance: 394b1955-c448-42c2-a718-28df7bd366e9] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 675.915649] env[61957]: DEBUG nova.network.neutron [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] [instance: 394b1955-c448-42c2-a718-28df7bd366e9] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 676.039583] env[61957]: INFO nova.scheduler.client.report [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] Deleted allocations for instance b541c437-93b7-4e4d-9ff8-bfc9cdf0e237 [ 676.147200] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ed24a32-e71b-4390-a3fd-df73f5b744d4 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.155513] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3ca495f-4797-42f4-b960-ddf53dbc2de4 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.187411] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f775cff-30f9-4e46-a466-edcd4441f312 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.195013] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7466656b-8f47-4e93-8ec2-32c367b9d81d {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.207661] env[61957]: DEBUG nova.compute.provider_tree [None req-af238887-1f40-4ff3-bfc6-fdb11b6cd8bf tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 676.419067] env[61957]: DEBUG nova.network.neutron [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] [instance: 394b1955-c448-42c2-a718-28df7bd366e9] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 676.548298] env[61957]: DEBUG oslo_concurrency.lockutils [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] Lock "b541c437-93b7-4e4d-9ff8-bfc9cdf0e237" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 82.674s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 676.711312] env[61957]: DEBUG nova.scheduler.client.report [None req-af238887-1f40-4ff3-bfc6-fdb11b6cd8bf tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 676.923211] env[61957]: INFO nova.compute.manager [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] [instance: 394b1955-c448-42c2-a718-28df7bd366e9] Took 1.02 seconds to deallocate network for instance. [ 677.050622] env[61957]: DEBUG nova.compute.manager [None req-456a04ab-263c-4c55-898a-70a189651712 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] Starting instance... {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 677.216305] env[61957]: DEBUG oslo_concurrency.lockutils [None req-af238887-1f40-4ff3-bfc6-fdb11b6cd8bf tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.424s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 677.216779] env[61957]: DEBUG nova.compute.manager [None req-af238887-1f40-4ff3-bfc6-fdb11b6cd8bf tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] [instance: 188b0fd3-5d71-4feb-aca5-75a2bd28895a] Start building networks asynchronously for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 677.219328] env[61957]: DEBUG oslo_concurrency.lockutils [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 16.259s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 677.569110] env[61957]: DEBUG oslo_concurrency.lockutils [None req-456a04ab-263c-4c55-898a-70a189651712 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 677.722801] env[61957]: DEBUG nova.compute.utils [None req-af238887-1f40-4ff3-bfc6-fdb11b6cd8bf tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Using /dev/sd instead of None {{(pid=61957) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 677.726449] env[61957]: DEBUG nova.compute.manager [None req-af238887-1f40-4ff3-bfc6-fdb11b6cd8bf tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] [instance: 188b0fd3-5d71-4feb-aca5-75a2bd28895a] Not allocating networking since 'none' was specified. {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 677.950143] env[61957]: INFO nova.scheduler.client.report [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] Deleted allocations for instance 394b1955-c448-42c2-a718-28df7bd366e9 [ 678.227826] env[61957]: DEBUG nova.compute.manager [None req-af238887-1f40-4ff3-bfc6-fdb11b6cd8bf tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] [instance: 188b0fd3-5d71-4feb-aca5-75a2bd28895a] Start building block device mappings for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 678.249529] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Instance 49e99297-ffb9-4104-bd06-911243908828 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61957) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 678.249697] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Instance 9df88f23-0219-43e0-b28a-e78f30a473a7 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61957) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 678.249821] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Instance 2b151571-348f-4543-a0c5-afe6458b1973 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61957) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 678.249939] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Instance 188b0fd3-5d71-4feb-aca5-75a2bd28895a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61957) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 678.458261] env[61957]: DEBUG oslo_concurrency.lockutils [None req-24c33218-7a5b-4137-b42e-28ce934d8b64 tempest-ListServersNegativeTestJSON-1224372416 tempest-ListServersNegativeTestJSON-1224372416-project-member] Lock "394b1955-c448-42c2-a718-28df7bd366e9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 84.521s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 678.752602] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Instance 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61957) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 678.960713] env[61957]: DEBUG nova.compute.manager [None req-ca1e8d1e-8c18-4652-924a-ab1cdb751a05 tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] [instance: 825130d4-b1f6-434e-be3e-dca952abe930] Starting instance... {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 679.242087] env[61957]: DEBUG nova.compute.manager [None req-af238887-1f40-4ff3-bfc6-fdb11b6cd8bf tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] [instance: 188b0fd3-5d71-4feb-aca5-75a2bd28895a] Start spawning the instance on the hypervisor. {{(pid=61957) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 679.260551] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Instance 8257a521-2c25-45a2-a2e5-c735ece03da2 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61957) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 679.271256] env[61957]: DEBUG nova.virt.hardware [None req-af238887-1f40-4ff3-bfc6-fdb11b6cd8bf tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T17:22:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T17:21:50Z,direct_url=,disk_format='vmdk',id=11c76a2c-f705-470a-ba9d-4657858bab38,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='81d3096d0f094373913fc3dc8f5c3c6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T17:21:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 679.271256] env[61957]: DEBUG nova.virt.hardware [None req-af238887-1f40-4ff3-bfc6-fdb11b6cd8bf tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Flavor limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 679.271256] env[61957]: DEBUG nova.virt.hardware [None req-af238887-1f40-4ff3-bfc6-fdb11b6cd8bf tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Image limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 679.271256] env[61957]: DEBUG nova.virt.hardware [None req-af238887-1f40-4ff3-bfc6-fdb11b6cd8bf tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Flavor pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 679.271479] env[61957]: DEBUG nova.virt.hardware [None req-af238887-1f40-4ff3-bfc6-fdb11b6cd8bf tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Image pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 679.271511] env[61957]: DEBUG nova.virt.hardware [None req-af238887-1f40-4ff3-bfc6-fdb11b6cd8bf tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 679.271698] env[61957]: DEBUG nova.virt.hardware [None req-af238887-1f40-4ff3-bfc6-fdb11b6cd8bf tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 679.271853] env[61957]: DEBUG nova.virt.hardware [None req-af238887-1f40-4ff3-bfc6-fdb11b6cd8bf tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 679.272043] env[61957]: DEBUG nova.virt.hardware [None req-af238887-1f40-4ff3-bfc6-fdb11b6cd8bf tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Got 1 possible topologies {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 679.272234] env[61957]: DEBUG nova.virt.hardware [None req-af238887-1f40-4ff3-bfc6-fdb11b6cd8bf tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 679.272398] env[61957]: DEBUG nova.virt.hardware [None req-af238887-1f40-4ff3-bfc6-fdb11b6cd8bf tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 679.273269] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bb35730-c7ee-40a1-b40d-31ddc262df88 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.287209] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64ccf51f-80c7-4c3e-b5e1-396e4a930325 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.302871] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-af238887-1f40-4ff3-bfc6-fdb11b6cd8bf tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] [instance: 188b0fd3-5d71-4feb-aca5-75a2bd28895a] Instance VIF info [] {{(pid=61957) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 679.308444] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-af238887-1f40-4ff3-bfc6-fdb11b6cd8bf tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Creating folder: Project (019df735e3484811bb9a9642a8329709). Parent ref: group-v274445. {{(pid=61957) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 679.308713] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b013fb5b-58ef-4321-b3e0-135a6706b5f9 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.320081] env[61957]: INFO nova.virt.vmwareapi.vm_util [None req-af238887-1f40-4ff3-bfc6-fdb11b6cd8bf tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Created folder: Project (019df735e3484811bb9a9642a8329709) in parent group-v274445. [ 679.320275] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-af238887-1f40-4ff3-bfc6-fdb11b6cd8bf tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Creating folder: Instances. Parent ref: group-v274460. {{(pid=61957) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 679.320495] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f070fdab-04c9-47f3-b444-ca3df178fe47 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.328942] env[61957]: INFO nova.virt.vmwareapi.vm_util [None req-af238887-1f40-4ff3-bfc6-fdb11b6cd8bf tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Created folder: Instances in parent group-v274460. [ 679.329205] env[61957]: DEBUG oslo.service.loopingcall [None req-af238887-1f40-4ff3-bfc6-fdb11b6cd8bf tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 679.329391] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 188b0fd3-5d71-4feb-aca5-75a2bd28895a] Creating VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 679.329652] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6d265ed6-6360-437a-a644-e4671b4579d1 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.345819] env[61957]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 679.345819] env[61957]: value = "task-1277340" [ 679.345819] env[61957]: _type = "Task" [ 679.345819] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 679.353489] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277340, 'name': CreateVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 679.485815] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ca1e8d1e-8c18-4652-924a-ab1cdb751a05 tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 679.765280] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Instance d0bde8ec-da55-42b5-8d2c-9df2b90b88a7 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61957) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 679.856501] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277340, 'name': CreateVM_Task, 'duration_secs': 0.24556} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 679.856822] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 188b0fd3-5d71-4feb-aca5-75a2bd28895a] Created VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 679.857959] env[61957]: DEBUG oslo_concurrency.lockutils [None req-af238887-1f40-4ff3-bfc6-fdb11b6cd8bf tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 679.857959] env[61957]: DEBUG oslo_concurrency.lockutils [None req-af238887-1f40-4ff3-bfc6-fdb11b6cd8bf tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 679.857959] env[61957]: DEBUG oslo_concurrency.lockutils [None req-af238887-1f40-4ff3-bfc6-fdb11b6cd8bf tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 679.858187] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4afc6a4c-02b6-451b-a65d-647380088b9a {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.862751] env[61957]: DEBUG oslo_vmware.api [None req-af238887-1f40-4ff3-bfc6-fdb11b6cd8bf tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Waiting for the task: (returnval){ [ 679.862751] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]527f3e95-24be-adcb-7ae7-a3b98e9301d9" [ 679.862751] env[61957]: _type = "Task" [ 679.862751] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 679.870430] env[61957]: DEBUG oslo_vmware.api [None req-af238887-1f40-4ff3-bfc6-fdb11b6cd8bf tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]527f3e95-24be-adcb-7ae7-a3b98e9301d9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 680.270539] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Instance c5024e22-6499-4f2e-8d1b-99ca34b33104 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61957) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 680.373092] env[61957]: DEBUG oslo_vmware.api [None req-af238887-1f40-4ff3-bfc6-fdb11b6cd8bf tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]527f3e95-24be-adcb-7ae7-a3b98e9301d9, 'name': SearchDatastore_Task, 'duration_secs': 0.008879} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 680.374025] env[61957]: DEBUG oslo_concurrency.lockutils [None req-af238887-1f40-4ff3-bfc6-fdb11b6cd8bf tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 680.374025] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-af238887-1f40-4ff3-bfc6-fdb11b6cd8bf tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] [instance: 188b0fd3-5d71-4feb-aca5-75a2bd28895a] Processing image 11c76a2c-f705-470a-ba9d-4657858bab38 {{(pid=61957) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 680.374025] env[61957]: DEBUG oslo_concurrency.lockutils [None req-af238887-1f40-4ff3-bfc6-fdb11b6cd8bf tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 680.374025] env[61957]: DEBUG oslo_concurrency.lockutils [None req-af238887-1f40-4ff3-bfc6-fdb11b6cd8bf tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 680.374198] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-af238887-1f40-4ff3-bfc6-fdb11b6cd8bf tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 680.374388] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e5611971-703f-40d6-90f0-32998ad0ba02 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.382087] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-af238887-1f40-4ff3-bfc6-fdb11b6cd8bf tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 680.382268] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-af238887-1f40-4ff3-bfc6-fdb11b6cd8bf tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61957) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 680.383045] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b7a25f6d-a4df-4b75-b2fc-d32231c885b1 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.388079] env[61957]: DEBUG oslo_vmware.api [None req-af238887-1f40-4ff3-bfc6-fdb11b6cd8bf tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Waiting for the task: (returnval){ [ 680.388079] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52a13cae-a887-d393-5f65-d2845b5a319e" [ 680.388079] env[61957]: _type = "Task" [ 680.388079] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 680.395726] env[61957]: DEBUG oslo_vmware.api [None req-af238887-1f40-4ff3-bfc6-fdb11b6cd8bf tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52a13cae-a887-d393-5f65-d2845b5a319e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 680.772118] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Instance 6241fb44-ce1c-4881-bc32-7d19ddc83a61 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61957) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 680.899791] env[61957]: DEBUG oslo_vmware.api [None req-af238887-1f40-4ff3-bfc6-fdb11b6cd8bf tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52a13cae-a887-d393-5f65-d2845b5a319e, 'name': SearchDatastore_Task, 'duration_secs': 0.009518} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 680.900663] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f497e86c-ab16-4d8c-b428-b3e5025b7198 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.906297] env[61957]: DEBUG oslo_vmware.api [None req-af238887-1f40-4ff3-bfc6-fdb11b6cd8bf tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Waiting for the task: (returnval){ [ 680.906297] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52b17647-017a-f507-b59e-223da8f8fd3c" [ 680.906297] env[61957]: _type = "Task" [ 680.906297] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 680.914358] env[61957]: DEBUG oslo_vmware.api [None req-af238887-1f40-4ff3-bfc6-fdb11b6cd8bf tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52b17647-017a-f507-b59e-223da8f8fd3c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 681.275329] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Instance 825130d4-b1f6-434e-be3e-dca952abe930 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61957) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 681.417325] env[61957]: DEBUG oslo_vmware.api [None req-af238887-1f40-4ff3-bfc6-fdb11b6cd8bf tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52b17647-017a-f507-b59e-223da8f8fd3c, 'name': SearchDatastore_Task, 'duration_secs': 0.008737} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 681.417597] env[61957]: DEBUG oslo_concurrency.lockutils [None req-af238887-1f40-4ff3-bfc6-fdb11b6cd8bf tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 681.417846] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-af238887-1f40-4ff3-bfc6-fdb11b6cd8bf tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore2] 188b0fd3-5d71-4feb-aca5-75a2bd28895a/188b0fd3-5d71-4feb-aca5-75a2bd28895a.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 681.418134] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-fc44a247-95ef-4c9b-8e85-e6199ab4f47e {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.424281] env[61957]: DEBUG oslo_vmware.api [None req-af238887-1f40-4ff3-bfc6-fdb11b6cd8bf tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Waiting for the task: (returnval){ [ 681.424281] env[61957]: value = "task-1277341" [ 681.424281] env[61957]: _type = "Task" [ 681.424281] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 681.431352] env[61957]: DEBUG oslo_vmware.api [None req-af238887-1f40-4ff3-bfc6-fdb11b6cd8bf tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Task: {'id': task-1277341, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 681.779129] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Instance c369d148-2093-4724-82cc-d80a67131fea has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61957) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 681.934942] env[61957]: DEBUG oslo_vmware.api [None req-af238887-1f40-4ff3-bfc6-fdb11b6cd8bf tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Task: {'id': task-1277341, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.439504} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 681.935243] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-af238887-1f40-4ff3-bfc6-fdb11b6cd8bf tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore2] 188b0fd3-5d71-4feb-aca5-75a2bd28895a/188b0fd3-5d71-4feb-aca5-75a2bd28895a.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 681.935459] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-af238887-1f40-4ff3-bfc6-fdb11b6cd8bf tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] [instance: 188b0fd3-5d71-4feb-aca5-75a2bd28895a] Extending root virtual disk to 1048576 {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 681.935699] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-bae1f6fe-ba08-4371-94ac-dff3951aeaf6 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.942200] env[61957]: DEBUG oslo_vmware.api [None req-af238887-1f40-4ff3-bfc6-fdb11b6cd8bf tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Waiting for the task: (returnval){ [ 681.942200] env[61957]: value = "task-1277342" [ 681.942200] env[61957]: _type = "Task" [ 681.942200] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 681.950641] env[61957]: DEBUG oslo_vmware.api [None req-af238887-1f40-4ff3-bfc6-fdb11b6cd8bf tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Task: {'id': task-1277342, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 682.283255] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Instance 2252ef6e-19ec-4356-8221-33168c01bef0 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61957) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 682.452568] env[61957]: DEBUG oslo_vmware.api [None req-af238887-1f40-4ff3-bfc6-fdb11b6cd8bf tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Task: {'id': task-1277342, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.054522} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 682.452896] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-af238887-1f40-4ff3-bfc6-fdb11b6cd8bf tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] [instance: 188b0fd3-5d71-4feb-aca5-75a2bd28895a] Extended root virtual disk {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 682.453700] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-325e36d8-d7e4-4d7c-9d43-2e185ad9c9b6 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.473154] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-af238887-1f40-4ff3-bfc6-fdb11b6cd8bf tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] [instance: 188b0fd3-5d71-4feb-aca5-75a2bd28895a] Reconfiguring VM instance instance-0000001b to attach disk [datastore2] 188b0fd3-5d71-4feb-aca5-75a2bd28895a/188b0fd3-5d71-4feb-aca5-75a2bd28895a.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 682.473365] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fe8c5277-1bf3-4f28-97d7-62b35c3bc0cf {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.492424] env[61957]: DEBUG oslo_vmware.api [None req-af238887-1f40-4ff3-bfc6-fdb11b6cd8bf tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Waiting for the task: (returnval){ [ 682.492424] env[61957]: value = "task-1277343" [ 682.492424] env[61957]: _type = "Task" [ 682.492424] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 682.500076] env[61957]: DEBUG oslo_vmware.api [None req-af238887-1f40-4ff3-bfc6-fdb11b6cd8bf tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Task: {'id': task-1277343, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 682.785951] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Instance f8235178-7ae0-4733-bb05-3515d5489a5b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61957) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 683.002494] env[61957]: DEBUG oslo_vmware.api [None req-af238887-1f40-4ff3-bfc6-fdb11b6cd8bf tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Task: {'id': task-1277343, 'name': ReconfigVM_Task, 'duration_secs': 0.261932} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 683.002795] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-af238887-1f40-4ff3-bfc6-fdb11b6cd8bf tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] [instance: 188b0fd3-5d71-4feb-aca5-75a2bd28895a] Reconfigured VM instance instance-0000001b to attach disk [datastore2] 188b0fd3-5d71-4feb-aca5-75a2bd28895a/188b0fd3-5d71-4feb-aca5-75a2bd28895a.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 683.003603] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b362d713-2abd-46e3-b824-2148913e85b2 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.009608] env[61957]: DEBUG oslo_vmware.api [None req-af238887-1f40-4ff3-bfc6-fdb11b6cd8bf tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Waiting for the task: (returnval){ [ 683.009608] env[61957]: value = "task-1277344" [ 683.009608] env[61957]: _type = "Task" [ 683.009608] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 683.017388] env[61957]: DEBUG oslo_vmware.api [None req-af238887-1f40-4ff3-bfc6-fdb11b6cd8bf tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Task: {'id': task-1277344, 'name': Rename_Task} progress is 5%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 683.289043] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Instance 115c1bb4-0e6b-4c6c-9d38-9584debea01f has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61957) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 683.519562] env[61957]: DEBUG oslo_vmware.api [None req-af238887-1f40-4ff3-bfc6-fdb11b6cd8bf tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Task: {'id': task-1277344, 'name': Rename_Task, 'duration_secs': 0.136121} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 683.519830] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-af238887-1f40-4ff3-bfc6-fdb11b6cd8bf tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] [instance: 188b0fd3-5d71-4feb-aca5-75a2bd28895a] Powering on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 683.521541] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4b88546f-11c7-4684-b596-23c48df1c046 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.525898] env[61957]: DEBUG oslo_vmware.api [None req-af238887-1f40-4ff3-bfc6-fdb11b6cd8bf tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Waiting for the task: (returnval){ [ 683.525898] env[61957]: value = "task-1277345" [ 683.525898] env[61957]: _type = "Task" [ 683.525898] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 683.535687] env[61957]: DEBUG oslo_vmware.api [None req-af238887-1f40-4ff3-bfc6-fdb11b6cd8bf tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Task: {'id': task-1277345, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 683.792045] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Instance 348ea580-2268-4082-a5f1-33c954063e4d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61957) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 684.035653] env[61957]: DEBUG oslo_vmware.api [None req-af238887-1f40-4ff3-bfc6-fdb11b6cd8bf tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Task: {'id': task-1277345, 'name': PowerOnVM_Task, 'duration_secs': 0.393473} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 684.035653] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-af238887-1f40-4ff3-bfc6-fdb11b6cd8bf tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] [instance: 188b0fd3-5d71-4feb-aca5-75a2bd28895a] Powered on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 684.035770] env[61957]: INFO nova.compute.manager [None req-af238887-1f40-4ff3-bfc6-fdb11b6cd8bf tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] [instance: 188b0fd3-5d71-4feb-aca5-75a2bd28895a] Took 4.79 seconds to spawn the instance on the hypervisor. [ 684.035926] env[61957]: DEBUG nova.compute.manager [None req-af238887-1f40-4ff3-bfc6-fdb11b6cd8bf tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] [instance: 188b0fd3-5d71-4feb-aca5-75a2bd28895a] Checking state {{(pid=61957) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 684.036868] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5b868d9-7ac5-4550-be7e-7531764ed0f7 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.295664] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Instance 7f3850c0-a455-4e5b-a329-fb3b02f9725f has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61957) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 684.555046] env[61957]: INFO nova.compute.manager [None req-af238887-1f40-4ff3-bfc6-fdb11b6cd8bf tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] [instance: 188b0fd3-5d71-4feb-aca5-75a2bd28895a] Took 25.65 seconds to build instance. [ 684.799254] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Instance 042425a0-5398-47a6-9842-380d5b3299c0 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61957) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 684.889787] env[61957]: INFO nova.compute.manager [None req-eb81388f-07fa-4a13-ac24-ee048434d451 tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] [instance: 188b0fd3-5d71-4feb-aca5-75a2bd28895a] Rebuilding instance [ 684.930139] env[61957]: DEBUG nova.compute.manager [None req-eb81388f-07fa-4a13-ac24-ee048434d451 tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] [instance: 188b0fd3-5d71-4feb-aca5-75a2bd28895a] Checking state {{(pid=61957) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 684.931030] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49d67531-5946-48db-9dbd-2742c830dd98 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.056415] env[61957]: DEBUG oslo_concurrency.lockutils [None req-af238887-1f40-4ff3-bfc6-fdb11b6cd8bf tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Lock "188b0fd3-5d71-4feb-aca5-75a2bd28895a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 78.825s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 685.303099] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Instance 60ec4271-2614-470d-bf40-47b9955f544c has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61957) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 685.444346] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-eb81388f-07fa-4a13-ac24-ee048434d451 tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] [instance: 188b0fd3-5d71-4feb-aca5-75a2bd28895a] Powering off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 685.444667] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6da5d478-930c-4554-9fe0-4293b73be73e {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.451987] env[61957]: DEBUG oslo_vmware.api [None req-eb81388f-07fa-4a13-ac24-ee048434d451 tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Waiting for the task: (returnval){ [ 685.451987] env[61957]: value = "task-1277346" [ 685.451987] env[61957]: _type = "Task" [ 685.451987] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 685.459855] env[61957]: DEBUG oslo_vmware.api [None req-eb81388f-07fa-4a13-ac24-ee048434d451 tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Task: {'id': task-1277346, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 685.558557] env[61957]: DEBUG nova.compute.manager [None req-58fe0400-f1d0-4e1d-a459-7541ba7aeeed tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] [instance: c369d148-2093-4724-82cc-d80a67131fea] Starting instance... {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 685.806105] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Instance e69a7833-97b1-4063-a6e8-c4558980ee1e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61957) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 685.962756] env[61957]: DEBUG oslo_vmware.api [None req-eb81388f-07fa-4a13-ac24-ee048434d451 tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Task: {'id': task-1277346, 'name': PowerOffVM_Task, 'duration_secs': 0.173071} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 685.963245] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-eb81388f-07fa-4a13-ac24-ee048434d451 tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] [instance: 188b0fd3-5d71-4feb-aca5-75a2bd28895a] Powered off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 685.963630] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-eb81388f-07fa-4a13-ac24-ee048434d451 tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] [instance: 188b0fd3-5d71-4feb-aca5-75a2bd28895a] Destroying instance {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 685.964408] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a581b66-22aa-4d85-bece-ac2b4c6bf4b1 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.971125] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-eb81388f-07fa-4a13-ac24-ee048434d451 tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] [instance: 188b0fd3-5d71-4feb-aca5-75a2bd28895a] Unregistering the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 685.971372] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e3989e65-61f3-4762-b147-cc8043aa7970 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.999854] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-eb81388f-07fa-4a13-ac24-ee048434d451 tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] [instance: 188b0fd3-5d71-4feb-aca5-75a2bd28895a] Unregistered the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 686.000757] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-eb81388f-07fa-4a13-ac24-ee048434d451 tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] [instance: 188b0fd3-5d71-4feb-aca5-75a2bd28895a] Deleting contents of the VM from datastore datastore2 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 686.000757] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-eb81388f-07fa-4a13-ac24-ee048434d451 tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Deleting the datastore file [datastore2] 188b0fd3-5d71-4feb-aca5-75a2bd28895a {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 686.000757] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2758e956-6b45-48b5-a605-78ade10cf530 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.006738] env[61957]: DEBUG oslo_vmware.api [None req-eb81388f-07fa-4a13-ac24-ee048434d451 tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Waiting for the task: (returnval){ [ 686.006738] env[61957]: value = "task-1277348" [ 686.006738] env[61957]: _type = "Task" [ 686.006738] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 686.013901] env[61957]: DEBUG oslo_vmware.api [None req-eb81388f-07fa-4a13-ac24-ee048434d451 tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Task: {'id': task-1277348, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 686.076965] env[61957]: DEBUG oslo_concurrency.lockutils [None req-58fe0400-f1d0-4e1d-a459-7541ba7aeeed tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 686.309995] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Instance d77276ba-6d37-440b-a113-c4f62b2e7946 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61957) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 686.517561] env[61957]: DEBUG oslo_vmware.api [None req-eb81388f-07fa-4a13-ac24-ee048434d451 tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Task: {'id': task-1277348, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.096627} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 686.517818] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-eb81388f-07fa-4a13-ac24-ee048434d451 tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Deleted the datastore file {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 686.518075] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-eb81388f-07fa-4a13-ac24-ee048434d451 tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] [instance: 188b0fd3-5d71-4feb-aca5-75a2bd28895a] Deleted contents of the VM from datastore datastore2 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 686.518279] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-eb81388f-07fa-4a13-ac24-ee048434d451 tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] [instance: 188b0fd3-5d71-4feb-aca5-75a2bd28895a] Instance destroyed {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 686.813050] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Instance 5dc993e9-5970-43ae-90be-f3a53d1f090a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61957) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 687.316447] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Instance 517eeb71-e9e3-4284-a144-dc4b63319d2d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61957) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 687.550563] env[61957]: DEBUG nova.virt.hardware [None req-eb81388f-07fa-4a13-ac24-ee048434d451 tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T17:22:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T17:21:50Z,direct_url=,disk_format='vmdk',id=11c76a2c-f705-470a-ba9d-4657858bab38,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='81d3096d0f094373913fc3dc8f5c3c6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T17:21:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 687.550819] env[61957]: DEBUG nova.virt.hardware [None req-eb81388f-07fa-4a13-ac24-ee048434d451 tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Flavor limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 687.550978] env[61957]: DEBUG nova.virt.hardware [None req-eb81388f-07fa-4a13-ac24-ee048434d451 tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Image limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 687.551175] env[61957]: DEBUG nova.virt.hardware [None req-eb81388f-07fa-4a13-ac24-ee048434d451 tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Flavor pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 687.551329] env[61957]: DEBUG nova.virt.hardware [None req-eb81388f-07fa-4a13-ac24-ee048434d451 tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Image pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 687.551478] env[61957]: DEBUG nova.virt.hardware [None req-eb81388f-07fa-4a13-ac24-ee048434d451 tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 687.551682] env[61957]: DEBUG nova.virt.hardware [None req-eb81388f-07fa-4a13-ac24-ee048434d451 tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 687.551838] env[61957]: DEBUG nova.virt.hardware [None req-eb81388f-07fa-4a13-ac24-ee048434d451 tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 687.552007] env[61957]: DEBUG nova.virt.hardware [None req-eb81388f-07fa-4a13-ac24-ee048434d451 tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Got 1 possible topologies {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 687.552179] env[61957]: DEBUG nova.virt.hardware [None req-eb81388f-07fa-4a13-ac24-ee048434d451 tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 687.552350] env[61957]: DEBUG nova.virt.hardware [None req-eb81388f-07fa-4a13-ac24-ee048434d451 tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 687.553204] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c8a0ebd-9c1d-4f95-80f3-a06dfff47368 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.561132] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da289c05-f3df-4856-943d-1efbafe937b0 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.573789] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-eb81388f-07fa-4a13-ac24-ee048434d451 tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] [instance: 188b0fd3-5d71-4feb-aca5-75a2bd28895a] Instance VIF info [] {{(pid=61957) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 687.579290] env[61957]: DEBUG oslo.service.loopingcall [None req-eb81388f-07fa-4a13-ac24-ee048434d451 tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 687.579518] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 188b0fd3-5d71-4feb-aca5-75a2bd28895a] Creating VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 687.579711] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ce9cd87a-d040-4399-88cc-844667fb1c3e {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.596310] env[61957]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 687.596310] env[61957]: value = "task-1277349" [ 687.596310] env[61957]: _type = "Task" [ 687.596310] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 687.603345] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277349, 'name': CreateVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 687.819684] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Instance 5cb09b60-70b6-4d04-850e-049612ec4a89 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61957) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 688.105645] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277349, 'name': CreateVM_Task} progress is 99%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 688.322864] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Instance 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61957) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 688.606238] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277349, 'name': CreateVM_Task} progress is 99%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 688.826116] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Instance 6372a384-61a8-43e8-8f6c-376e2a735045 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61957) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 689.106804] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277349, 'name': CreateVM_Task, 'duration_secs': 1.261381} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 689.106804] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 188b0fd3-5d71-4feb-aca5-75a2bd28895a] Created VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 689.107923] env[61957]: DEBUG oslo_concurrency.lockutils [None req-eb81388f-07fa-4a13-ac24-ee048434d451 tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 689.107923] env[61957]: DEBUG oslo_concurrency.lockutils [None req-eb81388f-07fa-4a13-ac24-ee048434d451 tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 689.107923] env[61957]: DEBUG oslo_concurrency.lockutils [None req-eb81388f-07fa-4a13-ac24-ee048434d451 tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 689.107923] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0ef29fae-7c7e-4a1b-83a7-ff521b56c306 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.112166] env[61957]: DEBUG oslo_vmware.api [None req-eb81388f-07fa-4a13-ac24-ee048434d451 tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Waiting for the task: (returnval){ [ 689.112166] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]521769d7-5e6a-1212-b576-251f5cc53c95" [ 689.112166] env[61957]: _type = "Task" [ 689.112166] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 689.120502] env[61957]: DEBUG oslo_vmware.api [None req-eb81388f-07fa-4a13-ac24-ee048434d451 tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]521769d7-5e6a-1212-b576-251f5cc53c95, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 689.328788] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Instance b0af6294-f30c-4266-ae46-6fb03dc0cbeb has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61957) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 689.622127] env[61957]: DEBUG oslo_vmware.api [None req-eb81388f-07fa-4a13-ac24-ee048434d451 tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]521769d7-5e6a-1212-b576-251f5cc53c95, 'name': SearchDatastore_Task, 'duration_secs': 0.010052} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 689.622455] env[61957]: DEBUG oslo_concurrency.lockutils [None req-eb81388f-07fa-4a13-ac24-ee048434d451 tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 689.622683] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-eb81388f-07fa-4a13-ac24-ee048434d451 tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] [instance: 188b0fd3-5d71-4feb-aca5-75a2bd28895a] Processing image 11c76a2c-f705-470a-ba9d-4657858bab38 {{(pid=61957) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 689.622908] env[61957]: DEBUG oslo_concurrency.lockutils [None req-eb81388f-07fa-4a13-ac24-ee048434d451 tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 689.623062] env[61957]: DEBUG oslo_concurrency.lockutils [None req-eb81388f-07fa-4a13-ac24-ee048434d451 tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 689.623236] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-eb81388f-07fa-4a13-ac24-ee048434d451 tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 689.623470] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4c6647a1-1f71-4646-9fa0-4c7f5e2f5562 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.630989] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-eb81388f-07fa-4a13-ac24-ee048434d451 tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 689.631174] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-eb81388f-07fa-4a13-ac24-ee048434d451 tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61957) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 689.631830] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-25b3bc14-4be8-4b0b-be53-284d41491bb9 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.636465] env[61957]: DEBUG oslo_vmware.api [None req-eb81388f-07fa-4a13-ac24-ee048434d451 tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Waiting for the task: (returnval){ [ 689.636465] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]527a2204-0699-d3e8-bbcf-43ef1f476c8b" [ 689.636465] env[61957]: _type = "Task" [ 689.636465] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 689.643296] env[61957]: DEBUG oslo_vmware.api [None req-eb81388f-07fa-4a13-ac24-ee048434d451 tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]527a2204-0699-d3e8-bbcf-43ef1f476c8b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 689.832319] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Instance 6e7bd89c-2c2a-450a-9858-3526d96c28ab has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61957) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 690.151017] env[61957]: DEBUG oslo_vmware.api [None req-eb81388f-07fa-4a13-ac24-ee048434d451 tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]527a2204-0699-d3e8-bbcf-43ef1f476c8b, 'name': SearchDatastore_Task, 'duration_secs': 0.007973} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 690.151107] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4a9a68da-813f-4f7c-a145-7cc2c736fd6c {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.156403] env[61957]: DEBUG oslo_vmware.api [None req-eb81388f-07fa-4a13-ac24-ee048434d451 tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Waiting for the task: (returnval){ [ 690.156403] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52fcb0e3-e418-97ba-2b59-0e23a7d34678" [ 690.156403] env[61957]: _type = "Task" [ 690.156403] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 690.164197] env[61957]: DEBUG oslo_vmware.api [None req-eb81388f-07fa-4a13-ac24-ee048434d451 tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52fcb0e3-e418-97ba-2b59-0e23a7d34678, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 690.335385] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Instance e0249936-d616-4ffb-8f77-d8107633c42a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61957) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 690.666890] env[61957]: DEBUG oslo_vmware.api [None req-eb81388f-07fa-4a13-ac24-ee048434d451 tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52fcb0e3-e418-97ba-2b59-0e23a7d34678, 'name': SearchDatastore_Task, 'duration_secs': 0.008408} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 690.667159] env[61957]: DEBUG oslo_concurrency.lockutils [None req-eb81388f-07fa-4a13-ac24-ee048434d451 tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 690.667407] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-eb81388f-07fa-4a13-ac24-ee048434d451 tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore2] 188b0fd3-5d71-4feb-aca5-75a2bd28895a/188b0fd3-5d71-4feb-aca5-75a2bd28895a.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 690.667644] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-285d71a2-0ee6-44af-bf33-0378dd66449b {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.673421] env[61957]: DEBUG oslo_vmware.api [None req-eb81388f-07fa-4a13-ac24-ee048434d451 tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Waiting for the task: (returnval){ [ 690.673421] env[61957]: value = "task-1277350" [ 690.673421] env[61957]: _type = "Task" [ 690.673421] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 690.680684] env[61957]: DEBUG oslo_vmware.api [None req-eb81388f-07fa-4a13-ac24-ee048434d451 tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Task: {'id': task-1277350, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 690.838905] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Instance 615dbbed-2b02-4351-9e03-8c13f424a133 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61957) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 690.839197] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Total usable vcpus: 48, total allocated vcpus: 4 {{(pid=61957) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 690.839317] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1280MB phys_disk=200GB used_disk=4GB total_vcpus=48 used_vcpus=4 pci_stats=[] {{(pid=61957) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 691.184015] env[61957]: DEBUG oslo_vmware.api [None req-eb81388f-07fa-4a13-ac24-ee048434d451 tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Task: {'id': task-1277350, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.432266} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 691.186498] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-eb81388f-07fa-4a13-ac24-ee048434d451 tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore2] 188b0fd3-5d71-4feb-aca5-75a2bd28895a/188b0fd3-5d71-4feb-aca5-75a2bd28895a.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 691.186711] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-eb81388f-07fa-4a13-ac24-ee048434d451 tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] [instance: 188b0fd3-5d71-4feb-aca5-75a2bd28895a] Extending root virtual disk to 1048576 {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 691.187146] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3e63a038-ce92-4b63-aaba-fcde9494fe4a {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.192976] env[61957]: DEBUG oslo_vmware.api [None req-eb81388f-07fa-4a13-ac24-ee048434d451 tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Waiting for the task: (returnval){ [ 691.192976] env[61957]: value = "task-1277351" [ 691.192976] env[61957]: _type = "Task" [ 691.192976] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 691.204182] env[61957]: DEBUG oslo_vmware.api [None req-eb81388f-07fa-4a13-ac24-ee048434d451 tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Task: {'id': task-1277351, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 691.238384] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3adec50-6e03-4a0b-8dc2-f5f2467f090a {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.245163] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f9bb94d-a20d-4384-b9ab-96ca3c421e4a {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.276588] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bb78910-4706-42c2-8c67-2d9d011ee7ab {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.284135] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5af8cbe2-4718-468b-ab62-bc95f1e39860 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.297394] env[61957]: DEBUG nova.compute.provider_tree [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 691.703032] env[61957]: DEBUG oslo_vmware.api [None req-eb81388f-07fa-4a13-ac24-ee048434d451 tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Task: {'id': task-1277351, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.062713} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 691.703313] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-eb81388f-07fa-4a13-ac24-ee048434d451 tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] [instance: 188b0fd3-5d71-4feb-aca5-75a2bd28895a] Extended root virtual disk {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 691.703981] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4adf028-ccf4-4be0-b03d-c7deefdd0fd6 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.723329] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-eb81388f-07fa-4a13-ac24-ee048434d451 tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] [instance: 188b0fd3-5d71-4feb-aca5-75a2bd28895a] Reconfiguring VM instance instance-0000001b to attach disk [datastore2] 188b0fd3-5d71-4feb-aca5-75a2bd28895a/188b0fd3-5d71-4feb-aca5-75a2bd28895a.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 691.723557] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-94b8fa8b-de1e-4a25-9998-546031537cef {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.742528] env[61957]: DEBUG oslo_vmware.api [None req-eb81388f-07fa-4a13-ac24-ee048434d451 tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Waiting for the task: (returnval){ [ 691.742528] env[61957]: value = "task-1277352" [ 691.742528] env[61957]: _type = "Task" [ 691.742528] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 691.749832] env[61957]: DEBUG oslo_vmware.api [None req-eb81388f-07fa-4a13-ac24-ee048434d451 tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Task: {'id': task-1277352, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 691.800999] env[61957]: DEBUG nova.scheduler.client.report [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 692.252690] env[61957]: DEBUG oslo_vmware.api [None req-eb81388f-07fa-4a13-ac24-ee048434d451 tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Task: {'id': task-1277352, 'name': ReconfigVM_Task, 'duration_secs': 0.276028} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 692.252996] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-eb81388f-07fa-4a13-ac24-ee048434d451 tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] [instance: 188b0fd3-5d71-4feb-aca5-75a2bd28895a] Reconfigured VM instance instance-0000001b to attach disk [datastore2] 188b0fd3-5d71-4feb-aca5-75a2bd28895a/188b0fd3-5d71-4feb-aca5-75a2bd28895a.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 692.253650] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-dff142cd-a12d-4e2d-a46c-8eb7b2cd5047 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.261025] env[61957]: DEBUG oslo_vmware.api [None req-eb81388f-07fa-4a13-ac24-ee048434d451 tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Waiting for the task: (returnval){ [ 692.261025] env[61957]: value = "task-1277353" [ 692.261025] env[61957]: _type = "Task" [ 692.261025] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 692.268375] env[61957]: DEBUG oslo_vmware.api [None req-eb81388f-07fa-4a13-ac24-ee048434d451 tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Task: {'id': task-1277353, 'name': Rename_Task} progress is 5%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 692.308233] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61957) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 692.308233] env[61957]: DEBUG oslo_concurrency.lockutils [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 15.087s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 692.308233] env[61957]: DEBUG oslo_concurrency.lockutils [None req-74fa7fa3-fb2c-4064-80f1-1405089fca10 tempest-ServerMetadataNegativeTestJSON-1373515003 tempest-ServerMetadataNegativeTestJSON-1373515003-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 31.314s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 692.308474] env[61957]: INFO nova.compute.claims [None req-74fa7fa3-fb2c-4064-80f1-1405089fca10 tempest-ServerMetadataNegativeTestJSON-1373515003 tempest-ServerMetadataNegativeTestJSON-1373515003-project-member] [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 692.771940] env[61957]: DEBUG oslo_vmware.api [None req-eb81388f-07fa-4a13-ac24-ee048434d451 tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Task: {'id': task-1277353, 'name': Rename_Task, 'duration_secs': 0.145396} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 692.772250] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-eb81388f-07fa-4a13-ac24-ee048434d451 tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] [instance: 188b0fd3-5d71-4feb-aca5-75a2bd28895a] Powering on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 692.772466] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7c73237f-d214-40bd-bc94-212f470c6786 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.778557] env[61957]: DEBUG oslo_vmware.api [None req-eb81388f-07fa-4a13-ac24-ee048434d451 tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Waiting for the task: (returnval){ [ 692.778557] env[61957]: value = "task-1277354" [ 692.778557] env[61957]: _type = "Task" [ 692.778557] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 692.785560] env[61957]: DEBUG oslo_vmware.api [None req-eb81388f-07fa-4a13-ac24-ee048434d451 tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Task: {'id': task-1277354, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 693.289382] env[61957]: DEBUG oslo_vmware.api [None req-eb81388f-07fa-4a13-ac24-ee048434d451 tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Task: {'id': task-1277354, 'name': PowerOnVM_Task, 'duration_secs': 0.492039} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 693.289667] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-eb81388f-07fa-4a13-ac24-ee048434d451 tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] [instance: 188b0fd3-5d71-4feb-aca5-75a2bd28895a] Powered on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 693.289869] env[61957]: DEBUG nova.compute.manager [None req-eb81388f-07fa-4a13-ac24-ee048434d451 tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] [instance: 188b0fd3-5d71-4feb-aca5-75a2bd28895a] Checking state {{(pid=61957) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 693.290666] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6dd1da6f-68ca-45ee-837f-bdc04d730210 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.641836] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cd21400-9f29-4c56-a216-51b8f2f72fb4 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.648847] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08667f9b-6392-49d1-998a-0cf35e787289 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.678010] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-678d339d-ef56-4c6b-8fef-82379c5edcda {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.686047] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-849294c0-6621-465b-98bc-c70cca6cfcbc {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.699441] env[61957]: DEBUG nova.compute.provider_tree [None req-74fa7fa3-fb2c-4064-80f1-1405089fca10 tempest-ServerMetadataNegativeTestJSON-1373515003 tempest-ServerMetadataNegativeTestJSON-1373515003-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 693.809085] env[61957]: DEBUG oslo_concurrency.lockutils [None req-eb81388f-07fa-4a13-ac24-ee048434d451 tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 694.202373] env[61957]: DEBUG nova.scheduler.client.report [None req-74fa7fa3-fb2c-4064-80f1-1405089fca10 tempest-ServerMetadataNegativeTestJSON-1373515003 tempest-ServerMetadataNegativeTestJSON-1373515003-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 694.707157] env[61957]: DEBUG oslo_concurrency.lockutils [None req-74fa7fa3-fb2c-4064-80f1-1405089fca10 tempest-ServerMetadataNegativeTestJSON-1373515003 tempest-ServerMetadataNegativeTestJSON-1373515003-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.400s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 694.707707] env[61957]: DEBUG nova.compute.manager [None req-74fa7fa3-fb2c-4064-80f1-1405089fca10 tempest-ServerMetadataNegativeTestJSON-1373515003 tempest-ServerMetadataNegativeTestJSON-1373515003-project-member] [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] Start building networks asynchronously for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 694.710276] env[61957]: DEBUG oslo_concurrency.lockutils [None req-b1f62379-e574-4175-a099-f938af5ddaa8 tempest-ServerPasswordTestJSON-1543821420 tempest-ServerPasswordTestJSON-1543821420-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 30.732s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 694.885661] env[61957]: INFO nova.compute.manager [None req-2d3c8bae-3f1e-41b1-82a9-b293bdd893bd tempest-ServersAdmin275Test-2089451214 tempest-ServersAdmin275Test-2089451214-project-admin] [instance: 188b0fd3-5d71-4feb-aca5-75a2bd28895a] Rebuilding instance [ 694.926298] env[61957]: DEBUG nova.compute.manager [None req-2d3c8bae-3f1e-41b1-82a9-b293bdd893bd tempest-ServersAdmin275Test-2089451214 tempest-ServersAdmin275Test-2089451214-project-admin] [instance: 188b0fd3-5d71-4feb-aca5-75a2bd28895a] Checking state {{(pid=61957) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 694.927151] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1e810be-c478-44f1-ace4-edf1a920ae00 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.215435] env[61957]: DEBUG nova.compute.utils [None req-74fa7fa3-fb2c-4064-80f1-1405089fca10 tempest-ServerMetadataNegativeTestJSON-1373515003 tempest-ServerMetadataNegativeTestJSON-1373515003-project-member] Using /dev/sd instead of None {{(pid=61957) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 695.219948] env[61957]: DEBUG nova.compute.manager [None req-74fa7fa3-fb2c-4064-80f1-1405089fca10 tempest-ServerMetadataNegativeTestJSON-1373515003 tempest-ServerMetadataNegativeTestJSON-1373515003-project-member] [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] Allocating IP information in the background. {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 695.220311] env[61957]: DEBUG nova.network.neutron [None req-74fa7fa3-fb2c-4064-80f1-1405089fca10 tempest-ServerMetadataNegativeTestJSON-1373515003 tempest-ServerMetadataNegativeTestJSON-1373515003-project-member] [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] allocate_for_instance() {{(pid=61957) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 695.268504] env[61957]: DEBUG nova.policy [None req-74fa7fa3-fb2c-4064-80f1-1405089fca10 tempest-ServerMetadataNegativeTestJSON-1373515003 tempest-ServerMetadataNegativeTestJSON-1373515003-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'db8333e65e7440dab8336b18826fc154', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9994b9ad2cf844aaa8c85977ebf3c275', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61957) authorize /opt/stack/nova/nova/policy.py:203}} [ 695.440666] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-2d3c8bae-3f1e-41b1-82a9-b293bdd893bd tempest-ServersAdmin275Test-2089451214 tempest-ServersAdmin275Test-2089451214-project-admin] [instance: 188b0fd3-5d71-4feb-aca5-75a2bd28895a] Powering off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 695.441855] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a061a953-69d6-4e45-b935-b85cee7f3ba5 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.449643] env[61957]: DEBUG oslo_vmware.api [None req-2d3c8bae-3f1e-41b1-82a9-b293bdd893bd tempest-ServersAdmin275Test-2089451214 tempest-ServersAdmin275Test-2089451214-project-admin] Waiting for the task: (returnval){ [ 695.449643] env[61957]: value = "task-1277355" [ 695.449643] env[61957]: _type = "Task" [ 695.449643] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 695.466535] env[61957]: DEBUG oslo_vmware.api [None req-2d3c8bae-3f1e-41b1-82a9-b293bdd893bd tempest-ServersAdmin275Test-2089451214 tempest-ServersAdmin275Test-2089451214-project-admin] Task: {'id': task-1277355, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 695.554982] env[61957]: DEBUG nova.network.neutron [None req-74fa7fa3-fb2c-4064-80f1-1405089fca10 tempest-ServerMetadataNegativeTestJSON-1373515003 tempest-ServerMetadataNegativeTestJSON-1373515003-project-member] [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] Successfully created port: ddfc7e7f-b45c-426a-8485-f28a208b52e8 {{(pid=61957) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 695.697889] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb27afd0-4321-4894-b696-48a66f9d5839 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.705943] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0cbd8b9-e55f-427c-af67-ee5ecbb28388 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.740144] env[61957]: DEBUG nova.compute.manager [None req-74fa7fa3-fb2c-4064-80f1-1405089fca10 tempest-ServerMetadataNegativeTestJSON-1373515003 tempest-ServerMetadataNegativeTestJSON-1373515003-project-member] [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] Start building block device mappings for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 695.746017] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81e426bb-ec14-47f1-9e77-7aefad276b93 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.752149] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44f88eed-1c1a-4436-b460-84085fd904a4 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.765361] env[61957]: DEBUG nova.compute.provider_tree [None req-b1f62379-e574-4175-a099-f938af5ddaa8 tempest-ServerPasswordTestJSON-1543821420 tempest-ServerPasswordTestJSON-1543821420-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 695.964682] env[61957]: DEBUG oslo_vmware.api [None req-2d3c8bae-3f1e-41b1-82a9-b293bdd893bd tempest-ServersAdmin275Test-2089451214 tempest-ServersAdmin275Test-2089451214-project-admin] Task: {'id': task-1277355, 'name': PowerOffVM_Task, 'duration_secs': 0.1221} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 695.964996] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-2d3c8bae-3f1e-41b1-82a9-b293bdd893bd tempest-ServersAdmin275Test-2089451214 tempest-ServersAdmin275Test-2089451214-project-admin] [instance: 188b0fd3-5d71-4feb-aca5-75a2bd28895a] Powered off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 695.966192] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-2d3c8bae-3f1e-41b1-82a9-b293bdd893bd tempest-ServersAdmin275Test-2089451214 tempest-ServersAdmin275Test-2089451214-project-admin] [instance: 188b0fd3-5d71-4feb-aca5-75a2bd28895a] Destroying instance {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 695.966968] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59b40e4a-610e-4efd-b2db-ed0bfd167f63 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.973553] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-2d3c8bae-3f1e-41b1-82a9-b293bdd893bd tempest-ServersAdmin275Test-2089451214 tempest-ServersAdmin275Test-2089451214-project-admin] [instance: 188b0fd3-5d71-4feb-aca5-75a2bd28895a] Unregistering the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 695.974299] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-cccb46e4-a6f7-41d5-a399-e948dedfda0c {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.998533] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-2d3c8bae-3f1e-41b1-82a9-b293bdd893bd tempest-ServersAdmin275Test-2089451214 tempest-ServersAdmin275Test-2089451214-project-admin] [instance: 188b0fd3-5d71-4feb-aca5-75a2bd28895a] Unregistered the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 695.998741] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-2d3c8bae-3f1e-41b1-82a9-b293bdd893bd tempest-ServersAdmin275Test-2089451214 tempest-ServersAdmin275Test-2089451214-project-admin] [instance: 188b0fd3-5d71-4feb-aca5-75a2bd28895a] Deleting contents of the VM from datastore datastore2 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 695.998970] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-2d3c8bae-3f1e-41b1-82a9-b293bdd893bd tempest-ServersAdmin275Test-2089451214 tempest-ServersAdmin275Test-2089451214-project-admin] Deleting the datastore file [datastore2] 188b0fd3-5d71-4feb-aca5-75a2bd28895a {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 695.999313] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e42b436a-804c-498e-a2b5-48d6f5c01fab {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.005411] env[61957]: DEBUG oslo_vmware.api [None req-2d3c8bae-3f1e-41b1-82a9-b293bdd893bd tempest-ServersAdmin275Test-2089451214 tempest-ServersAdmin275Test-2089451214-project-admin] Waiting for the task: (returnval){ [ 696.005411] env[61957]: value = "task-1277357" [ 696.005411] env[61957]: _type = "Task" [ 696.005411] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 696.012606] env[61957]: DEBUG oslo_vmware.api [None req-2d3c8bae-3f1e-41b1-82a9-b293bdd893bd tempest-ServersAdmin275Test-2089451214 tempest-ServersAdmin275Test-2089451214-project-admin] Task: {'id': task-1277357, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 696.241858] env[61957]: DEBUG nova.compute.manager [req-c63c3191-2145-4fe1-ad75-3aea8b41552e req-ac7a17ff-b022-48b0-9a45-79d910ffa16e service nova] [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] Received event network-changed-ddfc7e7f-b45c-426a-8485-f28a208b52e8 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 696.242082] env[61957]: DEBUG nova.compute.manager [req-c63c3191-2145-4fe1-ad75-3aea8b41552e req-ac7a17ff-b022-48b0-9a45-79d910ffa16e service nova] [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] Refreshing instance network info cache due to event network-changed-ddfc7e7f-b45c-426a-8485-f28a208b52e8. {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 696.242297] env[61957]: DEBUG oslo_concurrency.lockutils [req-c63c3191-2145-4fe1-ad75-3aea8b41552e req-ac7a17ff-b022-48b0-9a45-79d910ffa16e service nova] Acquiring lock "refresh_cache-3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 696.242436] env[61957]: DEBUG oslo_concurrency.lockutils [req-c63c3191-2145-4fe1-ad75-3aea8b41552e req-ac7a17ff-b022-48b0-9a45-79d910ffa16e service nova] Acquired lock "refresh_cache-3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 696.242590] env[61957]: DEBUG nova.network.neutron [req-c63c3191-2145-4fe1-ad75-3aea8b41552e req-ac7a17ff-b022-48b0-9a45-79d910ffa16e service nova] [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] Refreshing network info cache for port ddfc7e7f-b45c-426a-8485-f28a208b52e8 {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 696.268319] env[61957]: DEBUG nova.scheduler.client.report [None req-b1f62379-e574-4175-a099-f938af5ddaa8 tempest-ServerPasswordTestJSON-1543821420 tempest-ServerPasswordTestJSON-1543821420-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 696.429552] env[61957]: ERROR nova.compute.manager [None req-74fa7fa3-fb2c-4064-80f1-1405089fca10 tempest-ServerMetadataNegativeTestJSON-1373515003 tempest-ServerMetadataNegativeTestJSON-1373515003-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port ddfc7e7f-b45c-426a-8485-f28a208b52e8, please check neutron logs for more information. [ 696.429552] env[61957]: ERROR nova.compute.manager Traceback (most recent call last): [ 696.429552] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 696.429552] env[61957]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 696.429552] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 696.429552] env[61957]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 696.429552] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 696.429552] env[61957]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 696.429552] env[61957]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 696.429552] env[61957]: ERROR nova.compute.manager self.force_reraise() [ 696.429552] env[61957]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 696.429552] env[61957]: ERROR nova.compute.manager raise self.value [ 696.429552] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 696.429552] env[61957]: ERROR nova.compute.manager updated_port = self._update_port( [ 696.429552] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 696.429552] env[61957]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 696.430072] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 696.430072] env[61957]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 696.430072] env[61957]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port ddfc7e7f-b45c-426a-8485-f28a208b52e8, please check neutron logs for more information. [ 696.430072] env[61957]: ERROR nova.compute.manager [ 696.430072] env[61957]: Traceback (most recent call last): [ 696.430072] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 696.430072] env[61957]: listener.cb(fileno) [ 696.430072] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 696.430072] env[61957]: result = function(*args, **kwargs) [ 696.430072] env[61957]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 696.430072] env[61957]: return func(*args, **kwargs) [ 696.430072] env[61957]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 696.430072] env[61957]: raise e [ 696.430072] env[61957]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 696.430072] env[61957]: nwinfo = self.network_api.allocate_for_instance( [ 696.430072] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 696.430072] env[61957]: created_port_ids = self._update_ports_for_instance( [ 696.430072] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 696.430072] env[61957]: with excutils.save_and_reraise_exception(): [ 696.430072] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 696.430072] env[61957]: self.force_reraise() [ 696.430072] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 696.430072] env[61957]: raise self.value [ 696.430072] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 696.430072] env[61957]: updated_port = self._update_port( [ 696.430072] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 696.430072] env[61957]: _ensure_no_port_binding_failure(port) [ 696.430072] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 696.430072] env[61957]: raise exception.PortBindingFailed(port_id=port['id']) [ 696.430892] env[61957]: nova.exception.PortBindingFailed: Binding failed for port ddfc7e7f-b45c-426a-8485-f28a208b52e8, please check neutron logs for more information. [ 696.430892] env[61957]: Removing descriptor: 17 [ 696.515976] env[61957]: DEBUG oslo_vmware.api [None req-2d3c8bae-3f1e-41b1-82a9-b293bdd893bd tempest-ServersAdmin275Test-2089451214 tempest-ServersAdmin275Test-2089451214-project-admin] Task: {'id': task-1277357, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.086835} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 696.516310] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-2d3c8bae-3f1e-41b1-82a9-b293bdd893bd tempest-ServersAdmin275Test-2089451214 tempest-ServersAdmin275Test-2089451214-project-admin] Deleted the datastore file {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 696.516512] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-2d3c8bae-3f1e-41b1-82a9-b293bdd893bd tempest-ServersAdmin275Test-2089451214 tempest-ServersAdmin275Test-2089451214-project-admin] [instance: 188b0fd3-5d71-4feb-aca5-75a2bd28895a] Deleted contents of the VM from datastore datastore2 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 696.516688] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-2d3c8bae-3f1e-41b1-82a9-b293bdd893bd tempest-ServersAdmin275Test-2089451214 tempest-ServersAdmin275Test-2089451214-project-admin] [instance: 188b0fd3-5d71-4feb-aca5-75a2bd28895a] Instance destroyed {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 696.754302] env[61957]: DEBUG nova.compute.manager [None req-74fa7fa3-fb2c-4064-80f1-1405089fca10 tempest-ServerMetadataNegativeTestJSON-1373515003 tempest-ServerMetadataNegativeTestJSON-1373515003-project-member] [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] Start spawning the instance on the hypervisor. {{(pid=61957) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 696.761909] env[61957]: DEBUG nova.network.neutron [req-c63c3191-2145-4fe1-ad75-3aea8b41552e req-ac7a17ff-b022-48b0-9a45-79d910ffa16e service nova] [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 696.772671] env[61957]: DEBUG oslo_concurrency.lockutils [None req-b1f62379-e574-4175-a099-f938af5ddaa8 tempest-ServerPasswordTestJSON-1543821420 tempest-ServerPasswordTestJSON-1543821420-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.062s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 696.773285] env[61957]: ERROR nova.compute.manager [None req-b1f62379-e574-4175-a099-f938af5ddaa8 tempest-ServerPasswordTestJSON-1543821420 tempest-ServerPasswordTestJSON-1543821420-project-member] [instance: 49e99297-ffb9-4104-bd06-911243908828] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 4372cc6c-40b2-4067-93d2-12710b09600f, please check neutron logs for more information. [ 696.773285] env[61957]: ERROR nova.compute.manager [instance: 49e99297-ffb9-4104-bd06-911243908828] Traceback (most recent call last): [ 696.773285] env[61957]: ERROR nova.compute.manager [instance: 49e99297-ffb9-4104-bd06-911243908828] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 696.773285] env[61957]: ERROR nova.compute.manager [instance: 49e99297-ffb9-4104-bd06-911243908828] self.driver.spawn(context, instance, image_meta, [ 696.773285] env[61957]: ERROR nova.compute.manager [instance: 49e99297-ffb9-4104-bd06-911243908828] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 696.773285] env[61957]: ERROR nova.compute.manager [instance: 49e99297-ffb9-4104-bd06-911243908828] self._vmops.spawn(context, instance, image_meta, injected_files, [ 696.773285] env[61957]: ERROR nova.compute.manager [instance: 49e99297-ffb9-4104-bd06-911243908828] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 696.773285] env[61957]: ERROR nova.compute.manager [instance: 49e99297-ffb9-4104-bd06-911243908828] vm_ref = self.build_virtual_machine(instance, [ 696.773285] env[61957]: ERROR nova.compute.manager [instance: 49e99297-ffb9-4104-bd06-911243908828] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 696.773285] env[61957]: ERROR nova.compute.manager [instance: 49e99297-ffb9-4104-bd06-911243908828] vif_infos = vmwarevif.get_vif_info(self._session, [ 696.773285] env[61957]: ERROR nova.compute.manager [instance: 49e99297-ffb9-4104-bd06-911243908828] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 696.773578] env[61957]: ERROR nova.compute.manager [instance: 49e99297-ffb9-4104-bd06-911243908828] for vif in network_info: [ 696.773578] env[61957]: ERROR nova.compute.manager [instance: 49e99297-ffb9-4104-bd06-911243908828] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 696.773578] env[61957]: ERROR nova.compute.manager [instance: 49e99297-ffb9-4104-bd06-911243908828] return self._sync_wrapper(fn, *args, **kwargs) [ 696.773578] env[61957]: ERROR nova.compute.manager [instance: 49e99297-ffb9-4104-bd06-911243908828] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 696.773578] env[61957]: ERROR nova.compute.manager [instance: 49e99297-ffb9-4104-bd06-911243908828] self.wait() [ 696.773578] env[61957]: ERROR nova.compute.manager [instance: 49e99297-ffb9-4104-bd06-911243908828] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 696.773578] env[61957]: ERROR nova.compute.manager [instance: 49e99297-ffb9-4104-bd06-911243908828] self[:] = self._gt.wait() [ 696.773578] env[61957]: ERROR nova.compute.manager [instance: 49e99297-ffb9-4104-bd06-911243908828] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 696.773578] env[61957]: ERROR nova.compute.manager [instance: 49e99297-ffb9-4104-bd06-911243908828] return self._exit_event.wait() [ 696.773578] env[61957]: ERROR nova.compute.manager [instance: 49e99297-ffb9-4104-bd06-911243908828] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 696.773578] env[61957]: ERROR nova.compute.manager [instance: 49e99297-ffb9-4104-bd06-911243908828] current.throw(*self._exc) [ 696.773578] env[61957]: ERROR nova.compute.manager [instance: 49e99297-ffb9-4104-bd06-911243908828] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 696.773578] env[61957]: ERROR nova.compute.manager [instance: 49e99297-ffb9-4104-bd06-911243908828] result = function(*args, **kwargs) [ 696.773903] env[61957]: ERROR nova.compute.manager [instance: 49e99297-ffb9-4104-bd06-911243908828] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 696.773903] env[61957]: ERROR nova.compute.manager [instance: 49e99297-ffb9-4104-bd06-911243908828] return func(*args, **kwargs) [ 696.773903] env[61957]: ERROR nova.compute.manager [instance: 49e99297-ffb9-4104-bd06-911243908828] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 696.773903] env[61957]: ERROR nova.compute.manager [instance: 49e99297-ffb9-4104-bd06-911243908828] raise e [ 696.773903] env[61957]: ERROR nova.compute.manager [instance: 49e99297-ffb9-4104-bd06-911243908828] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 696.773903] env[61957]: ERROR nova.compute.manager [instance: 49e99297-ffb9-4104-bd06-911243908828] nwinfo = self.network_api.allocate_for_instance( [ 696.773903] env[61957]: ERROR nova.compute.manager [instance: 49e99297-ffb9-4104-bd06-911243908828] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 696.773903] env[61957]: ERROR nova.compute.manager [instance: 49e99297-ffb9-4104-bd06-911243908828] created_port_ids = self._update_ports_for_instance( [ 696.773903] env[61957]: ERROR nova.compute.manager [instance: 49e99297-ffb9-4104-bd06-911243908828] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 696.773903] env[61957]: ERROR nova.compute.manager [instance: 49e99297-ffb9-4104-bd06-911243908828] with excutils.save_and_reraise_exception(): [ 696.773903] env[61957]: ERROR nova.compute.manager [instance: 49e99297-ffb9-4104-bd06-911243908828] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 696.773903] env[61957]: ERROR nova.compute.manager [instance: 49e99297-ffb9-4104-bd06-911243908828] self.force_reraise() [ 696.773903] env[61957]: ERROR nova.compute.manager [instance: 49e99297-ffb9-4104-bd06-911243908828] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 696.774243] env[61957]: ERROR nova.compute.manager [instance: 49e99297-ffb9-4104-bd06-911243908828] raise self.value [ 696.774243] env[61957]: ERROR nova.compute.manager [instance: 49e99297-ffb9-4104-bd06-911243908828] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 696.774243] env[61957]: ERROR nova.compute.manager [instance: 49e99297-ffb9-4104-bd06-911243908828] updated_port = self._update_port( [ 696.774243] env[61957]: ERROR nova.compute.manager [instance: 49e99297-ffb9-4104-bd06-911243908828] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 696.774243] env[61957]: ERROR nova.compute.manager [instance: 49e99297-ffb9-4104-bd06-911243908828] _ensure_no_port_binding_failure(port) [ 696.774243] env[61957]: ERROR nova.compute.manager [instance: 49e99297-ffb9-4104-bd06-911243908828] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 696.774243] env[61957]: ERROR nova.compute.manager [instance: 49e99297-ffb9-4104-bd06-911243908828] raise exception.PortBindingFailed(port_id=port['id']) [ 696.774243] env[61957]: ERROR nova.compute.manager [instance: 49e99297-ffb9-4104-bd06-911243908828] nova.exception.PortBindingFailed: Binding failed for port 4372cc6c-40b2-4067-93d2-12710b09600f, please check neutron logs for more information. [ 696.774243] env[61957]: ERROR nova.compute.manager [instance: 49e99297-ffb9-4104-bd06-911243908828] [ 696.774243] env[61957]: DEBUG nova.compute.utils [None req-b1f62379-e574-4175-a099-f938af5ddaa8 tempest-ServerPasswordTestJSON-1543821420 tempest-ServerPasswordTestJSON-1543821420-project-member] [instance: 49e99297-ffb9-4104-bd06-911243908828] Binding failed for port 4372cc6c-40b2-4067-93d2-12710b09600f, please check neutron logs for more information. {{(pid=61957) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 696.776940] env[61957]: DEBUG nova.virt.hardware [None req-74fa7fa3-fb2c-4064-80f1-1405089fca10 tempest-ServerMetadataNegativeTestJSON-1373515003 tempest-ServerMetadataNegativeTestJSON-1373515003-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T17:22:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T17:21:50Z,direct_url=,disk_format='vmdk',id=11c76a2c-f705-470a-ba9d-4657858bab38,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='81d3096d0f094373913fc3dc8f5c3c6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T17:21:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 696.777177] env[61957]: DEBUG nova.virt.hardware [None req-74fa7fa3-fb2c-4064-80f1-1405089fca10 tempest-ServerMetadataNegativeTestJSON-1373515003 tempest-ServerMetadataNegativeTestJSON-1373515003-project-member] Flavor limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 696.777359] env[61957]: DEBUG nova.virt.hardware [None req-74fa7fa3-fb2c-4064-80f1-1405089fca10 tempest-ServerMetadataNegativeTestJSON-1373515003 tempest-ServerMetadataNegativeTestJSON-1373515003-project-member] Image limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 696.777513] env[61957]: DEBUG nova.virt.hardware [None req-74fa7fa3-fb2c-4064-80f1-1405089fca10 tempest-ServerMetadataNegativeTestJSON-1373515003 tempest-ServerMetadataNegativeTestJSON-1373515003-project-member] Flavor pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 696.777653] env[61957]: DEBUG nova.virt.hardware [None req-74fa7fa3-fb2c-4064-80f1-1405089fca10 tempest-ServerMetadataNegativeTestJSON-1373515003 tempest-ServerMetadataNegativeTestJSON-1373515003-project-member] Image pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 696.777793] env[61957]: DEBUG nova.virt.hardware [None req-74fa7fa3-fb2c-4064-80f1-1405089fca10 tempest-ServerMetadataNegativeTestJSON-1373515003 tempest-ServerMetadataNegativeTestJSON-1373515003-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 696.777994] env[61957]: DEBUG nova.virt.hardware [None req-74fa7fa3-fb2c-4064-80f1-1405089fca10 tempest-ServerMetadataNegativeTestJSON-1373515003 tempest-ServerMetadataNegativeTestJSON-1373515003-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 696.778192] env[61957]: DEBUG nova.virt.hardware [None req-74fa7fa3-fb2c-4064-80f1-1405089fca10 tempest-ServerMetadataNegativeTestJSON-1373515003 tempest-ServerMetadataNegativeTestJSON-1373515003-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 696.778367] env[61957]: DEBUG nova.virt.hardware [None req-74fa7fa3-fb2c-4064-80f1-1405089fca10 tempest-ServerMetadataNegativeTestJSON-1373515003 tempest-ServerMetadataNegativeTestJSON-1373515003-project-member] Got 1 possible topologies {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 696.778532] env[61957]: DEBUG nova.virt.hardware [None req-74fa7fa3-fb2c-4064-80f1-1405089fca10 tempest-ServerMetadataNegativeTestJSON-1373515003 tempest-ServerMetadataNegativeTestJSON-1373515003-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 696.778691] env[61957]: DEBUG nova.virt.hardware [None req-74fa7fa3-fb2c-4064-80f1-1405089fca10 tempest-ServerMetadataNegativeTestJSON-1373515003 tempest-ServerMetadataNegativeTestJSON-1373515003-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 696.778989] env[61957]: DEBUG oslo_concurrency.lockutils [None req-dbfaab8a-0784-4441-ad61-79e56d9c3c60 tempest-FloatingIPsAssociationTestJSON-1786153562 tempest-FloatingIPsAssociationTestJSON-1786153562-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.620s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 696.780485] env[61957]: INFO nova.compute.claims [None req-dbfaab8a-0784-4441-ad61-79e56d9c3c60 tempest-FloatingIPsAssociationTestJSON-1786153562 tempest-FloatingIPsAssociationTestJSON-1786153562-project-member] [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 696.783436] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57b99271-6aa6-475f-abe7-87d4740abade {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.786508] env[61957]: DEBUG nova.compute.manager [None req-b1f62379-e574-4175-a099-f938af5ddaa8 tempest-ServerPasswordTestJSON-1543821420 tempest-ServerPasswordTestJSON-1543821420-project-member] [instance: 49e99297-ffb9-4104-bd06-911243908828] Build of instance 49e99297-ffb9-4104-bd06-911243908828 was re-scheduled: Binding failed for port 4372cc6c-40b2-4067-93d2-12710b09600f, please check neutron logs for more information. {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 696.786961] env[61957]: DEBUG nova.compute.manager [None req-b1f62379-e574-4175-a099-f938af5ddaa8 tempest-ServerPasswordTestJSON-1543821420 tempest-ServerPasswordTestJSON-1543821420-project-member] [instance: 49e99297-ffb9-4104-bd06-911243908828] Unplugging VIFs for instance {{(pid=61957) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 696.787210] env[61957]: DEBUG oslo_concurrency.lockutils [None req-b1f62379-e574-4175-a099-f938af5ddaa8 tempest-ServerPasswordTestJSON-1543821420 tempest-ServerPasswordTestJSON-1543821420-project-member] Acquiring lock "refresh_cache-49e99297-ffb9-4104-bd06-911243908828" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 696.787312] env[61957]: DEBUG oslo_concurrency.lockutils [None req-b1f62379-e574-4175-a099-f938af5ddaa8 tempest-ServerPasswordTestJSON-1543821420 tempest-ServerPasswordTestJSON-1543821420-project-member] Acquired lock "refresh_cache-49e99297-ffb9-4104-bd06-911243908828" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 696.791099] env[61957]: DEBUG nova.network.neutron [None req-b1f62379-e574-4175-a099-f938af5ddaa8 tempest-ServerPasswordTestJSON-1543821420 tempest-ServerPasswordTestJSON-1543821420-project-member] [instance: 49e99297-ffb9-4104-bd06-911243908828] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 696.797750] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80a05b86-bc60-4b0f-9ab2-7df013ca6ae6 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.818852] env[61957]: ERROR nova.compute.manager [None req-74fa7fa3-fb2c-4064-80f1-1405089fca10 tempest-ServerMetadataNegativeTestJSON-1373515003 tempest-ServerMetadataNegativeTestJSON-1373515003-project-member] [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port ddfc7e7f-b45c-426a-8485-f28a208b52e8, please check neutron logs for more information. [ 696.818852] env[61957]: ERROR nova.compute.manager [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] Traceback (most recent call last): [ 696.818852] env[61957]: ERROR nova.compute.manager [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 696.818852] env[61957]: ERROR nova.compute.manager [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] yield resources [ 696.818852] env[61957]: ERROR nova.compute.manager [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 696.818852] env[61957]: ERROR nova.compute.manager [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] self.driver.spawn(context, instance, image_meta, [ 696.818852] env[61957]: ERROR nova.compute.manager [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 696.818852] env[61957]: ERROR nova.compute.manager [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] self._vmops.spawn(context, instance, image_meta, injected_files, [ 696.818852] env[61957]: ERROR nova.compute.manager [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 696.818852] env[61957]: ERROR nova.compute.manager [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] vm_ref = self.build_virtual_machine(instance, [ 696.818852] env[61957]: ERROR nova.compute.manager [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 696.819643] env[61957]: ERROR nova.compute.manager [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] vif_infos = vmwarevif.get_vif_info(self._session, [ 696.819643] env[61957]: ERROR nova.compute.manager [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 696.819643] env[61957]: ERROR nova.compute.manager [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] for vif in network_info: [ 696.819643] env[61957]: ERROR nova.compute.manager [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 696.819643] env[61957]: ERROR nova.compute.manager [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] return self._sync_wrapper(fn, *args, **kwargs) [ 696.819643] env[61957]: ERROR nova.compute.manager [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 696.819643] env[61957]: ERROR nova.compute.manager [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] self.wait() [ 696.819643] env[61957]: ERROR nova.compute.manager [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 696.819643] env[61957]: ERROR nova.compute.manager [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] self[:] = self._gt.wait() [ 696.819643] env[61957]: ERROR nova.compute.manager [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 696.819643] env[61957]: ERROR nova.compute.manager [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] return self._exit_event.wait() [ 696.819643] env[61957]: ERROR nova.compute.manager [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 696.819643] env[61957]: ERROR nova.compute.manager [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] current.throw(*self._exc) [ 696.819953] env[61957]: ERROR nova.compute.manager [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 696.819953] env[61957]: ERROR nova.compute.manager [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] result = function(*args, **kwargs) [ 696.819953] env[61957]: ERROR nova.compute.manager [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 696.819953] env[61957]: ERROR nova.compute.manager [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] return func(*args, **kwargs) [ 696.819953] env[61957]: ERROR nova.compute.manager [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 696.819953] env[61957]: ERROR nova.compute.manager [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] raise e [ 696.819953] env[61957]: ERROR nova.compute.manager [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 696.819953] env[61957]: ERROR nova.compute.manager [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] nwinfo = self.network_api.allocate_for_instance( [ 696.819953] env[61957]: ERROR nova.compute.manager [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 696.819953] env[61957]: ERROR nova.compute.manager [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] created_port_ids = self._update_ports_for_instance( [ 696.819953] env[61957]: ERROR nova.compute.manager [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 696.819953] env[61957]: ERROR nova.compute.manager [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] with excutils.save_and_reraise_exception(): [ 696.819953] env[61957]: ERROR nova.compute.manager [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 696.820295] env[61957]: ERROR nova.compute.manager [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] self.force_reraise() [ 696.820295] env[61957]: ERROR nova.compute.manager [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 696.820295] env[61957]: ERROR nova.compute.manager [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] raise self.value [ 696.820295] env[61957]: ERROR nova.compute.manager [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 696.820295] env[61957]: ERROR nova.compute.manager [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] updated_port = self._update_port( [ 696.820295] env[61957]: ERROR nova.compute.manager [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 696.820295] env[61957]: ERROR nova.compute.manager [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] _ensure_no_port_binding_failure(port) [ 696.820295] env[61957]: ERROR nova.compute.manager [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 696.820295] env[61957]: ERROR nova.compute.manager [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] raise exception.PortBindingFailed(port_id=port['id']) [ 696.820295] env[61957]: ERROR nova.compute.manager [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] nova.exception.PortBindingFailed: Binding failed for port ddfc7e7f-b45c-426a-8485-f28a208b52e8, please check neutron logs for more information. [ 696.820295] env[61957]: ERROR nova.compute.manager [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] [ 696.820295] env[61957]: INFO nova.compute.manager [None req-74fa7fa3-fb2c-4064-80f1-1405089fca10 tempest-ServerMetadataNegativeTestJSON-1373515003 tempest-ServerMetadataNegativeTestJSON-1373515003-project-member] [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] Terminating instance [ 696.820603] env[61957]: DEBUG oslo_concurrency.lockutils [None req-74fa7fa3-fb2c-4064-80f1-1405089fca10 tempest-ServerMetadataNegativeTestJSON-1373515003 tempest-ServerMetadataNegativeTestJSON-1373515003-project-member] Acquiring lock "refresh_cache-3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 696.840000] env[61957]: DEBUG nova.network.neutron [req-c63c3191-2145-4fe1-ad75-3aea8b41552e req-ac7a17ff-b022-48b0-9a45-79d910ffa16e service nova] [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 697.310844] env[61957]: DEBUG nova.network.neutron [None req-b1f62379-e574-4175-a099-f938af5ddaa8 tempest-ServerPasswordTestJSON-1543821420 tempest-ServerPasswordTestJSON-1543821420-project-member] [instance: 49e99297-ffb9-4104-bd06-911243908828] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 697.342039] env[61957]: DEBUG oslo_concurrency.lockutils [req-c63c3191-2145-4fe1-ad75-3aea8b41552e req-ac7a17ff-b022-48b0-9a45-79d910ffa16e service nova] Releasing lock "refresh_cache-3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 697.342660] env[61957]: DEBUG oslo_concurrency.lockutils [None req-74fa7fa3-fb2c-4064-80f1-1405089fca10 tempest-ServerMetadataNegativeTestJSON-1373515003 tempest-ServerMetadataNegativeTestJSON-1373515003-project-member] Acquired lock "refresh_cache-3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 697.342660] env[61957]: DEBUG nova.network.neutron [None req-74fa7fa3-fb2c-4064-80f1-1405089fca10 tempest-ServerMetadataNegativeTestJSON-1373515003 tempest-ServerMetadataNegativeTestJSON-1373515003-project-member] [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 697.387620] env[61957]: DEBUG nova.network.neutron [None req-b1f62379-e574-4175-a099-f938af5ddaa8 tempest-ServerPasswordTestJSON-1543821420 tempest-ServerPasswordTestJSON-1543821420-project-member] [instance: 49e99297-ffb9-4104-bd06-911243908828] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 697.545383] env[61957]: DEBUG nova.virt.hardware [None req-2d3c8bae-3f1e-41b1-82a9-b293bdd893bd tempest-ServersAdmin275Test-2089451214 tempest-ServersAdmin275Test-2089451214-project-admin] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T17:22:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T17:21:50Z,direct_url=,disk_format='vmdk',id=11c76a2c-f705-470a-ba9d-4657858bab38,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='81d3096d0f094373913fc3dc8f5c3c6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T17:21:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 697.545630] env[61957]: DEBUG nova.virt.hardware [None req-2d3c8bae-3f1e-41b1-82a9-b293bdd893bd tempest-ServersAdmin275Test-2089451214 tempest-ServersAdmin275Test-2089451214-project-admin] Flavor limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 697.545785] env[61957]: DEBUG nova.virt.hardware [None req-2d3c8bae-3f1e-41b1-82a9-b293bdd893bd tempest-ServersAdmin275Test-2089451214 tempest-ServersAdmin275Test-2089451214-project-admin] Image limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 697.545966] env[61957]: DEBUG nova.virt.hardware [None req-2d3c8bae-3f1e-41b1-82a9-b293bdd893bd tempest-ServersAdmin275Test-2089451214 tempest-ServersAdmin275Test-2089451214-project-admin] Flavor pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 697.546177] env[61957]: DEBUG nova.virt.hardware [None req-2d3c8bae-3f1e-41b1-82a9-b293bdd893bd tempest-ServersAdmin275Test-2089451214 tempest-ServersAdmin275Test-2089451214-project-admin] Image pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 697.546324] env[61957]: DEBUG nova.virt.hardware [None req-2d3c8bae-3f1e-41b1-82a9-b293bdd893bd tempest-ServersAdmin275Test-2089451214 tempest-ServersAdmin275Test-2089451214-project-admin] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 697.546529] env[61957]: DEBUG nova.virt.hardware [None req-2d3c8bae-3f1e-41b1-82a9-b293bdd893bd tempest-ServersAdmin275Test-2089451214 tempest-ServersAdmin275Test-2089451214-project-admin] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 697.546685] env[61957]: DEBUG nova.virt.hardware [None req-2d3c8bae-3f1e-41b1-82a9-b293bdd893bd tempest-ServersAdmin275Test-2089451214 tempest-ServersAdmin275Test-2089451214-project-admin] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 697.546845] env[61957]: DEBUG nova.virt.hardware [None req-2d3c8bae-3f1e-41b1-82a9-b293bdd893bd tempest-ServersAdmin275Test-2089451214 tempest-ServersAdmin275Test-2089451214-project-admin] Got 1 possible topologies {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 697.547011] env[61957]: DEBUG nova.virt.hardware [None req-2d3c8bae-3f1e-41b1-82a9-b293bdd893bd tempest-ServersAdmin275Test-2089451214 tempest-ServersAdmin275Test-2089451214-project-admin] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 697.547197] env[61957]: DEBUG nova.virt.hardware [None req-2d3c8bae-3f1e-41b1-82a9-b293bdd893bd tempest-ServersAdmin275Test-2089451214 tempest-ServersAdmin275Test-2089451214-project-admin] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 697.548100] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f02db5f-bc8b-40b1-b2a0-27c444c9720e {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.555726] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae7d7443-c214-45bd-be2a-a1ac5614c324 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.568877] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-2d3c8bae-3f1e-41b1-82a9-b293bdd893bd tempest-ServersAdmin275Test-2089451214 tempest-ServersAdmin275Test-2089451214-project-admin] [instance: 188b0fd3-5d71-4feb-aca5-75a2bd28895a] Instance VIF info [] {{(pid=61957) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 697.574322] env[61957]: DEBUG oslo.service.loopingcall [None req-2d3c8bae-3f1e-41b1-82a9-b293bdd893bd tempest-ServersAdmin275Test-2089451214 tempest-ServersAdmin275Test-2089451214-project-admin] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 697.574541] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 188b0fd3-5d71-4feb-aca5-75a2bd28895a] Creating VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 697.574732] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c29c550a-e4ff-460a-bc9e-749ec1fe974b {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.591165] env[61957]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 697.591165] env[61957]: value = "task-1277358" [ 697.591165] env[61957]: _type = "Task" [ 697.591165] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 697.597981] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277358, 'name': CreateVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 697.861665] env[61957]: DEBUG nova.network.neutron [None req-74fa7fa3-fb2c-4064-80f1-1405089fca10 tempest-ServerMetadataNegativeTestJSON-1373515003 tempest-ServerMetadataNegativeTestJSON-1373515003-project-member] [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 697.891135] env[61957]: DEBUG oslo_concurrency.lockutils [None req-b1f62379-e574-4175-a099-f938af5ddaa8 tempest-ServerPasswordTestJSON-1543821420 tempest-ServerPasswordTestJSON-1543821420-project-member] Releasing lock "refresh_cache-49e99297-ffb9-4104-bd06-911243908828" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 697.891375] env[61957]: DEBUG nova.compute.manager [None req-b1f62379-e574-4175-a099-f938af5ddaa8 tempest-ServerPasswordTestJSON-1543821420 tempest-ServerPasswordTestJSON-1543821420-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61957) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 697.891553] env[61957]: DEBUG nova.compute.manager [None req-b1f62379-e574-4175-a099-f938af5ddaa8 tempest-ServerPasswordTestJSON-1543821420 tempest-ServerPasswordTestJSON-1543821420-project-member] [instance: 49e99297-ffb9-4104-bd06-911243908828] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 697.891719] env[61957]: DEBUG nova.network.neutron [None req-b1f62379-e574-4175-a099-f938af5ddaa8 tempest-ServerPasswordTestJSON-1543821420 tempest-ServerPasswordTestJSON-1543821420-project-member] [instance: 49e99297-ffb9-4104-bd06-911243908828] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 697.938618] env[61957]: DEBUG nova.network.neutron [None req-74fa7fa3-fb2c-4064-80f1-1405089fca10 tempest-ServerMetadataNegativeTestJSON-1373515003 tempest-ServerMetadataNegativeTestJSON-1373515003-project-member] [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 697.942276] env[61957]: DEBUG nova.network.neutron [None req-b1f62379-e574-4175-a099-f938af5ddaa8 tempest-ServerPasswordTestJSON-1543821420 tempest-ServerPasswordTestJSON-1543821420-project-member] [instance: 49e99297-ffb9-4104-bd06-911243908828] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 698.101907] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277358, 'name': CreateVM_Task} progress is 99%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 698.136872] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10b618ca-175d-427e-b4a4-4a80078f5bb6 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.143956] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec56aa63-885d-44ee-a197-db572cd0d441 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.174247] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f33abe4c-73bf-4156-a3c1-a76b70bb2677 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.181178] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d9dfc85-6545-4264-adcf-7eeb117ffcab {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.193940] env[61957]: DEBUG nova.compute.provider_tree [None req-dbfaab8a-0784-4441-ad61-79e56d9c3c60 tempest-FloatingIPsAssociationTestJSON-1786153562 tempest-FloatingIPsAssociationTestJSON-1786153562-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 698.266647] env[61957]: DEBUG nova.compute.manager [req-22c1dcef-0abc-44c3-a0ab-4fe01ce26ae3 req-eddd4d68-a0a1-481f-bebe-8c046c8adfbf service nova] [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] Received event network-vif-deleted-ddfc7e7f-b45c-426a-8485-f28a208b52e8 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 698.444174] env[61957]: DEBUG oslo_concurrency.lockutils [None req-74fa7fa3-fb2c-4064-80f1-1405089fca10 tempest-ServerMetadataNegativeTestJSON-1373515003 tempest-ServerMetadataNegativeTestJSON-1373515003-project-member] Releasing lock "refresh_cache-3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 698.444585] env[61957]: DEBUG nova.compute.manager [None req-74fa7fa3-fb2c-4064-80f1-1405089fca10 tempest-ServerMetadataNegativeTestJSON-1373515003 tempest-ServerMetadataNegativeTestJSON-1373515003-project-member] [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] Start destroying the instance on the hypervisor. {{(pid=61957) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 698.444784] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-74fa7fa3-fb2c-4064-80f1-1405089fca10 tempest-ServerMetadataNegativeTestJSON-1373515003 tempest-ServerMetadataNegativeTestJSON-1373515003-project-member] [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] Destroying instance {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 698.445095] env[61957]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a786e196-f65c-4173-8099-c6314c4d4bd4 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.448031] env[61957]: DEBUG nova.network.neutron [None req-b1f62379-e574-4175-a099-f938af5ddaa8 tempest-ServerPasswordTestJSON-1543821420 tempest-ServerPasswordTestJSON-1543821420-project-member] [instance: 49e99297-ffb9-4104-bd06-911243908828] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 698.455839] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea9a62f1-4182-4090-8657-7f24ae577558 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.477623] env[61957]: WARNING nova.virt.vmwareapi.vmops [None req-74fa7fa3-fb2c-4064-80f1-1405089fca10 tempest-ServerMetadataNegativeTestJSON-1373515003 tempest-ServerMetadataNegativeTestJSON-1373515003-project-member] [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1 could not be found. [ 698.477831] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-74fa7fa3-fb2c-4064-80f1-1405089fca10 tempest-ServerMetadataNegativeTestJSON-1373515003 tempest-ServerMetadataNegativeTestJSON-1373515003-project-member] [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] Instance destroyed {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 698.478015] env[61957]: INFO nova.compute.manager [None req-74fa7fa3-fb2c-4064-80f1-1405089fca10 tempest-ServerMetadataNegativeTestJSON-1373515003 tempest-ServerMetadataNegativeTestJSON-1373515003-project-member] [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] Took 0.03 seconds to destroy the instance on the hypervisor. [ 698.478322] env[61957]: DEBUG oslo.service.loopingcall [None req-74fa7fa3-fb2c-4064-80f1-1405089fca10 tempest-ServerMetadataNegativeTestJSON-1373515003 tempest-ServerMetadataNegativeTestJSON-1373515003-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 698.478538] env[61957]: DEBUG nova.compute.manager [-] [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 698.478632] env[61957]: DEBUG nova.network.neutron [-] [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 698.493991] env[61957]: DEBUG nova.network.neutron [-] [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 698.601521] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277358, 'name': CreateVM_Task} progress is 99%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 698.697077] env[61957]: DEBUG nova.scheduler.client.report [None req-dbfaab8a-0784-4441-ad61-79e56d9c3c60 tempest-FloatingIPsAssociationTestJSON-1786153562 tempest-FloatingIPsAssociationTestJSON-1786153562-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 698.949745] env[61957]: INFO nova.compute.manager [None req-b1f62379-e574-4175-a099-f938af5ddaa8 tempest-ServerPasswordTestJSON-1543821420 tempest-ServerPasswordTestJSON-1543821420-project-member] [instance: 49e99297-ffb9-4104-bd06-911243908828] Took 1.06 seconds to deallocate network for instance. [ 698.995771] env[61957]: DEBUG nova.network.neutron [-] [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 699.102545] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277358, 'name': CreateVM_Task, 'duration_secs': 1.237912} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 699.102722] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 188b0fd3-5d71-4feb-aca5-75a2bd28895a] Created VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 699.103140] env[61957]: DEBUG oslo_concurrency.lockutils [None req-2d3c8bae-3f1e-41b1-82a9-b293bdd893bd tempest-ServersAdmin275Test-2089451214 tempest-ServersAdmin275Test-2089451214-project-admin] Acquiring lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 699.103372] env[61957]: DEBUG oslo_concurrency.lockutils [None req-2d3c8bae-3f1e-41b1-82a9-b293bdd893bd tempest-ServersAdmin275Test-2089451214 tempest-ServersAdmin275Test-2089451214-project-admin] Acquired lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 699.103599] env[61957]: DEBUG oslo_concurrency.lockutils [None req-2d3c8bae-3f1e-41b1-82a9-b293bdd893bd tempest-ServersAdmin275Test-2089451214 tempest-ServersAdmin275Test-2089451214-project-admin] Acquired external semaphore "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 699.103837] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fdd677cf-7ea5-4001-bedc-ef58a1c1c177 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.108057] env[61957]: DEBUG oslo_vmware.api [None req-2d3c8bae-3f1e-41b1-82a9-b293bdd893bd tempest-ServersAdmin275Test-2089451214 tempest-ServersAdmin275Test-2089451214-project-admin] Waiting for the task: (returnval){ [ 699.108057] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52efbb42-646f-155a-9ab6-b8c531b2d82c" [ 699.108057] env[61957]: _type = "Task" [ 699.108057] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 699.115228] env[61957]: DEBUG oslo_vmware.api [None req-2d3c8bae-3f1e-41b1-82a9-b293bdd893bd tempest-ServersAdmin275Test-2089451214 tempest-ServersAdmin275Test-2089451214-project-admin] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52efbb42-646f-155a-9ab6-b8c531b2d82c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 699.201833] env[61957]: DEBUG oslo_concurrency.lockutils [None req-dbfaab8a-0784-4441-ad61-79e56d9c3c60 tempest-FloatingIPsAssociationTestJSON-1786153562 tempest-FloatingIPsAssociationTestJSON-1786153562-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.423s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 699.202532] env[61957]: DEBUG nova.compute.manager [None req-dbfaab8a-0784-4441-ad61-79e56d9c3c60 tempest-FloatingIPsAssociationTestJSON-1786153562 tempest-FloatingIPsAssociationTestJSON-1786153562-project-member] [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] Start building networks asynchronously for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 699.205401] env[61957]: DEBUG oslo_concurrency.lockutils [None req-fc6b77c8-b5cd-4608-beab-d2cb5c9f7440 tempest-ImagesNegativeTestJSON-1548152881 tempest-ImagesNegativeTestJSON-1548152881-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 30.682s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 699.498038] env[61957]: INFO nova.compute.manager [-] [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] Took 1.02 seconds to deallocate network for instance. [ 699.501789] env[61957]: DEBUG nova.compute.claims [None req-74fa7fa3-fb2c-4064-80f1-1405089fca10 tempest-ServerMetadataNegativeTestJSON-1373515003 tempest-ServerMetadataNegativeTestJSON-1373515003-project-member] [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] Aborting claim: {{(pid=61957) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 699.501967] env[61957]: DEBUG oslo_concurrency.lockutils [None req-74fa7fa3-fb2c-4064-80f1-1405089fca10 tempest-ServerMetadataNegativeTestJSON-1373515003 tempest-ServerMetadataNegativeTestJSON-1373515003-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 699.619814] env[61957]: DEBUG oslo_vmware.api [None req-2d3c8bae-3f1e-41b1-82a9-b293bdd893bd tempest-ServersAdmin275Test-2089451214 tempest-ServersAdmin275Test-2089451214-project-admin] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52efbb42-646f-155a-9ab6-b8c531b2d82c, 'name': SearchDatastore_Task, 'duration_secs': 0.010518} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 699.620121] env[61957]: DEBUG oslo_concurrency.lockutils [None req-2d3c8bae-3f1e-41b1-82a9-b293bdd893bd tempest-ServersAdmin275Test-2089451214 tempest-ServersAdmin275Test-2089451214-project-admin] Releasing lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 699.620358] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-2d3c8bae-3f1e-41b1-82a9-b293bdd893bd tempest-ServersAdmin275Test-2089451214 tempest-ServersAdmin275Test-2089451214-project-admin] [instance: 188b0fd3-5d71-4feb-aca5-75a2bd28895a] Processing image 11c76a2c-f705-470a-ba9d-4657858bab38 {{(pid=61957) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 699.620589] env[61957]: DEBUG oslo_concurrency.lockutils [None req-2d3c8bae-3f1e-41b1-82a9-b293bdd893bd tempest-ServersAdmin275Test-2089451214 tempest-ServersAdmin275Test-2089451214-project-admin] Acquiring lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 699.620733] env[61957]: DEBUG oslo_concurrency.lockutils [None req-2d3c8bae-3f1e-41b1-82a9-b293bdd893bd tempest-ServersAdmin275Test-2089451214 tempest-ServersAdmin275Test-2089451214-project-admin] Acquired lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 699.620908] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-2d3c8bae-3f1e-41b1-82a9-b293bdd893bd tempest-ServersAdmin275Test-2089451214 tempest-ServersAdmin275Test-2089451214-project-admin] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 699.621182] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b09ffcd0-19e1-489d-b264-5c6803be4c38 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.628572] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-2d3c8bae-3f1e-41b1-82a9-b293bdd893bd tempest-ServersAdmin275Test-2089451214 tempest-ServersAdmin275Test-2089451214-project-admin] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 699.628737] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-2d3c8bae-3f1e-41b1-82a9-b293bdd893bd tempest-ServersAdmin275Test-2089451214 tempest-ServersAdmin275Test-2089451214-project-admin] Folder [datastore2] devstack-image-cache_base created. {{(pid=61957) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 699.629527] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e13b0de0-2a40-4d24-8d20-26052cc8e621 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.635287] env[61957]: DEBUG oslo_vmware.api [None req-2d3c8bae-3f1e-41b1-82a9-b293bdd893bd tempest-ServersAdmin275Test-2089451214 tempest-ServersAdmin275Test-2089451214-project-admin] Waiting for the task: (returnval){ [ 699.635287] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]5296d4a9-0176-1915-0c90-7fdb1b68bd3d" [ 699.635287] env[61957]: _type = "Task" [ 699.635287] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 699.642660] env[61957]: DEBUG oslo_vmware.api [None req-2d3c8bae-3f1e-41b1-82a9-b293bdd893bd tempest-ServersAdmin275Test-2089451214 tempest-ServersAdmin275Test-2089451214-project-admin] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]5296d4a9-0176-1915-0c90-7fdb1b68bd3d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 699.710936] env[61957]: DEBUG nova.compute.utils [None req-dbfaab8a-0784-4441-ad61-79e56d9c3c60 tempest-FloatingIPsAssociationTestJSON-1786153562 tempest-FloatingIPsAssociationTestJSON-1786153562-project-member] Using /dev/sd instead of None {{(pid=61957) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 699.717196] env[61957]: DEBUG nova.compute.manager [None req-dbfaab8a-0784-4441-ad61-79e56d9c3c60 tempest-FloatingIPsAssociationTestJSON-1786153562 tempest-FloatingIPsAssociationTestJSON-1786153562-project-member] [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] Allocating IP information in the background. {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 699.718054] env[61957]: DEBUG nova.network.neutron [None req-dbfaab8a-0784-4441-ad61-79e56d9c3c60 tempest-FloatingIPsAssociationTestJSON-1786153562 tempest-FloatingIPsAssociationTestJSON-1786153562-project-member] [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] allocate_for_instance() {{(pid=61957) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 699.767708] env[61957]: DEBUG nova.policy [None req-dbfaab8a-0784-4441-ad61-79e56d9c3c60 tempest-FloatingIPsAssociationTestJSON-1786153562 tempest-FloatingIPsAssociationTestJSON-1786153562-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2dacb3f45da04508a1f8a64d5246508a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8b8ea8d81a474ff2ae7fc0ab08cb1780', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61957) authorize /opt/stack/nova/nova/policy.py:203}} [ 699.986017] env[61957]: INFO nova.scheduler.client.report [None req-b1f62379-e574-4175-a099-f938af5ddaa8 tempest-ServerPasswordTestJSON-1543821420 tempest-ServerPasswordTestJSON-1543821420-project-member] Deleted allocations for instance 49e99297-ffb9-4104-bd06-911243908828 [ 700.072171] env[61957]: DEBUG nova.network.neutron [None req-dbfaab8a-0784-4441-ad61-79e56d9c3c60 tempest-FloatingIPsAssociationTestJSON-1786153562 tempest-FloatingIPsAssociationTestJSON-1786153562-project-member] [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] Successfully created port: b04626cb-e752-4764-b985-7ae70cc01fd7 {{(pid=61957) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 700.108998] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c29f805b-b933-403c-bb7d-87fe6eab1913 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.116270] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bea913a-38e4-488e-954a-7286e4f2c3e8 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.153755] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45c96b55-18f9-49f2-a975-3d3adac54dc0 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.163353] env[61957]: DEBUG oslo_vmware.api [None req-2d3c8bae-3f1e-41b1-82a9-b293bdd893bd tempest-ServersAdmin275Test-2089451214 tempest-ServersAdmin275Test-2089451214-project-admin] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]5296d4a9-0176-1915-0c90-7fdb1b68bd3d, 'name': SearchDatastore_Task, 'duration_secs': 0.008074} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 700.166610] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f9879dc9-532f-472c-8d17-b1d6cd2e2b75 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.170272] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1ecad1c-a9e1-4925-a65e-e42623bd3cf8 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.187836] env[61957]: DEBUG nova.compute.provider_tree [None req-fc6b77c8-b5cd-4608-beab-d2cb5c9f7440 tempest-ImagesNegativeTestJSON-1548152881 tempest-ImagesNegativeTestJSON-1548152881-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 700.189610] env[61957]: DEBUG oslo_vmware.api [None req-2d3c8bae-3f1e-41b1-82a9-b293bdd893bd tempest-ServersAdmin275Test-2089451214 tempest-ServersAdmin275Test-2089451214-project-admin] Waiting for the task: (returnval){ [ 700.189610] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]529aa783-787d-3ca2-9219-c27c35f49537" [ 700.189610] env[61957]: _type = "Task" [ 700.189610] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 700.199324] env[61957]: DEBUG oslo_vmware.api [None req-2d3c8bae-3f1e-41b1-82a9-b293bdd893bd tempest-ServersAdmin275Test-2089451214 tempest-ServersAdmin275Test-2089451214-project-admin] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]529aa783-787d-3ca2-9219-c27c35f49537, 'name': SearchDatastore_Task, 'duration_secs': 0.008671} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 700.200653] env[61957]: DEBUG oslo_concurrency.lockutils [None req-2d3c8bae-3f1e-41b1-82a9-b293bdd893bd tempest-ServersAdmin275Test-2089451214 tempest-ServersAdmin275Test-2089451214-project-admin] Releasing lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 700.200653] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-2d3c8bae-3f1e-41b1-82a9-b293bdd893bd tempest-ServersAdmin275Test-2089451214 tempest-ServersAdmin275Test-2089451214-project-admin] Copying Virtual Disk [datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore2] 188b0fd3-5d71-4feb-aca5-75a2bd28895a/188b0fd3-5d71-4feb-aca5-75a2bd28895a.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 700.200653] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-46ab002c-0dff-44fb-9ad4-6b19ea5436ca {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.206661] env[61957]: DEBUG oslo_vmware.api [None req-2d3c8bae-3f1e-41b1-82a9-b293bdd893bd tempest-ServersAdmin275Test-2089451214 tempest-ServersAdmin275Test-2089451214-project-admin] Waiting for the task: (returnval){ [ 700.206661] env[61957]: value = "task-1277359" [ 700.206661] env[61957]: _type = "Task" [ 700.206661] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 700.215632] env[61957]: DEBUG oslo_vmware.api [None req-2d3c8bae-3f1e-41b1-82a9-b293bdd893bd tempest-ServersAdmin275Test-2089451214 tempest-ServersAdmin275Test-2089451214-project-admin] Task: {'id': task-1277359, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 700.217237] env[61957]: DEBUG nova.compute.manager [None req-dbfaab8a-0784-4441-ad61-79e56d9c3c60 tempest-FloatingIPsAssociationTestJSON-1786153562 tempest-FloatingIPsAssociationTestJSON-1786153562-project-member] [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] Start building block device mappings for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 700.497150] env[61957]: DEBUG oslo_concurrency.lockutils [None req-b1f62379-e574-4175-a099-f938af5ddaa8 tempest-ServerPasswordTestJSON-1543821420 tempest-ServerPasswordTestJSON-1543821420-project-member] Lock "49e99297-ffb9-4104-bd06-911243908828" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 102.978s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 700.691837] env[61957]: DEBUG nova.scheduler.client.report [None req-fc6b77c8-b5cd-4608-beab-d2cb5c9f7440 tempest-ImagesNegativeTestJSON-1548152881 tempest-ImagesNegativeTestJSON-1548152881-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 700.718574] env[61957]: DEBUG oslo_vmware.api [None req-2d3c8bae-3f1e-41b1-82a9-b293bdd893bd tempest-ServersAdmin275Test-2089451214 tempest-ServersAdmin275Test-2089451214-project-admin] Task: {'id': task-1277359, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.452149} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 700.718977] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-2d3c8bae-3f1e-41b1-82a9-b293bdd893bd tempest-ServersAdmin275Test-2089451214 tempest-ServersAdmin275Test-2089451214-project-admin] Copied Virtual Disk [datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore2] 188b0fd3-5d71-4feb-aca5-75a2bd28895a/188b0fd3-5d71-4feb-aca5-75a2bd28895a.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 700.719261] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-2d3c8bae-3f1e-41b1-82a9-b293bdd893bd tempest-ServersAdmin275Test-2089451214 tempest-ServersAdmin275Test-2089451214-project-admin] [instance: 188b0fd3-5d71-4feb-aca5-75a2bd28895a] Extending root virtual disk to 1048576 {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 700.719549] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-542ba19b-eeb2-4833-b81b-5040e9375980 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.729470] env[61957]: DEBUG oslo_vmware.api [None req-2d3c8bae-3f1e-41b1-82a9-b293bdd893bd tempest-ServersAdmin275Test-2089451214 tempest-ServersAdmin275Test-2089451214-project-admin] Waiting for the task: (returnval){ [ 700.729470] env[61957]: value = "task-1277360" [ 700.729470] env[61957]: _type = "Task" [ 700.729470] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 700.742612] env[61957]: DEBUG oslo_vmware.api [None req-2d3c8bae-3f1e-41b1-82a9-b293bdd893bd tempest-ServersAdmin275Test-2089451214 tempest-ServersAdmin275Test-2089451214-project-admin] Task: {'id': task-1277360, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 701.000102] env[61957]: DEBUG nova.compute.manager [None req-707bfa11-3857-45e5-a4be-581ffade0e04 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] Starting instance... {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 701.017294] env[61957]: DEBUG nova.compute.manager [req-9cb1c589-e370-4490-950b-049df041558f req-d88ef57a-4b31-4f71-b7b4-4626b0f61e8d service nova] [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] Received event network-changed-b04626cb-e752-4764-b985-7ae70cc01fd7 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 701.017454] env[61957]: DEBUG nova.compute.manager [req-9cb1c589-e370-4490-950b-049df041558f req-d88ef57a-4b31-4f71-b7b4-4626b0f61e8d service nova] [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] Refreshing instance network info cache due to event network-changed-b04626cb-e752-4764-b985-7ae70cc01fd7. {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 701.017665] env[61957]: DEBUG oslo_concurrency.lockutils [req-9cb1c589-e370-4490-950b-049df041558f req-d88ef57a-4b31-4f71-b7b4-4626b0f61e8d service nova] Acquiring lock "refresh_cache-8257a521-2c25-45a2-a2e5-c735ece03da2" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 701.017805] env[61957]: DEBUG oslo_concurrency.lockutils [req-9cb1c589-e370-4490-950b-049df041558f req-d88ef57a-4b31-4f71-b7b4-4626b0f61e8d service nova] Acquired lock "refresh_cache-8257a521-2c25-45a2-a2e5-c735ece03da2" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 701.017959] env[61957]: DEBUG nova.network.neutron [req-9cb1c589-e370-4490-950b-049df041558f req-d88ef57a-4b31-4f71-b7b4-4626b0f61e8d service nova] [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] Refreshing network info cache for port b04626cb-e752-4764-b985-7ae70cc01fd7 {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 701.150451] env[61957]: ERROR nova.compute.manager [None req-dbfaab8a-0784-4441-ad61-79e56d9c3c60 tempest-FloatingIPsAssociationTestJSON-1786153562 tempest-FloatingIPsAssociationTestJSON-1786153562-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port b04626cb-e752-4764-b985-7ae70cc01fd7, please check neutron logs for more information. [ 701.150451] env[61957]: ERROR nova.compute.manager Traceback (most recent call last): [ 701.150451] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 701.150451] env[61957]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 701.150451] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 701.150451] env[61957]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 701.150451] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 701.150451] env[61957]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 701.150451] env[61957]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 701.150451] env[61957]: ERROR nova.compute.manager self.force_reraise() [ 701.150451] env[61957]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 701.150451] env[61957]: ERROR nova.compute.manager raise self.value [ 701.150451] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 701.150451] env[61957]: ERROR nova.compute.manager updated_port = self._update_port( [ 701.150451] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 701.150451] env[61957]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 701.150937] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 701.150937] env[61957]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 701.150937] env[61957]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port b04626cb-e752-4764-b985-7ae70cc01fd7, please check neutron logs for more information. [ 701.150937] env[61957]: ERROR nova.compute.manager [ 701.150937] env[61957]: Traceback (most recent call last): [ 701.150937] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 701.150937] env[61957]: listener.cb(fileno) [ 701.150937] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 701.150937] env[61957]: result = function(*args, **kwargs) [ 701.150937] env[61957]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 701.150937] env[61957]: return func(*args, **kwargs) [ 701.150937] env[61957]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 701.150937] env[61957]: raise e [ 701.150937] env[61957]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 701.150937] env[61957]: nwinfo = self.network_api.allocate_for_instance( [ 701.150937] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 701.150937] env[61957]: created_port_ids = self._update_ports_for_instance( [ 701.150937] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 701.150937] env[61957]: with excutils.save_and_reraise_exception(): [ 701.150937] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 701.150937] env[61957]: self.force_reraise() [ 701.150937] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 701.150937] env[61957]: raise self.value [ 701.150937] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 701.150937] env[61957]: updated_port = self._update_port( [ 701.150937] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 701.150937] env[61957]: _ensure_no_port_binding_failure(port) [ 701.150937] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 701.150937] env[61957]: raise exception.PortBindingFailed(port_id=port['id']) [ 701.151777] env[61957]: nova.exception.PortBindingFailed: Binding failed for port b04626cb-e752-4764-b985-7ae70cc01fd7, please check neutron logs for more information. [ 701.151777] env[61957]: Removing descriptor: 16 [ 701.197507] env[61957]: DEBUG oslo_concurrency.lockutils [None req-fc6b77c8-b5cd-4608-beab-d2cb5c9f7440 tempest-ImagesNegativeTestJSON-1548152881 tempest-ImagesNegativeTestJSON-1548152881-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.992s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 701.198022] env[61957]: ERROR nova.compute.manager [None req-fc6b77c8-b5cd-4608-beab-d2cb5c9f7440 tempest-ImagesNegativeTestJSON-1548152881 tempest-ImagesNegativeTestJSON-1548152881-project-member] [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 7ebf5e95-b147-40b2-9271-c193cd89e949, please check neutron logs for more information. [ 701.198022] env[61957]: ERROR nova.compute.manager [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] Traceback (most recent call last): [ 701.198022] env[61957]: ERROR nova.compute.manager [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 701.198022] env[61957]: ERROR nova.compute.manager [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] self.driver.spawn(context, instance, image_meta, [ 701.198022] env[61957]: ERROR nova.compute.manager [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 701.198022] env[61957]: ERROR nova.compute.manager [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] self._vmops.spawn(context, instance, image_meta, injected_files, [ 701.198022] env[61957]: ERROR nova.compute.manager [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 701.198022] env[61957]: ERROR nova.compute.manager [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] vm_ref = self.build_virtual_machine(instance, [ 701.198022] env[61957]: ERROR nova.compute.manager [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 701.198022] env[61957]: ERROR nova.compute.manager [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] vif_infos = vmwarevif.get_vif_info(self._session, [ 701.198022] env[61957]: ERROR nova.compute.manager [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 701.198410] env[61957]: ERROR nova.compute.manager [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] for vif in network_info: [ 701.198410] env[61957]: ERROR nova.compute.manager [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 701.198410] env[61957]: ERROR nova.compute.manager [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] return self._sync_wrapper(fn, *args, **kwargs) [ 701.198410] env[61957]: ERROR nova.compute.manager [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 701.198410] env[61957]: ERROR nova.compute.manager [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] self.wait() [ 701.198410] env[61957]: ERROR nova.compute.manager [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 701.198410] env[61957]: ERROR nova.compute.manager [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] self[:] = self._gt.wait() [ 701.198410] env[61957]: ERROR nova.compute.manager [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 701.198410] env[61957]: ERROR nova.compute.manager [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] return self._exit_event.wait() [ 701.198410] env[61957]: ERROR nova.compute.manager [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 701.198410] env[61957]: ERROR nova.compute.manager [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] current.throw(*self._exc) [ 701.198410] env[61957]: ERROR nova.compute.manager [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 701.198410] env[61957]: ERROR nova.compute.manager [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] result = function(*args, **kwargs) [ 701.198748] env[61957]: ERROR nova.compute.manager [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 701.198748] env[61957]: ERROR nova.compute.manager [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] return func(*args, **kwargs) [ 701.198748] env[61957]: ERROR nova.compute.manager [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 701.198748] env[61957]: ERROR nova.compute.manager [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] raise e [ 701.198748] env[61957]: ERROR nova.compute.manager [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 701.198748] env[61957]: ERROR nova.compute.manager [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] nwinfo = self.network_api.allocate_for_instance( [ 701.198748] env[61957]: ERROR nova.compute.manager [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 701.198748] env[61957]: ERROR nova.compute.manager [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] created_port_ids = self._update_ports_for_instance( [ 701.198748] env[61957]: ERROR nova.compute.manager [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 701.198748] env[61957]: ERROR nova.compute.manager [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] with excutils.save_and_reraise_exception(): [ 701.198748] env[61957]: ERROR nova.compute.manager [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 701.198748] env[61957]: ERROR nova.compute.manager [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] self.force_reraise() [ 701.198748] env[61957]: ERROR nova.compute.manager [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 701.199099] env[61957]: ERROR nova.compute.manager [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] raise self.value [ 701.199099] env[61957]: ERROR nova.compute.manager [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 701.199099] env[61957]: ERROR nova.compute.manager [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] updated_port = self._update_port( [ 701.199099] env[61957]: ERROR nova.compute.manager [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 701.199099] env[61957]: ERROR nova.compute.manager [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] _ensure_no_port_binding_failure(port) [ 701.199099] env[61957]: ERROR nova.compute.manager [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 701.199099] env[61957]: ERROR nova.compute.manager [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] raise exception.PortBindingFailed(port_id=port['id']) [ 701.199099] env[61957]: ERROR nova.compute.manager [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] nova.exception.PortBindingFailed: Binding failed for port 7ebf5e95-b147-40b2-9271-c193cd89e949, please check neutron logs for more information. [ 701.199099] env[61957]: ERROR nova.compute.manager [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] [ 701.199099] env[61957]: DEBUG nova.compute.utils [None req-fc6b77c8-b5cd-4608-beab-d2cb5c9f7440 tempest-ImagesNegativeTestJSON-1548152881 tempest-ImagesNegativeTestJSON-1548152881-project-member] [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] Binding failed for port 7ebf5e95-b147-40b2-9271-c193cd89e949, please check neutron logs for more information. {{(pid=61957) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 701.200061] env[61957]: DEBUG oslo_concurrency.lockutils [None req-c2f4fc75-b8bd-417b-8703-3b9048472a92 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.176s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 701.201550] env[61957]: INFO nova.compute.claims [None req-c2f4fc75-b8bd-417b-8703-3b9048472a92 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 701.204167] env[61957]: DEBUG nova.compute.manager [None req-fc6b77c8-b5cd-4608-beab-d2cb5c9f7440 tempest-ImagesNegativeTestJSON-1548152881 tempest-ImagesNegativeTestJSON-1548152881-project-member] [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] Build of instance 9df88f23-0219-43e0-b28a-e78f30a473a7 was re-scheduled: Binding failed for port 7ebf5e95-b147-40b2-9271-c193cd89e949, please check neutron logs for more information. {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 701.204599] env[61957]: DEBUG nova.compute.manager [None req-fc6b77c8-b5cd-4608-beab-d2cb5c9f7440 tempest-ImagesNegativeTestJSON-1548152881 tempest-ImagesNegativeTestJSON-1548152881-project-member] [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] Unplugging VIFs for instance {{(pid=61957) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 701.204817] env[61957]: DEBUG oslo_concurrency.lockutils [None req-fc6b77c8-b5cd-4608-beab-d2cb5c9f7440 tempest-ImagesNegativeTestJSON-1548152881 tempest-ImagesNegativeTestJSON-1548152881-project-member] Acquiring lock "refresh_cache-9df88f23-0219-43e0-b28a-e78f30a473a7" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 701.204980] env[61957]: DEBUG oslo_concurrency.lockutils [None req-fc6b77c8-b5cd-4608-beab-d2cb5c9f7440 tempest-ImagesNegativeTestJSON-1548152881 tempest-ImagesNegativeTestJSON-1548152881-project-member] Acquired lock "refresh_cache-9df88f23-0219-43e0-b28a-e78f30a473a7" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 701.205148] env[61957]: DEBUG nova.network.neutron [None req-fc6b77c8-b5cd-4608-beab-d2cb5c9f7440 tempest-ImagesNegativeTestJSON-1548152881 tempest-ImagesNegativeTestJSON-1548152881-project-member] [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 701.228175] env[61957]: DEBUG nova.compute.manager [None req-dbfaab8a-0784-4441-ad61-79e56d9c3c60 tempest-FloatingIPsAssociationTestJSON-1786153562 tempest-FloatingIPsAssociationTestJSON-1786153562-project-member] [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] Start spawning the instance on the hypervisor. {{(pid=61957) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 701.240180] env[61957]: DEBUG oslo_vmware.api [None req-2d3c8bae-3f1e-41b1-82a9-b293bdd893bd tempest-ServersAdmin275Test-2089451214 tempest-ServersAdmin275Test-2089451214-project-admin] Task: {'id': task-1277360, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.062002} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 701.240180] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-2d3c8bae-3f1e-41b1-82a9-b293bdd893bd tempest-ServersAdmin275Test-2089451214 tempest-ServersAdmin275Test-2089451214-project-admin] [instance: 188b0fd3-5d71-4feb-aca5-75a2bd28895a] Extended root virtual disk {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 701.241043] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c69bace-05f7-4f76-a367-20f0d70a548d {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.263302] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-2d3c8bae-3f1e-41b1-82a9-b293bdd893bd tempest-ServersAdmin275Test-2089451214 tempest-ServersAdmin275Test-2089451214-project-admin] [instance: 188b0fd3-5d71-4feb-aca5-75a2bd28895a] Reconfiguring VM instance instance-0000001b to attach disk [datastore2] 188b0fd3-5d71-4feb-aca5-75a2bd28895a/188b0fd3-5d71-4feb-aca5-75a2bd28895a.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 701.265933] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bfe63186-bc59-4009-8766-4d1a320f31a3 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.292183] env[61957]: DEBUG nova.virt.hardware [None req-dbfaab8a-0784-4441-ad61-79e56d9c3c60 tempest-FloatingIPsAssociationTestJSON-1786153562 tempest-FloatingIPsAssociationTestJSON-1786153562-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T17:22:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T17:21:50Z,direct_url=,disk_format='vmdk',id=11c76a2c-f705-470a-ba9d-4657858bab38,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='81d3096d0f094373913fc3dc8f5c3c6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T17:21:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 701.292617] env[61957]: DEBUG nova.virt.hardware [None req-dbfaab8a-0784-4441-ad61-79e56d9c3c60 tempest-FloatingIPsAssociationTestJSON-1786153562 tempest-FloatingIPsAssociationTestJSON-1786153562-project-member] Flavor limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 701.292617] env[61957]: DEBUG nova.virt.hardware [None req-dbfaab8a-0784-4441-ad61-79e56d9c3c60 tempest-FloatingIPsAssociationTestJSON-1786153562 tempest-FloatingIPsAssociationTestJSON-1786153562-project-member] Image limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 701.292761] env[61957]: DEBUG nova.virt.hardware [None req-dbfaab8a-0784-4441-ad61-79e56d9c3c60 tempest-FloatingIPsAssociationTestJSON-1786153562 tempest-FloatingIPsAssociationTestJSON-1786153562-project-member] Flavor pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 701.292901] env[61957]: DEBUG nova.virt.hardware [None req-dbfaab8a-0784-4441-ad61-79e56d9c3c60 tempest-FloatingIPsAssociationTestJSON-1786153562 tempest-FloatingIPsAssociationTestJSON-1786153562-project-member] Image pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 701.293099] env[61957]: DEBUG nova.virt.hardware [None req-dbfaab8a-0784-4441-ad61-79e56d9c3c60 tempest-FloatingIPsAssociationTestJSON-1786153562 tempest-FloatingIPsAssociationTestJSON-1786153562-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 701.293272] env[61957]: DEBUG nova.virt.hardware [None req-dbfaab8a-0784-4441-ad61-79e56d9c3c60 tempest-FloatingIPsAssociationTestJSON-1786153562 tempest-FloatingIPsAssociationTestJSON-1786153562-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 701.293431] env[61957]: DEBUG nova.virt.hardware [None req-dbfaab8a-0784-4441-ad61-79e56d9c3c60 tempest-FloatingIPsAssociationTestJSON-1786153562 tempest-FloatingIPsAssociationTestJSON-1786153562-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 701.293593] env[61957]: DEBUG nova.virt.hardware [None req-dbfaab8a-0784-4441-ad61-79e56d9c3c60 tempest-FloatingIPsAssociationTestJSON-1786153562 tempest-FloatingIPsAssociationTestJSON-1786153562-project-member] Got 1 possible topologies {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 701.293750] env[61957]: DEBUG nova.virt.hardware [None req-dbfaab8a-0784-4441-ad61-79e56d9c3c60 tempest-FloatingIPsAssociationTestJSON-1786153562 tempest-FloatingIPsAssociationTestJSON-1786153562-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 701.293918] env[61957]: DEBUG nova.virt.hardware [None req-dbfaab8a-0784-4441-ad61-79e56d9c3c60 tempest-FloatingIPsAssociationTestJSON-1786153562 tempest-FloatingIPsAssociationTestJSON-1786153562-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 701.297110] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd014bda-958a-45ac-b6fc-ee0ac396d777 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.300151] env[61957]: DEBUG oslo_vmware.api [None req-2d3c8bae-3f1e-41b1-82a9-b293bdd893bd tempest-ServersAdmin275Test-2089451214 tempest-ServersAdmin275Test-2089451214-project-admin] Waiting for the task: (returnval){ [ 701.300151] env[61957]: value = "task-1277361" [ 701.300151] env[61957]: _type = "Task" [ 701.300151] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 701.308212] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f71e13f5-02b5-44c0-90a0-65a3c2494352 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.314509] env[61957]: DEBUG oslo_vmware.api [None req-2d3c8bae-3f1e-41b1-82a9-b293bdd893bd tempest-ServersAdmin275Test-2089451214 tempest-ServersAdmin275Test-2089451214-project-admin] Task: {'id': task-1277361, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 701.324527] env[61957]: ERROR nova.compute.manager [None req-dbfaab8a-0784-4441-ad61-79e56d9c3c60 tempest-FloatingIPsAssociationTestJSON-1786153562 tempest-FloatingIPsAssociationTestJSON-1786153562-project-member] [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port b04626cb-e752-4764-b985-7ae70cc01fd7, please check neutron logs for more information. [ 701.324527] env[61957]: ERROR nova.compute.manager [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] Traceback (most recent call last): [ 701.324527] env[61957]: ERROR nova.compute.manager [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 701.324527] env[61957]: ERROR nova.compute.manager [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] yield resources [ 701.324527] env[61957]: ERROR nova.compute.manager [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 701.324527] env[61957]: ERROR nova.compute.manager [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] self.driver.spawn(context, instance, image_meta, [ 701.324527] env[61957]: ERROR nova.compute.manager [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 701.324527] env[61957]: ERROR nova.compute.manager [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] self._vmops.spawn(context, instance, image_meta, injected_files, [ 701.324527] env[61957]: ERROR nova.compute.manager [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 701.324527] env[61957]: ERROR nova.compute.manager [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] vm_ref = self.build_virtual_machine(instance, [ 701.324527] env[61957]: ERROR nova.compute.manager [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 701.324863] env[61957]: ERROR nova.compute.manager [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] vif_infos = vmwarevif.get_vif_info(self._session, [ 701.324863] env[61957]: ERROR nova.compute.manager [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 701.324863] env[61957]: ERROR nova.compute.manager [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] for vif in network_info: [ 701.324863] env[61957]: ERROR nova.compute.manager [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 701.324863] env[61957]: ERROR nova.compute.manager [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] return self._sync_wrapper(fn, *args, **kwargs) [ 701.324863] env[61957]: ERROR nova.compute.manager [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 701.324863] env[61957]: ERROR nova.compute.manager [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] self.wait() [ 701.324863] env[61957]: ERROR nova.compute.manager [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 701.324863] env[61957]: ERROR nova.compute.manager [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] self[:] = self._gt.wait() [ 701.324863] env[61957]: ERROR nova.compute.manager [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 701.324863] env[61957]: ERROR nova.compute.manager [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] return self._exit_event.wait() [ 701.324863] env[61957]: ERROR nova.compute.manager [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 701.324863] env[61957]: ERROR nova.compute.manager [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] current.throw(*self._exc) [ 701.325217] env[61957]: ERROR nova.compute.manager [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 701.325217] env[61957]: ERROR nova.compute.manager [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] result = function(*args, **kwargs) [ 701.325217] env[61957]: ERROR nova.compute.manager [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 701.325217] env[61957]: ERROR nova.compute.manager [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] return func(*args, **kwargs) [ 701.325217] env[61957]: ERROR nova.compute.manager [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 701.325217] env[61957]: ERROR nova.compute.manager [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] raise e [ 701.325217] env[61957]: ERROR nova.compute.manager [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 701.325217] env[61957]: ERROR nova.compute.manager [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] nwinfo = self.network_api.allocate_for_instance( [ 701.325217] env[61957]: ERROR nova.compute.manager [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 701.325217] env[61957]: ERROR nova.compute.manager [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] created_port_ids = self._update_ports_for_instance( [ 701.325217] env[61957]: ERROR nova.compute.manager [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 701.325217] env[61957]: ERROR nova.compute.manager [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] with excutils.save_and_reraise_exception(): [ 701.325217] env[61957]: ERROR nova.compute.manager [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 701.325563] env[61957]: ERROR nova.compute.manager [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] self.force_reraise() [ 701.325563] env[61957]: ERROR nova.compute.manager [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 701.325563] env[61957]: ERROR nova.compute.manager [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] raise self.value [ 701.325563] env[61957]: ERROR nova.compute.manager [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 701.325563] env[61957]: ERROR nova.compute.manager [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] updated_port = self._update_port( [ 701.325563] env[61957]: ERROR nova.compute.manager [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 701.325563] env[61957]: ERROR nova.compute.manager [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] _ensure_no_port_binding_failure(port) [ 701.325563] env[61957]: ERROR nova.compute.manager [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 701.325563] env[61957]: ERROR nova.compute.manager [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] raise exception.PortBindingFailed(port_id=port['id']) [ 701.325563] env[61957]: ERROR nova.compute.manager [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] nova.exception.PortBindingFailed: Binding failed for port b04626cb-e752-4764-b985-7ae70cc01fd7, please check neutron logs for more information. [ 701.325563] env[61957]: ERROR nova.compute.manager [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] [ 701.325563] env[61957]: INFO nova.compute.manager [None req-dbfaab8a-0784-4441-ad61-79e56d9c3c60 tempest-FloatingIPsAssociationTestJSON-1786153562 tempest-FloatingIPsAssociationTestJSON-1786153562-project-member] [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] Terminating instance [ 701.326790] env[61957]: DEBUG oslo_concurrency.lockutils [None req-dbfaab8a-0784-4441-ad61-79e56d9c3c60 tempest-FloatingIPsAssociationTestJSON-1786153562 tempest-FloatingIPsAssociationTestJSON-1786153562-project-member] Acquiring lock "refresh_cache-8257a521-2c25-45a2-a2e5-c735ece03da2" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 701.524167] env[61957]: DEBUG oslo_concurrency.lockutils [None req-707bfa11-3857-45e5-a4be-581ffade0e04 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 701.536446] env[61957]: DEBUG nova.network.neutron [req-9cb1c589-e370-4490-950b-049df041558f req-d88ef57a-4b31-4f71-b7b4-4626b0f61e8d service nova] [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 701.612898] env[61957]: DEBUG nova.network.neutron [req-9cb1c589-e370-4490-950b-049df041558f req-d88ef57a-4b31-4f71-b7b4-4626b0f61e8d service nova] [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 701.725645] env[61957]: DEBUG nova.network.neutron [None req-fc6b77c8-b5cd-4608-beab-d2cb5c9f7440 tempest-ImagesNegativeTestJSON-1548152881 tempest-ImagesNegativeTestJSON-1548152881-project-member] [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 701.800417] env[61957]: DEBUG nova.network.neutron [None req-fc6b77c8-b5cd-4608-beab-d2cb5c9f7440 tempest-ImagesNegativeTestJSON-1548152881 tempest-ImagesNegativeTestJSON-1548152881-project-member] [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 701.810953] env[61957]: DEBUG oslo_vmware.api [None req-2d3c8bae-3f1e-41b1-82a9-b293bdd893bd tempest-ServersAdmin275Test-2089451214 tempest-ServersAdmin275Test-2089451214-project-admin] Task: {'id': task-1277361, 'name': ReconfigVM_Task, 'duration_secs': 0.249881} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 701.811801] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-2d3c8bae-3f1e-41b1-82a9-b293bdd893bd tempest-ServersAdmin275Test-2089451214 tempest-ServersAdmin275Test-2089451214-project-admin] [instance: 188b0fd3-5d71-4feb-aca5-75a2bd28895a] Reconfigured VM instance instance-0000001b to attach disk [datastore2] 188b0fd3-5d71-4feb-aca5-75a2bd28895a/188b0fd3-5d71-4feb-aca5-75a2bd28895a.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 701.812423] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5608c900-c30a-4d6a-b5bb-b0a1c4f4667c {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.818794] env[61957]: DEBUG oslo_vmware.api [None req-2d3c8bae-3f1e-41b1-82a9-b293bdd893bd tempest-ServersAdmin275Test-2089451214 tempest-ServersAdmin275Test-2089451214-project-admin] Waiting for the task: (returnval){ [ 701.818794] env[61957]: value = "task-1277362" [ 701.818794] env[61957]: _type = "Task" [ 701.818794] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 701.826948] env[61957]: DEBUG oslo_vmware.api [None req-2d3c8bae-3f1e-41b1-82a9-b293bdd893bd tempest-ServersAdmin275Test-2089451214 tempest-ServersAdmin275Test-2089451214-project-admin] Task: {'id': task-1277362, 'name': Rename_Task} progress is 5%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 702.117313] env[61957]: DEBUG oslo_concurrency.lockutils [req-9cb1c589-e370-4490-950b-049df041558f req-d88ef57a-4b31-4f71-b7b4-4626b0f61e8d service nova] Releasing lock "refresh_cache-8257a521-2c25-45a2-a2e5-c735ece03da2" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 702.117313] env[61957]: DEBUG oslo_concurrency.lockutils [None req-dbfaab8a-0784-4441-ad61-79e56d9c3c60 tempest-FloatingIPsAssociationTestJSON-1786153562 tempest-FloatingIPsAssociationTestJSON-1786153562-project-member] Acquired lock "refresh_cache-8257a521-2c25-45a2-a2e5-c735ece03da2" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 702.117313] env[61957]: DEBUG nova.network.neutron [None req-dbfaab8a-0784-4441-ad61-79e56d9c3c60 tempest-FloatingIPsAssociationTestJSON-1786153562 tempest-FloatingIPsAssociationTestJSON-1786153562-project-member] [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 702.307099] env[61957]: DEBUG oslo_concurrency.lockutils [None req-fc6b77c8-b5cd-4608-beab-d2cb5c9f7440 tempest-ImagesNegativeTestJSON-1548152881 tempest-ImagesNegativeTestJSON-1548152881-project-member] Releasing lock "refresh_cache-9df88f23-0219-43e0-b28a-e78f30a473a7" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 702.307609] env[61957]: DEBUG nova.compute.manager [None req-fc6b77c8-b5cd-4608-beab-d2cb5c9f7440 tempest-ImagesNegativeTestJSON-1548152881 tempest-ImagesNegativeTestJSON-1548152881-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61957) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 702.307609] env[61957]: DEBUG nova.compute.manager [None req-fc6b77c8-b5cd-4608-beab-d2cb5c9f7440 tempest-ImagesNegativeTestJSON-1548152881 tempest-ImagesNegativeTestJSON-1548152881-project-member] [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 702.307759] env[61957]: DEBUG nova.network.neutron [None req-fc6b77c8-b5cd-4608-beab-d2cb5c9f7440 tempest-ImagesNegativeTestJSON-1548152881 tempest-ImagesNegativeTestJSON-1548152881-project-member] [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 702.327398] env[61957]: DEBUG nova.network.neutron [None req-fc6b77c8-b5cd-4608-beab-d2cb5c9f7440 tempest-ImagesNegativeTestJSON-1548152881 tempest-ImagesNegativeTestJSON-1548152881-project-member] [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 702.336320] env[61957]: DEBUG oslo_vmware.api [None req-2d3c8bae-3f1e-41b1-82a9-b293bdd893bd tempest-ServersAdmin275Test-2089451214 tempest-ServersAdmin275Test-2089451214-project-admin] Task: {'id': task-1277362, 'name': Rename_Task, 'duration_secs': 0.128931} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 702.337492] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-2d3c8bae-3f1e-41b1-82a9-b293bdd893bd tempest-ServersAdmin275Test-2089451214 tempest-ServersAdmin275Test-2089451214-project-admin] [instance: 188b0fd3-5d71-4feb-aca5-75a2bd28895a] Powering on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 702.337770] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c173eef7-10c2-46c1-8d50-4c922c95844d {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.345103] env[61957]: DEBUG oslo_vmware.api [None req-2d3c8bae-3f1e-41b1-82a9-b293bdd893bd tempest-ServersAdmin275Test-2089451214 tempest-ServersAdmin275Test-2089451214-project-admin] Waiting for the task: (returnval){ [ 702.345103] env[61957]: value = "task-1277363" [ 702.345103] env[61957]: _type = "Task" [ 702.345103] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 702.353524] env[61957]: DEBUG oslo_vmware.api [None req-2d3c8bae-3f1e-41b1-82a9-b293bdd893bd tempest-ServersAdmin275Test-2089451214 tempest-ServersAdmin275Test-2089451214-project-admin] Task: {'id': task-1277363, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 702.610395] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85f73681-c6e1-487d-9c60-f25c09879dcf {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.618088] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c060237-bbbe-4702-af67-a789a127bf75 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.652486] env[61957]: DEBUG nova.network.neutron [None req-dbfaab8a-0784-4441-ad61-79e56d9c3c60 tempest-FloatingIPsAssociationTestJSON-1786153562 tempest-FloatingIPsAssociationTestJSON-1786153562-project-member] [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 702.652486] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07d18d0d-1191-4189-968e-894ccf7e5c18 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.660859] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1da584f6-0825-47d8-b4d3-e569359b64aa {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.675033] env[61957]: DEBUG nova.compute.provider_tree [None req-c2f4fc75-b8bd-417b-8703-3b9048472a92 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 702.770907] env[61957]: DEBUG nova.network.neutron [None req-dbfaab8a-0784-4441-ad61-79e56d9c3c60 tempest-FloatingIPsAssociationTestJSON-1786153562 tempest-FloatingIPsAssociationTestJSON-1786153562-project-member] [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 702.832338] env[61957]: DEBUG nova.network.neutron [None req-fc6b77c8-b5cd-4608-beab-d2cb5c9f7440 tempest-ImagesNegativeTestJSON-1548152881 tempest-ImagesNegativeTestJSON-1548152881-project-member] [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 702.855120] env[61957]: DEBUG oslo_vmware.api [None req-2d3c8bae-3f1e-41b1-82a9-b293bdd893bd tempest-ServersAdmin275Test-2089451214 tempest-ServersAdmin275Test-2089451214-project-admin] Task: {'id': task-1277363, 'name': PowerOnVM_Task, 'duration_secs': 0.408784} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 702.855120] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-2d3c8bae-3f1e-41b1-82a9-b293bdd893bd tempest-ServersAdmin275Test-2089451214 tempest-ServersAdmin275Test-2089451214-project-admin] [instance: 188b0fd3-5d71-4feb-aca5-75a2bd28895a] Powered on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 702.855120] env[61957]: DEBUG nova.compute.manager [None req-2d3c8bae-3f1e-41b1-82a9-b293bdd893bd tempest-ServersAdmin275Test-2089451214 tempest-ServersAdmin275Test-2089451214-project-admin] [instance: 188b0fd3-5d71-4feb-aca5-75a2bd28895a] Checking state {{(pid=61957) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 702.855635] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edb0c30d-64a5-445e-bbfb-6e512e3267f3 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.039458] env[61957]: DEBUG nova.compute.manager [req-c31e2cb2-9276-4aa5-8c16-fe57808cd02e req-babae55d-685e-43aa-a0f0-1af04140b3cc service nova] [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] Received event network-vif-deleted-b04626cb-e752-4764-b985-7ae70cc01fd7 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 703.177966] env[61957]: DEBUG nova.scheduler.client.report [None req-c2f4fc75-b8bd-417b-8703-3b9048472a92 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 703.273708] env[61957]: DEBUG oslo_concurrency.lockutils [None req-dbfaab8a-0784-4441-ad61-79e56d9c3c60 tempest-FloatingIPsAssociationTestJSON-1786153562 tempest-FloatingIPsAssociationTestJSON-1786153562-project-member] Releasing lock "refresh_cache-8257a521-2c25-45a2-a2e5-c735ece03da2" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 703.274147] env[61957]: DEBUG nova.compute.manager [None req-dbfaab8a-0784-4441-ad61-79e56d9c3c60 tempest-FloatingIPsAssociationTestJSON-1786153562 tempest-FloatingIPsAssociationTestJSON-1786153562-project-member] [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] Start destroying the instance on the hypervisor. {{(pid=61957) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 703.274341] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-dbfaab8a-0784-4441-ad61-79e56d9c3c60 tempest-FloatingIPsAssociationTestJSON-1786153562 tempest-FloatingIPsAssociationTestJSON-1786153562-project-member] [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] Destroying instance {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 703.274627] env[61957]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8d3ed741-86a2-477d-94dc-8f0362d03ccf {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.284994] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-534bcfda-8f53-4159-b8d6-11353e3512c7 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.308106] env[61957]: WARNING nova.virt.vmwareapi.vmops [None req-dbfaab8a-0784-4441-ad61-79e56d9c3c60 tempest-FloatingIPsAssociationTestJSON-1786153562 tempest-FloatingIPsAssociationTestJSON-1786153562-project-member] [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 8257a521-2c25-45a2-a2e5-c735ece03da2 could not be found. [ 703.308396] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-dbfaab8a-0784-4441-ad61-79e56d9c3c60 tempest-FloatingIPsAssociationTestJSON-1786153562 tempest-FloatingIPsAssociationTestJSON-1786153562-project-member] [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] Instance destroyed {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 703.308613] env[61957]: INFO nova.compute.manager [None req-dbfaab8a-0784-4441-ad61-79e56d9c3c60 tempest-FloatingIPsAssociationTestJSON-1786153562 tempest-FloatingIPsAssociationTestJSON-1786153562-project-member] [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] Took 0.03 seconds to destroy the instance on the hypervisor. [ 703.308886] env[61957]: DEBUG oslo.service.loopingcall [None req-dbfaab8a-0784-4441-ad61-79e56d9c3c60 tempest-FloatingIPsAssociationTestJSON-1786153562 tempest-FloatingIPsAssociationTestJSON-1786153562-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 703.309174] env[61957]: DEBUG nova.compute.manager [-] [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 703.309327] env[61957]: DEBUG nova.network.neutron [-] [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 703.325985] env[61957]: DEBUG nova.network.neutron [-] [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 703.334508] env[61957]: INFO nova.compute.manager [None req-fc6b77c8-b5cd-4608-beab-d2cb5c9f7440 tempest-ImagesNegativeTestJSON-1548152881 tempest-ImagesNegativeTestJSON-1548152881-project-member] [instance: 9df88f23-0219-43e0-b28a-e78f30a473a7] Took 1.03 seconds to deallocate network for instance. [ 703.370105] env[61957]: DEBUG oslo_concurrency.lockutils [None req-2d3c8bae-3f1e-41b1-82a9-b293bdd893bd tempest-ServersAdmin275Test-2089451214 tempest-ServersAdmin275Test-2089451214-project-admin] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 703.682466] env[61957]: DEBUG oslo_concurrency.lockutils [None req-c2f4fc75-b8bd-417b-8703-3b9048472a92 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.482s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 703.683128] env[61957]: DEBUG nova.compute.manager [None req-c2f4fc75-b8bd-417b-8703-3b9048472a92 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] Start building networks asynchronously for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 703.685981] env[61957]: DEBUG oslo_concurrency.lockutils [None req-3d56c405-76b3-4bdb-9787-ecf239aeeb0d tempest-ImagesOneServerNegativeTestJSON-1581220397 tempest-ImagesOneServerNegativeTestJSON-1581220397-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 29.902s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 703.828949] env[61957]: DEBUG nova.network.neutron [-] [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 704.027748] env[61957]: DEBUG oslo_concurrency.lockutils [None req-4e657274-3c21-45ad-b242-d1090e30b7c2 tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Acquiring lock "188b0fd3-5d71-4feb-aca5-75a2bd28895a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 704.027948] env[61957]: DEBUG oslo_concurrency.lockutils [None req-4e657274-3c21-45ad-b242-d1090e30b7c2 tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Lock "188b0fd3-5d71-4feb-aca5-75a2bd28895a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 704.028214] env[61957]: DEBUG oslo_concurrency.lockutils [None req-4e657274-3c21-45ad-b242-d1090e30b7c2 tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Acquiring lock "188b0fd3-5d71-4feb-aca5-75a2bd28895a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 704.028403] env[61957]: DEBUG oslo_concurrency.lockutils [None req-4e657274-3c21-45ad-b242-d1090e30b7c2 tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Lock "188b0fd3-5d71-4feb-aca5-75a2bd28895a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 704.028572] env[61957]: DEBUG oslo_concurrency.lockutils [None req-4e657274-3c21-45ad-b242-d1090e30b7c2 tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Lock "188b0fd3-5d71-4feb-aca5-75a2bd28895a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 704.030777] env[61957]: INFO nova.compute.manager [None req-4e657274-3c21-45ad-b242-d1090e30b7c2 tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] [instance: 188b0fd3-5d71-4feb-aca5-75a2bd28895a] Terminating instance [ 704.032399] env[61957]: DEBUG oslo_concurrency.lockutils [None req-4e657274-3c21-45ad-b242-d1090e30b7c2 tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Acquiring lock "refresh_cache-188b0fd3-5d71-4feb-aca5-75a2bd28895a" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 704.032559] env[61957]: DEBUG oslo_concurrency.lockutils [None req-4e657274-3c21-45ad-b242-d1090e30b7c2 tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Acquired lock "refresh_cache-188b0fd3-5d71-4feb-aca5-75a2bd28895a" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 704.032734] env[61957]: DEBUG nova.network.neutron [None req-4e657274-3c21-45ad-b242-d1090e30b7c2 tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] [instance: 188b0fd3-5d71-4feb-aca5-75a2bd28895a] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 704.190527] env[61957]: DEBUG nova.compute.utils [None req-c2f4fc75-b8bd-417b-8703-3b9048472a92 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] Using /dev/sd instead of None {{(pid=61957) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 704.195850] env[61957]: DEBUG nova.compute.manager [None req-c2f4fc75-b8bd-417b-8703-3b9048472a92 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] Allocating IP information in the background. {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 704.195850] env[61957]: DEBUG nova.network.neutron [None req-c2f4fc75-b8bd-417b-8703-3b9048472a92 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] allocate_for_instance() {{(pid=61957) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 704.248264] env[61957]: DEBUG nova.policy [None req-c2f4fc75-b8bd-417b-8703-3b9048472a92 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '02106e108faa446cad15bec05f2cc763', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c33de8fc6166438dabcfb820e0513085', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61957) authorize /opt/stack/nova/nova/policy.py:203}} [ 704.331035] env[61957]: INFO nova.compute.manager [-] [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] Took 1.02 seconds to deallocate network for instance. [ 704.333258] env[61957]: DEBUG nova.compute.claims [None req-dbfaab8a-0784-4441-ad61-79e56d9c3c60 tempest-FloatingIPsAssociationTestJSON-1786153562 tempest-FloatingIPsAssociationTestJSON-1786153562-project-member] [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] Aborting claim: {{(pid=61957) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 704.333435] env[61957]: DEBUG oslo_concurrency.lockutils [None req-dbfaab8a-0784-4441-ad61-79e56d9c3c60 tempest-FloatingIPsAssociationTestJSON-1786153562 tempest-FloatingIPsAssociationTestJSON-1786153562-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 704.360152] env[61957]: INFO nova.scheduler.client.report [None req-fc6b77c8-b5cd-4608-beab-d2cb5c9f7440 tempest-ImagesNegativeTestJSON-1548152881 tempest-ImagesNegativeTestJSON-1548152881-project-member] Deleted allocations for instance 9df88f23-0219-43e0-b28a-e78f30a473a7 [ 704.516996] env[61957]: DEBUG nova.network.neutron [None req-c2f4fc75-b8bd-417b-8703-3b9048472a92 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] Successfully created port: b8d0e37e-0f54-4ef7-90b0-fb5903f71e0f {{(pid=61957) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 704.552105] env[61957]: DEBUG nova.network.neutron [None req-4e657274-3c21-45ad-b242-d1090e30b7c2 tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] [instance: 188b0fd3-5d71-4feb-aca5-75a2bd28895a] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 704.587332] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c3213ce-18cf-4325-889c-d737d8c4acb2 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.595050] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1e7a8bc-c772-429c-b881-eeeeebce1255 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.627079] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c473075-bef0-4f6f-aec6-fabf678357e8 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.633776] env[61957]: DEBUG nova.network.neutron [None req-4e657274-3c21-45ad-b242-d1090e30b7c2 tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] [instance: 188b0fd3-5d71-4feb-aca5-75a2bd28895a] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 704.635871] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8e29f5e-05ac-4506-9a59-9b585e9732b2 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.651210] env[61957]: DEBUG nova.compute.provider_tree [None req-3d56c405-76b3-4bdb-9787-ecf239aeeb0d tempest-ImagesOneServerNegativeTestJSON-1581220397 tempest-ImagesOneServerNegativeTestJSON-1581220397-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 704.695755] env[61957]: DEBUG nova.compute.manager [None req-c2f4fc75-b8bd-417b-8703-3b9048472a92 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] Start building block device mappings for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 704.871269] env[61957]: DEBUG oslo_concurrency.lockutils [None req-fc6b77c8-b5cd-4608-beab-d2cb5c9f7440 tempest-ImagesNegativeTestJSON-1548152881 tempest-ImagesNegativeTestJSON-1548152881-project-member] Lock "9df88f23-0219-43e0-b28a-e78f30a473a7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 105.886s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 705.141982] env[61957]: DEBUG oslo_concurrency.lockutils [None req-4e657274-3c21-45ad-b242-d1090e30b7c2 tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Releasing lock "refresh_cache-188b0fd3-5d71-4feb-aca5-75a2bd28895a" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 705.142448] env[61957]: DEBUG nova.compute.manager [None req-4e657274-3c21-45ad-b242-d1090e30b7c2 tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] [instance: 188b0fd3-5d71-4feb-aca5-75a2bd28895a] Start destroying the instance on the hypervisor. {{(pid=61957) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 705.142648] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-4e657274-3c21-45ad-b242-d1090e30b7c2 tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] [instance: 188b0fd3-5d71-4feb-aca5-75a2bd28895a] Destroying instance {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 705.143552] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8988565c-fa3a-4120-bcdc-180a6bf7b1d9 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.155020] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e657274-3c21-45ad-b242-d1090e30b7c2 tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] [instance: 188b0fd3-5d71-4feb-aca5-75a2bd28895a] Powering off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 705.155020] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b844e013-e770-4e54-abfa-05cb633fe115 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.155020] env[61957]: DEBUG nova.scheduler.client.report [None req-3d56c405-76b3-4bdb-9787-ecf239aeeb0d tempest-ImagesOneServerNegativeTestJSON-1581220397 tempest-ImagesOneServerNegativeTestJSON-1581220397-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 705.165252] env[61957]: DEBUG oslo_vmware.api [None req-4e657274-3c21-45ad-b242-d1090e30b7c2 tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Waiting for the task: (returnval){ [ 705.165252] env[61957]: value = "task-1277364" [ 705.165252] env[61957]: _type = "Task" [ 705.165252] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 705.173840] env[61957]: DEBUG oslo_vmware.api [None req-4e657274-3c21-45ad-b242-d1090e30b7c2 tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Task: {'id': task-1277364, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 705.232618] env[61957]: DEBUG nova.compute.manager [req-1a0d4f54-bc59-495d-b8de-55df956951fe req-b1a663d6-cb2f-4ac4-bc05-a886002a713d service nova] [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] Received event network-changed-b8d0e37e-0f54-4ef7-90b0-fb5903f71e0f {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 705.232808] env[61957]: DEBUG nova.compute.manager [req-1a0d4f54-bc59-495d-b8de-55df956951fe req-b1a663d6-cb2f-4ac4-bc05-a886002a713d service nova] [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] Refreshing instance network info cache due to event network-changed-b8d0e37e-0f54-4ef7-90b0-fb5903f71e0f. {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 705.233024] env[61957]: DEBUG oslo_concurrency.lockutils [req-1a0d4f54-bc59-495d-b8de-55df956951fe req-b1a663d6-cb2f-4ac4-bc05-a886002a713d service nova] Acquiring lock "refresh_cache-d0bde8ec-da55-42b5-8d2c-9df2b90b88a7" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 705.233170] env[61957]: DEBUG oslo_concurrency.lockutils [req-1a0d4f54-bc59-495d-b8de-55df956951fe req-b1a663d6-cb2f-4ac4-bc05-a886002a713d service nova] Acquired lock "refresh_cache-d0bde8ec-da55-42b5-8d2c-9df2b90b88a7" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 705.233327] env[61957]: DEBUG nova.network.neutron [req-1a0d4f54-bc59-495d-b8de-55df956951fe req-b1a663d6-cb2f-4ac4-bc05-a886002a713d service nova] [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] Refreshing network info cache for port b8d0e37e-0f54-4ef7-90b0-fb5903f71e0f {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 705.372696] env[61957]: DEBUG nova.compute.manager [None req-2b9492ff-1cb8-4ba9-ba5e-b513c4b108ad tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] Starting instance... {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 705.381127] env[61957]: ERROR nova.compute.manager [None req-c2f4fc75-b8bd-417b-8703-3b9048472a92 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port b8d0e37e-0f54-4ef7-90b0-fb5903f71e0f, please check neutron logs for more information. [ 705.381127] env[61957]: ERROR nova.compute.manager Traceback (most recent call last): [ 705.381127] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 705.381127] env[61957]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 705.381127] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 705.381127] env[61957]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 705.381127] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 705.381127] env[61957]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 705.381127] env[61957]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 705.381127] env[61957]: ERROR nova.compute.manager self.force_reraise() [ 705.381127] env[61957]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 705.381127] env[61957]: ERROR nova.compute.manager raise self.value [ 705.381127] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 705.381127] env[61957]: ERROR nova.compute.manager updated_port = self._update_port( [ 705.381127] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 705.381127] env[61957]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 705.381706] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 705.381706] env[61957]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 705.381706] env[61957]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port b8d0e37e-0f54-4ef7-90b0-fb5903f71e0f, please check neutron logs for more information. [ 705.381706] env[61957]: ERROR nova.compute.manager [ 705.381706] env[61957]: Traceback (most recent call last): [ 705.381706] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 705.381706] env[61957]: listener.cb(fileno) [ 705.381706] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 705.381706] env[61957]: result = function(*args, **kwargs) [ 705.381706] env[61957]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 705.381706] env[61957]: return func(*args, **kwargs) [ 705.381706] env[61957]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 705.381706] env[61957]: raise e [ 705.381706] env[61957]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 705.381706] env[61957]: nwinfo = self.network_api.allocate_for_instance( [ 705.381706] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 705.381706] env[61957]: created_port_ids = self._update_ports_for_instance( [ 705.381706] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 705.381706] env[61957]: with excutils.save_and_reraise_exception(): [ 705.381706] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 705.381706] env[61957]: self.force_reraise() [ 705.381706] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 705.381706] env[61957]: raise self.value [ 705.381706] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 705.381706] env[61957]: updated_port = self._update_port( [ 705.381706] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 705.381706] env[61957]: _ensure_no_port_binding_failure(port) [ 705.381706] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 705.381706] env[61957]: raise exception.PortBindingFailed(port_id=port['id']) [ 705.382442] env[61957]: nova.exception.PortBindingFailed: Binding failed for port b8d0e37e-0f54-4ef7-90b0-fb5903f71e0f, please check neutron logs for more information. [ 705.382442] env[61957]: Removing descriptor: 16 [ 705.659604] env[61957]: DEBUG oslo_concurrency.lockutils [None req-3d56c405-76b3-4bdb-9787-ecf239aeeb0d tempest-ImagesOneServerNegativeTestJSON-1581220397 tempest-ImagesOneServerNegativeTestJSON-1581220397-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.974s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 705.660152] env[61957]: ERROR nova.compute.manager [None req-3d56c405-76b3-4bdb-9787-ecf239aeeb0d tempest-ImagesOneServerNegativeTestJSON-1581220397 tempest-ImagesOneServerNegativeTestJSON-1581220397-project-member] [instance: 2b151571-348f-4543-a0c5-afe6458b1973] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 892c3e98-da9b-4830-8c81-221309a32b88, please check neutron logs for more information. [ 705.660152] env[61957]: ERROR nova.compute.manager [instance: 2b151571-348f-4543-a0c5-afe6458b1973] Traceback (most recent call last): [ 705.660152] env[61957]: ERROR nova.compute.manager [instance: 2b151571-348f-4543-a0c5-afe6458b1973] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 705.660152] env[61957]: ERROR nova.compute.manager [instance: 2b151571-348f-4543-a0c5-afe6458b1973] self.driver.spawn(context, instance, image_meta, [ 705.660152] env[61957]: ERROR nova.compute.manager [instance: 2b151571-348f-4543-a0c5-afe6458b1973] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 705.660152] env[61957]: ERROR nova.compute.manager [instance: 2b151571-348f-4543-a0c5-afe6458b1973] self._vmops.spawn(context, instance, image_meta, injected_files, [ 705.660152] env[61957]: ERROR nova.compute.manager [instance: 2b151571-348f-4543-a0c5-afe6458b1973] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 705.660152] env[61957]: ERROR nova.compute.manager [instance: 2b151571-348f-4543-a0c5-afe6458b1973] vm_ref = self.build_virtual_machine(instance, [ 705.660152] env[61957]: ERROR nova.compute.manager [instance: 2b151571-348f-4543-a0c5-afe6458b1973] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 705.660152] env[61957]: ERROR nova.compute.manager [instance: 2b151571-348f-4543-a0c5-afe6458b1973] vif_infos = vmwarevif.get_vif_info(self._session, [ 705.660152] env[61957]: ERROR nova.compute.manager [instance: 2b151571-348f-4543-a0c5-afe6458b1973] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 705.660508] env[61957]: ERROR nova.compute.manager [instance: 2b151571-348f-4543-a0c5-afe6458b1973] for vif in network_info: [ 705.660508] env[61957]: ERROR nova.compute.manager [instance: 2b151571-348f-4543-a0c5-afe6458b1973] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 705.660508] env[61957]: ERROR nova.compute.manager [instance: 2b151571-348f-4543-a0c5-afe6458b1973] return self._sync_wrapper(fn, *args, **kwargs) [ 705.660508] env[61957]: ERROR nova.compute.manager [instance: 2b151571-348f-4543-a0c5-afe6458b1973] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 705.660508] env[61957]: ERROR nova.compute.manager [instance: 2b151571-348f-4543-a0c5-afe6458b1973] self.wait() [ 705.660508] env[61957]: ERROR nova.compute.manager [instance: 2b151571-348f-4543-a0c5-afe6458b1973] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 705.660508] env[61957]: ERROR nova.compute.manager [instance: 2b151571-348f-4543-a0c5-afe6458b1973] self[:] = self._gt.wait() [ 705.660508] env[61957]: ERROR nova.compute.manager [instance: 2b151571-348f-4543-a0c5-afe6458b1973] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 705.660508] env[61957]: ERROR nova.compute.manager [instance: 2b151571-348f-4543-a0c5-afe6458b1973] return self._exit_event.wait() [ 705.660508] env[61957]: ERROR nova.compute.manager [instance: 2b151571-348f-4543-a0c5-afe6458b1973] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 705.660508] env[61957]: ERROR nova.compute.manager [instance: 2b151571-348f-4543-a0c5-afe6458b1973] current.throw(*self._exc) [ 705.660508] env[61957]: ERROR nova.compute.manager [instance: 2b151571-348f-4543-a0c5-afe6458b1973] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 705.660508] env[61957]: ERROR nova.compute.manager [instance: 2b151571-348f-4543-a0c5-afe6458b1973] result = function(*args, **kwargs) [ 705.660871] env[61957]: ERROR nova.compute.manager [instance: 2b151571-348f-4543-a0c5-afe6458b1973] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 705.660871] env[61957]: ERROR nova.compute.manager [instance: 2b151571-348f-4543-a0c5-afe6458b1973] return func(*args, **kwargs) [ 705.660871] env[61957]: ERROR nova.compute.manager [instance: 2b151571-348f-4543-a0c5-afe6458b1973] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 705.660871] env[61957]: ERROR nova.compute.manager [instance: 2b151571-348f-4543-a0c5-afe6458b1973] raise e [ 705.660871] env[61957]: ERROR nova.compute.manager [instance: 2b151571-348f-4543-a0c5-afe6458b1973] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 705.660871] env[61957]: ERROR nova.compute.manager [instance: 2b151571-348f-4543-a0c5-afe6458b1973] nwinfo = self.network_api.allocate_for_instance( [ 705.660871] env[61957]: ERROR nova.compute.manager [instance: 2b151571-348f-4543-a0c5-afe6458b1973] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 705.660871] env[61957]: ERROR nova.compute.manager [instance: 2b151571-348f-4543-a0c5-afe6458b1973] created_port_ids = self._update_ports_for_instance( [ 705.660871] env[61957]: ERROR nova.compute.manager [instance: 2b151571-348f-4543-a0c5-afe6458b1973] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 705.660871] env[61957]: ERROR nova.compute.manager [instance: 2b151571-348f-4543-a0c5-afe6458b1973] with excutils.save_and_reraise_exception(): [ 705.660871] env[61957]: ERROR nova.compute.manager [instance: 2b151571-348f-4543-a0c5-afe6458b1973] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 705.660871] env[61957]: ERROR nova.compute.manager [instance: 2b151571-348f-4543-a0c5-afe6458b1973] self.force_reraise() [ 705.660871] env[61957]: ERROR nova.compute.manager [instance: 2b151571-348f-4543-a0c5-afe6458b1973] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 705.661343] env[61957]: ERROR nova.compute.manager [instance: 2b151571-348f-4543-a0c5-afe6458b1973] raise self.value [ 705.661343] env[61957]: ERROR nova.compute.manager [instance: 2b151571-348f-4543-a0c5-afe6458b1973] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 705.661343] env[61957]: ERROR nova.compute.manager [instance: 2b151571-348f-4543-a0c5-afe6458b1973] updated_port = self._update_port( [ 705.661343] env[61957]: ERROR nova.compute.manager [instance: 2b151571-348f-4543-a0c5-afe6458b1973] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 705.661343] env[61957]: ERROR nova.compute.manager [instance: 2b151571-348f-4543-a0c5-afe6458b1973] _ensure_no_port_binding_failure(port) [ 705.661343] env[61957]: ERROR nova.compute.manager [instance: 2b151571-348f-4543-a0c5-afe6458b1973] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 705.661343] env[61957]: ERROR nova.compute.manager [instance: 2b151571-348f-4543-a0c5-afe6458b1973] raise exception.PortBindingFailed(port_id=port['id']) [ 705.661343] env[61957]: ERROR nova.compute.manager [instance: 2b151571-348f-4543-a0c5-afe6458b1973] nova.exception.PortBindingFailed: Binding failed for port 892c3e98-da9b-4830-8c81-221309a32b88, please check neutron logs for more information. [ 705.661343] env[61957]: ERROR nova.compute.manager [instance: 2b151571-348f-4543-a0c5-afe6458b1973] [ 705.661343] env[61957]: DEBUG nova.compute.utils [None req-3d56c405-76b3-4bdb-9787-ecf239aeeb0d tempest-ImagesOneServerNegativeTestJSON-1581220397 tempest-ImagesOneServerNegativeTestJSON-1581220397-project-member] [instance: 2b151571-348f-4543-a0c5-afe6458b1973] Binding failed for port 892c3e98-da9b-4830-8c81-221309a32b88, please check neutron logs for more information. {{(pid=61957) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 705.662899] env[61957]: DEBUG oslo_concurrency.lockutils [None req-691c932a-6913-4bb6-98f2-4142c7d131a8 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.072s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 705.664431] env[61957]: INFO nova.compute.claims [None req-691c932a-6913-4bb6-98f2-4142c7d131a8 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 705.667138] env[61957]: DEBUG nova.compute.manager [None req-3d56c405-76b3-4bdb-9787-ecf239aeeb0d tempest-ImagesOneServerNegativeTestJSON-1581220397 tempest-ImagesOneServerNegativeTestJSON-1581220397-project-member] [instance: 2b151571-348f-4543-a0c5-afe6458b1973] Build of instance 2b151571-348f-4543-a0c5-afe6458b1973 was re-scheduled: Binding failed for port 892c3e98-da9b-4830-8c81-221309a32b88, please check neutron logs for more information. {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 705.667576] env[61957]: DEBUG nova.compute.manager [None req-3d56c405-76b3-4bdb-9787-ecf239aeeb0d tempest-ImagesOneServerNegativeTestJSON-1581220397 tempest-ImagesOneServerNegativeTestJSON-1581220397-project-member] [instance: 2b151571-348f-4543-a0c5-afe6458b1973] Unplugging VIFs for instance {{(pid=61957) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 705.667796] env[61957]: DEBUG oslo_concurrency.lockutils [None req-3d56c405-76b3-4bdb-9787-ecf239aeeb0d tempest-ImagesOneServerNegativeTestJSON-1581220397 tempest-ImagesOneServerNegativeTestJSON-1581220397-project-member] Acquiring lock "refresh_cache-2b151571-348f-4543-a0c5-afe6458b1973" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 705.667939] env[61957]: DEBUG oslo_concurrency.lockutils [None req-3d56c405-76b3-4bdb-9787-ecf239aeeb0d tempest-ImagesOneServerNegativeTestJSON-1581220397 tempest-ImagesOneServerNegativeTestJSON-1581220397-project-member] Acquired lock "refresh_cache-2b151571-348f-4543-a0c5-afe6458b1973" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 705.668109] env[61957]: DEBUG nova.network.neutron [None req-3d56c405-76b3-4bdb-9787-ecf239aeeb0d tempest-ImagesOneServerNegativeTestJSON-1581220397 tempest-ImagesOneServerNegativeTestJSON-1581220397-project-member] [instance: 2b151571-348f-4543-a0c5-afe6458b1973] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 705.692703] env[61957]: DEBUG oslo_vmware.api [None req-4e657274-3c21-45ad-b242-d1090e30b7c2 tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Task: {'id': task-1277364, 'name': PowerOffVM_Task, 'duration_secs': 0.166794} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 705.692703] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e657274-3c21-45ad-b242-d1090e30b7c2 tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] [instance: 188b0fd3-5d71-4feb-aca5-75a2bd28895a] Powered off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 705.692703] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-4e657274-3c21-45ad-b242-d1090e30b7c2 tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] [instance: 188b0fd3-5d71-4feb-aca5-75a2bd28895a] Unregistering the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 705.692703] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2bcda29d-ff70-45d5-a0da-f2db7f8676f6 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.709195] env[61957]: DEBUG nova.compute.manager [None req-c2f4fc75-b8bd-417b-8703-3b9048472a92 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] Start spawning the instance on the hypervisor. {{(pid=61957) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 705.715708] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-4e657274-3c21-45ad-b242-d1090e30b7c2 tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] [instance: 188b0fd3-5d71-4feb-aca5-75a2bd28895a] Unregistered the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 705.716274] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-4e657274-3c21-45ad-b242-d1090e30b7c2 tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] [instance: 188b0fd3-5d71-4feb-aca5-75a2bd28895a] Deleting contents of the VM from datastore datastore2 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 705.716274] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-4e657274-3c21-45ad-b242-d1090e30b7c2 tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Deleting the datastore file [datastore2] 188b0fd3-5d71-4feb-aca5-75a2bd28895a {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 705.716415] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-957ce6d1-e674-496e-9a61-412e70bb937c {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.723868] env[61957]: DEBUG oslo_vmware.api [None req-4e657274-3c21-45ad-b242-d1090e30b7c2 tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Waiting for the task: (returnval){ [ 705.723868] env[61957]: value = "task-1277366" [ 705.723868] env[61957]: _type = "Task" [ 705.723868] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 705.743609] env[61957]: DEBUG oslo_vmware.api [None req-4e657274-3c21-45ad-b242-d1090e30b7c2 tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Task: {'id': task-1277366, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 705.749398] env[61957]: DEBUG nova.virt.hardware [None req-c2f4fc75-b8bd-417b-8703-3b9048472a92 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T17:22:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T17:21:50Z,direct_url=,disk_format='vmdk',id=11c76a2c-f705-470a-ba9d-4657858bab38,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='81d3096d0f094373913fc3dc8f5c3c6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T17:21:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 705.749398] env[61957]: DEBUG nova.virt.hardware [None req-c2f4fc75-b8bd-417b-8703-3b9048472a92 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] Flavor limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 705.749538] env[61957]: DEBUG nova.virt.hardware [None req-c2f4fc75-b8bd-417b-8703-3b9048472a92 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] Image limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 705.749611] env[61957]: DEBUG nova.virt.hardware [None req-c2f4fc75-b8bd-417b-8703-3b9048472a92 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] Flavor pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 705.749770] env[61957]: DEBUG nova.virt.hardware [None req-c2f4fc75-b8bd-417b-8703-3b9048472a92 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] Image pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 705.749927] env[61957]: DEBUG nova.virt.hardware [None req-c2f4fc75-b8bd-417b-8703-3b9048472a92 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 705.750178] env[61957]: DEBUG nova.virt.hardware [None req-c2f4fc75-b8bd-417b-8703-3b9048472a92 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 705.750346] env[61957]: DEBUG nova.virt.hardware [None req-c2f4fc75-b8bd-417b-8703-3b9048472a92 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 705.750532] env[61957]: DEBUG nova.virt.hardware [None req-c2f4fc75-b8bd-417b-8703-3b9048472a92 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] Got 1 possible topologies {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 705.750711] env[61957]: DEBUG nova.virt.hardware [None req-c2f4fc75-b8bd-417b-8703-3b9048472a92 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 705.750960] env[61957]: DEBUG nova.virt.hardware [None req-c2f4fc75-b8bd-417b-8703-3b9048472a92 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 705.752062] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd5066e9-6f55-4109-b39e-035fbf848d32 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.761145] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a7124d4-f142-479c-8c97-4ec66edffc74 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.765709] env[61957]: DEBUG nova.network.neutron [req-1a0d4f54-bc59-495d-b8de-55df956951fe req-b1a663d6-cb2f-4ac4-bc05-a886002a713d service nova] [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 705.778539] env[61957]: ERROR nova.compute.manager [None req-c2f4fc75-b8bd-417b-8703-3b9048472a92 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port b8d0e37e-0f54-4ef7-90b0-fb5903f71e0f, please check neutron logs for more information. [ 705.778539] env[61957]: ERROR nova.compute.manager [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] Traceback (most recent call last): [ 705.778539] env[61957]: ERROR nova.compute.manager [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 705.778539] env[61957]: ERROR nova.compute.manager [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] yield resources [ 705.778539] env[61957]: ERROR nova.compute.manager [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 705.778539] env[61957]: ERROR nova.compute.manager [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] self.driver.spawn(context, instance, image_meta, [ 705.778539] env[61957]: ERROR nova.compute.manager [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 705.778539] env[61957]: ERROR nova.compute.manager [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] self._vmops.spawn(context, instance, image_meta, injected_files, [ 705.778539] env[61957]: ERROR nova.compute.manager [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 705.778539] env[61957]: ERROR nova.compute.manager [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] vm_ref = self.build_virtual_machine(instance, [ 705.778539] env[61957]: ERROR nova.compute.manager [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 705.778920] env[61957]: ERROR nova.compute.manager [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] vif_infos = vmwarevif.get_vif_info(self._session, [ 705.778920] env[61957]: ERROR nova.compute.manager [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 705.778920] env[61957]: ERROR nova.compute.manager [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] for vif in network_info: [ 705.778920] env[61957]: ERROR nova.compute.manager [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 705.778920] env[61957]: ERROR nova.compute.manager [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] return self._sync_wrapper(fn, *args, **kwargs) [ 705.778920] env[61957]: ERROR nova.compute.manager [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 705.778920] env[61957]: ERROR nova.compute.manager [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] self.wait() [ 705.778920] env[61957]: ERROR nova.compute.manager [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 705.778920] env[61957]: ERROR nova.compute.manager [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] self[:] = self._gt.wait() [ 705.778920] env[61957]: ERROR nova.compute.manager [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 705.778920] env[61957]: ERROR nova.compute.manager [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] return self._exit_event.wait() [ 705.778920] env[61957]: ERROR nova.compute.manager [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 705.778920] env[61957]: ERROR nova.compute.manager [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] current.throw(*self._exc) [ 705.779328] env[61957]: ERROR nova.compute.manager [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 705.779328] env[61957]: ERROR nova.compute.manager [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] result = function(*args, **kwargs) [ 705.779328] env[61957]: ERROR nova.compute.manager [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 705.779328] env[61957]: ERROR nova.compute.manager [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] return func(*args, **kwargs) [ 705.779328] env[61957]: ERROR nova.compute.manager [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 705.779328] env[61957]: ERROR nova.compute.manager [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] raise e [ 705.779328] env[61957]: ERROR nova.compute.manager [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 705.779328] env[61957]: ERROR nova.compute.manager [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] nwinfo = self.network_api.allocate_for_instance( [ 705.779328] env[61957]: ERROR nova.compute.manager [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 705.779328] env[61957]: ERROR nova.compute.manager [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] created_port_ids = self._update_ports_for_instance( [ 705.779328] env[61957]: ERROR nova.compute.manager [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 705.779328] env[61957]: ERROR nova.compute.manager [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] with excutils.save_and_reraise_exception(): [ 705.779328] env[61957]: ERROR nova.compute.manager [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 705.779703] env[61957]: ERROR nova.compute.manager [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] self.force_reraise() [ 705.779703] env[61957]: ERROR nova.compute.manager [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 705.779703] env[61957]: ERROR nova.compute.manager [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] raise self.value [ 705.779703] env[61957]: ERROR nova.compute.manager [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 705.779703] env[61957]: ERROR nova.compute.manager [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] updated_port = self._update_port( [ 705.779703] env[61957]: ERROR nova.compute.manager [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 705.779703] env[61957]: ERROR nova.compute.manager [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] _ensure_no_port_binding_failure(port) [ 705.779703] env[61957]: ERROR nova.compute.manager [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 705.779703] env[61957]: ERROR nova.compute.manager [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] raise exception.PortBindingFailed(port_id=port['id']) [ 705.779703] env[61957]: ERROR nova.compute.manager [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] nova.exception.PortBindingFailed: Binding failed for port b8d0e37e-0f54-4ef7-90b0-fb5903f71e0f, please check neutron logs for more information. [ 705.779703] env[61957]: ERROR nova.compute.manager [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] [ 705.779703] env[61957]: INFO nova.compute.manager [None req-c2f4fc75-b8bd-417b-8703-3b9048472a92 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] Terminating instance [ 705.780999] env[61957]: DEBUG oslo_concurrency.lockutils [None req-c2f4fc75-b8bd-417b-8703-3b9048472a92 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] Acquiring lock "refresh_cache-d0bde8ec-da55-42b5-8d2c-9df2b90b88a7" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 705.853169] env[61957]: DEBUG nova.network.neutron [req-1a0d4f54-bc59-495d-b8de-55df956951fe req-b1a663d6-cb2f-4ac4-bc05-a886002a713d service nova] [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 705.895869] env[61957]: DEBUG oslo_concurrency.lockutils [None req-2b9492ff-1cb8-4ba9-ba5e-b513c4b108ad tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 706.192392] env[61957]: DEBUG nova.network.neutron [None req-3d56c405-76b3-4bdb-9787-ecf239aeeb0d tempest-ImagesOneServerNegativeTestJSON-1581220397 tempest-ImagesOneServerNegativeTestJSON-1581220397-project-member] [instance: 2b151571-348f-4543-a0c5-afe6458b1973] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 706.233821] env[61957]: DEBUG oslo_vmware.api [None req-4e657274-3c21-45ad-b242-d1090e30b7c2 tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Task: {'id': task-1277366, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.091225} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 706.234226] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-4e657274-3c21-45ad-b242-d1090e30b7c2 tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Deleted the datastore file {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 706.234453] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-4e657274-3c21-45ad-b242-d1090e30b7c2 tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] [instance: 188b0fd3-5d71-4feb-aca5-75a2bd28895a] Deleted contents of the VM from datastore datastore2 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 706.234629] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-4e657274-3c21-45ad-b242-d1090e30b7c2 tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] [instance: 188b0fd3-5d71-4feb-aca5-75a2bd28895a] Instance destroyed {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 706.234802] env[61957]: INFO nova.compute.manager [None req-4e657274-3c21-45ad-b242-d1090e30b7c2 tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] [instance: 188b0fd3-5d71-4feb-aca5-75a2bd28895a] Took 1.09 seconds to destroy the instance on the hypervisor. [ 706.235840] env[61957]: DEBUG oslo.service.loopingcall [None req-4e657274-3c21-45ad-b242-d1090e30b7c2 tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 706.236537] env[61957]: DEBUG nova.compute.manager [-] [instance: 188b0fd3-5d71-4feb-aca5-75a2bd28895a] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 706.236666] env[61957]: DEBUG nova.network.neutron [-] [instance: 188b0fd3-5d71-4feb-aca5-75a2bd28895a] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 706.264640] env[61957]: DEBUG nova.network.neutron [-] [instance: 188b0fd3-5d71-4feb-aca5-75a2bd28895a] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 706.325609] env[61957]: DEBUG nova.network.neutron [None req-3d56c405-76b3-4bdb-9787-ecf239aeeb0d tempest-ImagesOneServerNegativeTestJSON-1581220397 tempest-ImagesOneServerNegativeTestJSON-1581220397-project-member] [instance: 2b151571-348f-4543-a0c5-afe6458b1973] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 706.355653] env[61957]: DEBUG oslo_concurrency.lockutils [req-1a0d4f54-bc59-495d-b8de-55df956951fe req-b1a663d6-cb2f-4ac4-bc05-a886002a713d service nova] Releasing lock "refresh_cache-d0bde8ec-da55-42b5-8d2c-9df2b90b88a7" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 706.355653] env[61957]: DEBUG oslo_concurrency.lockutils [None req-c2f4fc75-b8bd-417b-8703-3b9048472a92 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] Acquired lock "refresh_cache-d0bde8ec-da55-42b5-8d2c-9df2b90b88a7" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 706.355846] env[61957]: DEBUG nova.network.neutron [None req-c2f4fc75-b8bd-417b-8703-3b9048472a92 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 706.771900] env[61957]: DEBUG nova.network.neutron [-] [instance: 188b0fd3-5d71-4feb-aca5-75a2bd28895a] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 706.827964] env[61957]: DEBUG oslo_concurrency.lockutils [None req-3d56c405-76b3-4bdb-9787-ecf239aeeb0d tempest-ImagesOneServerNegativeTestJSON-1581220397 tempest-ImagesOneServerNegativeTestJSON-1581220397-project-member] Releasing lock "refresh_cache-2b151571-348f-4543-a0c5-afe6458b1973" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 706.828370] env[61957]: DEBUG nova.compute.manager [None req-3d56c405-76b3-4bdb-9787-ecf239aeeb0d tempest-ImagesOneServerNegativeTestJSON-1581220397 tempest-ImagesOneServerNegativeTestJSON-1581220397-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61957) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 706.829915] env[61957]: DEBUG nova.compute.manager [None req-3d56c405-76b3-4bdb-9787-ecf239aeeb0d tempest-ImagesOneServerNegativeTestJSON-1581220397 tempest-ImagesOneServerNegativeTestJSON-1581220397-project-member] [instance: 2b151571-348f-4543-a0c5-afe6458b1973] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 706.830129] env[61957]: DEBUG nova.network.neutron [None req-3d56c405-76b3-4bdb-9787-ecf239aeeb0d tempest-ImagesOneServerNegativeTestJSON-1581220397 tempest-ImagesOneServerNegativeTestJSON-1581220397-project-member] [instance: 2b151571-348f-4543-a0c5-afe6458b1973] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 706.849763] env[61957]: DEBUG nova.network.neutron [None req-3d56c405-76b3-4bdb-9787-ecf239aeeb0d tempest-ImagesOneServerNegativeTestJSON-1581220397 tempest-ImagesOneServerNegativeTestJSON-1581220397-project-member] [instance: 2b151571-348f-4543-a0c5-afe6458b1973] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 706.881034] env[61957]: DEBUG nova.network.neutron [None req-c2f4fc75-b8bd-417b-8703-3b9048472a92 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 707.000362] env[61957]: DEBUG nova.network.neutron [None req-c2f4fc75-b8bd-417b-8703-3b9048472a92 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 707.070942] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-767ad5b3-3730-44a4-91ed-ab0da66632cf {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.079187] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-912ca51d-69ee-4c5e-a29a-b00c7c045bdd {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.113608] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b355657e-b4f3-4ec9-a594-2ea189759f24 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.122061] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e76272c2-33d9-4eec-90c0-2f33d00a2b13 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.133760] env[61957]: DEBUG nova.compute.provider_tree [None req-691c932a-6913-4bb6-98f2-4142c7d131a8 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 707.265835] env[61957]: DEBUG nova.compute.manager [req-bfbd7cb8-f6ad-4e5f-ab55-ad0bfed89299 req-0f6ceeef-692d-47b6-a6aa-49cebc364692 service nova] [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] Received event network-vif-deleted-b8d0e37e-0f54-4ef7-90b0-fb5903f71e0f {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 707.274008] env[61957]: INFO nova.compute.manager [-] [instance: 188b0fd3-5d71-4feb-aca5-75a2bd28895a] Took 1.04 seconds to deallocate network for instance. [ 707.352493] env[61957]: DEBUG nova.network.neutron [None req-3d56c405-76b3-4bdb-9787-ecf239aeeb0d tempest-ImagesOneServerNegativeTestJSON-1581220397 tempest-ImagesOneServerNegativeTestJSON-1581220397-project-member] [instance: 2b151571-348f-4543-a0c5-afe6458b1973] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 707.505444] env[61957]: DEBUG oslo_concurrency.lockutils [None req-c2f4fc75-b8bd-417b-8703-3b9048472a92 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] Releasing lock "refresh_cache-d0bde8ec-da55-42b5-8d2c-9df2b90b88a7" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 707.505774] env[61957]: DEBUG nova.compute.manager [None req-c2f4fc75-b8bd-417b-8703-3b9048472a92 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] Start destroying the instance on the hypervisor. {{(pid=61957) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 707.505968] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-c2f4fc75-b8bd-417b-8703-3b9048472a92 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] Destroying instance {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 707.506285] env[61957]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8b0a3964-ac9e-423f-bfaa-60160263bb5a {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.517908] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb9d2270-4428-49c6-b57e-ae55c336d9dd {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.537620] env[61957]: WARNING nova.virt.vmwareapi.vmops [None req-c2f4fc75-b8bd-417b-8703-3b9048472a92 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance d0bde8ec-da55-42b5-8d2c-9df2b90b88a7 could not be found. [ 707.537908] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-c2f4fc75-b8bd-417b-8703-3b9048472a92 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] Instance destroyed {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 707.538116] env[61957]: INFO nova.compute.manager [None req-c2f4fc75-b8bd-417b-8703-3b9048472a92 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] Took 0.03 seconds to destroy the instance on the hypervisor. [ 707.538418] env[61957]: DEBUG oslo.service.loopingcall [None req-c2f4fc75-b8bd-417b-8703-3b9048472a92 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 707.538639] env[61957]: DEBUG nova.compute.manager [-] [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 707.538732] env[61957]: DEBUG nova.network.neutron [-] [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 707.556469] env[61957]: DEBUG nova.network.neutron [-] [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 707.636977] env[61957]: DEBUG nova.scheduler.client.report [None req-691c932a-6913-4bb6-98f2-4142c7d131a8 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 707.779338] env[61957]: DEBUG oslo_concurrency.lockutils [None req-4e657274-3c21-45ad-b242-d1090e30b7c2 tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 707.855031] env[61957]: INFO nova.compute.manager [None req-3d56c405-76b3-4bdb-9787-ecf239aeeb0d tempest-ImagesOneServerNegativeTestJSON-1581220397 tempest-ImagesOneServerNegativeTestJSON-1581220397-project-member] [instance: 2b151571-348f-4543-a0c5-afe6458b1973] Took 1.02 seconds to deallocate network for instance. [ 708.059897] env[61957]: DEBUG nova.network.neutron [-] [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 708.143022] env[61957]: DEBUG oslo_concurrency.lockutils [None req-691c932a-6913-4bb6-98f2-4142c7d131a8 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.479s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 708.143135] env[61957]: DEBUG nova.compute.manager [None req-691c932a-6913-4bb6-98f2-4142c7d131a8 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] Start building networks asynchronously for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 708.146433] env[61957]: DEBUG oslo_concurrency.lockutils [None req-456a04ab-263c-4c55-898a-70a189651712 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.577s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 708.147810] env[61957]: INFO nova.compute.claims [None req-456a04ab-263c-4c55-898a-70a189651712 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 708.564104] env[61957]: INFO nova.compute.manager [-] [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] Took 1.03 seconds to deallocate network for instance. [ 708.568138] env[61957]: DEBUG nova.compute.claims [None req-c2f4fc75-b8bd-417b-8703-3b9048472a92 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] Aborting claim: {{(pid=61957) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 708.568339] env[61957]: DEBUG oslo_concurrency.lockutils [None req-c2f4fc75-b8bd-417b-8703-3b9048472a92 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 708.653286] env[61957]: DEBUG nova.compute.utils [None req-691c932a-6913-4bb6-98f2-4142c7d131a8 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Using /dev/sd instead of None {{(pid=61957) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 708.659022] env[61957]: DEBUG nova.compute.manager [None req-691c932a-6913-4bb6-98f2-4142c7d131a8 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] Allocating IP information in the background. {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 708.659022] env[61957]: DEBUG nova.network.neutron [None req-691c932a-6913-4bb6-98f2-4142c7d131a8 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] allocate_for_instance() {{(pid=61957) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 708.724894] env[61957]: DEBUG nova.policy [None req-691c932a-6913-4bb6-98f2-4142c7d131a8 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3136bf0d1db3411fb1ff1ebe9c6045f6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1dbdc699741a48af93c52068d88a5357', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61957) authorize /opt/stack/nova/nova/policy.py:203}} [ 708.887129] env[61957]: INFO nova.scheduler.client.report [None req-3d56c405-76b3-4bdb-9787-ecf239aeeb0d tempest-ImagesOneServerNegativeTestJSON-1581220397 tempest-ImagesOneServerNegativeTestJSON-1581220397-project-member] Deleted allocations for instance 2b151571-348f-4543-a0c5-afe6458b1973 [ 709.160328] env[61957]: DEBUG nova.compute.manager [None req-691c932a-6913-4bb6-98f2-4142c7d131a8 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] Start building block device mappings for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 709.340978] env[61957]: DEBUG nova.network.neutron [None req-691c932a-6913-4bb6-98f2-4142c7d131a8 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] Successfully created port: 75f87eff-42b6-4098-9846-409a7c48d82f {{(pid=61957) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 709.400038] env[61957]: DEBUG oslo_concurrency.lockutils [None req-3d56c405-76b3-4bdb-9787-ecf239aeeb0d tempest-ImagesOneServerNegativeTestJSON-1581220397 tempest-ImagesOneServerNegativeTestJSON-1581220397-project-member] Lock "2b151571-348f-4543-a0c5-afe6458b1973" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 105.408s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 709.521413] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12cb1ff9-0d70-48cb-baf5-79becc6ff0ad {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.528815] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e053f3f0-0386-4e04-83ec-5ec247e1cc42 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.559124] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-752d8d60-2d15-4025-8d6c-4615252099a2 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.566265] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8dd2ee2f-fed0-4342-8fbb-8c70adef693d {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.578993] env[61957]: DEBUG nova.compute.provider_tree [None req-456a04ab-263c-4c55-898a-70a189651712 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 709.904046] env[61957]: DEBUG nova.compute.manager [None req-f462e606-4c03-40e0-ac91-60efbd401559 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] Starting instance... {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 710.082024] env[61957]: DEBUG nova.scheduler.client.report [None req-456a04ab-263c-4c55-898a-70a189651712 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 710.172239] env[61957]: DEBUG nova.compute.manager [None req-691c932a-6913-4bb6-98f2-4142c7d131a8 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] Start spawning the instance on the hypervisor. {{(pid=61957) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 710.214117] env[61957]: DEBUG nova.virt.hardware [None req-691c932a-6913-4bb6-98f2-4142c7d131a8 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T17:22:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T17:21:50Z,direct_url=,disk_format='vmdk',id=11c76a2c-f705-470a-ba9d-4657858bab38,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='81d3096d0f094373913fc3dc8f5c3c6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T17:21:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 710.214665] env[61957]: DEBUG nova.virt.hardware [None req-691c932a-6913-4bb6-98f2-4142c7d131a8 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Flavor limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 710.214665] env[61957]: DEBUG nova.virt.hardware [None req-691c932a-6913-4bb6-98f2-4142c7d131a8 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Image limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 710.214827] env[61957]: DEBUG nova.virt.hardware [None req-691c932a-6913-4bb6-98f2-4142c7d131a8 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Flavor pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 710.214953] env[61957]: DEBUG nova.virt.hardware [None req-691c932a-6913-4bb6-98f2-4142c7d131a8 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Image pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 710.215093] env[61957]: DEBUG nova.virt.hardware [None req-691c932a-6913-4bb6-98f2-4142c7d131a8 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 710.215371] env[61957]: DEBUG nova.virt.hardware [None req-691c932a-6913-4bb6-98f2-4142c7d131a8 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 710.215531] env[61957]: DEBUG nova.virt.hardware [None req-691c932a-6913-4bb6-98f2-4142c7d131a8 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 710.215717] env[61957]: DEBUG nova.virt.hardware [None req-691c932a-6913-4bb6-98f2-4142c7d131a8 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Got 1 possible topologies {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 710.215902] env[61957]: DEBUG nova.virt.hardware [None req-691c932a-6913-4bb6-98f2-4142c7d131a8 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 710.216110] env[61957]: DEBUG nova.virt.hardware [None req-691c932a-6913-4bb6-98f2-4142c7d131a8 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 710.217091] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d8b0815-6e4e-4bf8-a955-52f64ad2edf7 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.228800] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-975b62e5-d49b-4c30-877d-7771a8f7b194 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.440929] env[61957]: DEBUG oslo_concurrency.lockutils [None req-f462e606-4c03-40e0-ac91-60efbd401559 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 710.593448] env[61957]: DEBUG oslo_concurrency.lockutils [None req-456a04ab-263c-4c55-898a-70a189651712 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.447s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 710.593976] env[61957]: DEBUG nova.compute.manager [None req-456a04ab-263c-4c55-898a-70a189651712 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] Start building networks asynchronously for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 710.599612] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ca1e8d1e-8c18-4652-924a-ab1cdb751a05 tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 31.114s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 710.601788] env[61957]: INFO nova.compute.claims [None req-ca1e8d1e-8c18-4652-924a-ab1cdb751a05 tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] [instance: 825130d4-b1f6-434e-be3e-dca952abe930] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 710.662402] env[61957]: DEBUG nova.compute.manager [req-a54afb02-2307-45f1-b357-e6ea69a9fcd7 req-7ee47b85-adcf-4362-b222-0e0d7e1a417f service nova] [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] Received event network-changed-75f87eff-42b6-4098-9846-409a7c48d82f {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 710.662402] env[61957]: DEBUG nova.compute.manager [req-a54afb02-2307-45f1-b357-e6ea69a9fcd7 req-7ee47b85-adcf-4362-b222-0e0d7e1a417f service nova] [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] Refreshing instance network info cache due to event network-changed-75f87eff-42b6-4098-9846-409a7c48d82f. {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 710.662402] env[61957]: DEBUG oslo_concurrency.lockutils [req-a54afb02-2307-45f1-b357-e6ea69a9fcd7 req-7ee47b85-adcf-4362-b222-0e0d7e1a417f service nova] Acquiring lock "refresh_cache-c5024e22-6499-4f2e-8d1b-99ca34b33104" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 710.662522] env[61957]: DEBUG oslo_concurrency.lockutils [req-a54afb02-2307-45f1-b357-e6ea69a9fcd7 req-7ee47b85-adcf-4362-b222-0e0d7e1a417f service nova] Acquired lock "refresh_cache-c5024e22-6499-4f2e-8d1b-99ca34b33104" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 710.665425] env[61957]: DEBUG nova.network.neutron [req-a54afb02-2307-45f1-b357-e6ea69a9fcd7 req-7ee47b85-adcf-4362-b222-0e0d7e1a417f service nova] [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] Refreshing network info cache for port 75f87eff-42b6-4098-9846-409a7c48d82f {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 710.913852] env[61957]: ERROR nova.compute.manager [None req-691c932a-6913-4bb6-98f2-4142c7d131a8 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 75f87eff-42b6-4098-9846-409a7c48d82f, please check neutron logs for more information. [ 710.913852] env[61957]: ERROR nova.compute.manager Traceback (most recent call last): [ 710.913852] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 710.913852] env[61957]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 710.913852] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 710.913852] env[61957]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 710.913852] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 710.913852] env[61957]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 710.913852] env[61957]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 710.913852] env[61957]: ERROR nova.compute.manager self.force_reraise() [ 710.913852] env[61957]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 710.913852] env[61957]: ERROR nova.compute.manager raise self.value [ 710.913852] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 710.913852] env[61957]: ERROR nova.compute.manager updated_port = self._update_port( [ 710.913852] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 710.913852] env[61957]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 710.914307] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 710.914307] env[61957]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 710.914307] env[61957]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 75f87eff-42b6-4098-9846-409a7c48d82f, please check neutron logs for more information. [ 710.914307] env[61957]: ERROR nova.compute.manager [ 710.914307] env[61957]: Traceback (most recent call last): [ 710.914307] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 710.914307] env[61957]: listener.cb(fileno) [ 710.914307] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 710.914307] env[61957]: result = function(*args, **kwargs) [ 710.914307] env[61957]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 710.914307] env[61957]: return func(*args, **kwargs) [ 710.914307] env[61957]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 710.914307] env[61957]: raise e [ 710.914307] env[61957]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 710.914307] env[61957]: nwinfo = self.network_api.allocate_for_instance( [ 710.914307] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 710.914307] env[61957]: created_port_ids = self._update_ports_for_instance( [ 710.914307] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 710.914307] env[61957]: with excutils.save_and_reraise_exception(): [ 710.914307] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 710.914307] env[61957]: self.force_reraise() [ 710.914307] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 710.914307] env[61957]: raise self.value [ 710.914307] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 710.914307] env[61957]: updated_port = self._update_port( [ 710.914307] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 710.914307] env[61957]: _ensure_no_port_binding_failure(port) [ 710.914307] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 710.914307] env[61957]: raise exception.PortBindingFailed(port_id=port['id']) [ 710.914996] env[61957]: nova.exception.PortBindingFailed: Binding failed for port 75f87eff-42b6-4098-9846-409a7c48d82f, please check neutron logs for more information. [ 710.914996] env[61957]: Removing descriptor: 16 [ 710.914996] env[61957]: ERROR nova.compute.manager [None req-691c932a-6913-4bb6-98f2-4142c7d131a8 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 75f87eff-42b6-4098-9846-409a7c48d82f, please check neutron logs for more information. [ 710.914996] env[61957]: ERROR nova.compute.manager [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] Traceback (most recent call last): [ 710.914996] env[61957]: ERROR nova.compute.manager [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 710.914996] env[61957]: ERROR nova.compute.manager [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] yield resources [ 710.914996] env[61957]: ERROR nova.compute.manager [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 710.914996] env[61957]: ERROR nova.compute.manager [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] self.driver.spawn(context, instance, image_meta, [ 710.914996] env[61957]: ERROR nova.compute.manager [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 710.914996] env[61957]: ERROR nova.compute.manager [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] self._vmops.spawn(context, instance, image_meta, injected_files, [ 710.914996] env[61957]: ERROR nova.compute.manager [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 710.914996] env[61957]: ERROR nova.compute.manager [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] vm_ref = self.build_virtual_machine(instance, [ 710.915440] env[61957]: ERROR nova.compute.manager [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 710.915440] env[61957]: ERROR nova.compute.manager [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] vif_infos = vmwarevif.get_vif_info(self._session, [ 710.915440] env[61957]: ERROR nova.compute.manager [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 710.915440] env[61957]: ERROR nova.compute.manager [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] for vif in network_info: [ 710.915440] env[61957]: ERROR nova.compute.manager [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 710.915440] env[61957]: ERROR nova.compute.manager [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] return self._sync_wrapper(fn, *args, **kwargs) [ 710.915440] env[61957]: ERROR nova.compute.manager [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 710.915440] env[61957]: ERROR nova.compute.manager [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] self.wait() [ 710.915440] env[61957]: ERROR nova.compute.manager [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 710.915440] env[61957]: ERROR nova.compute.manager [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] self[:] = self._gt.wait() [ 710.915440] env[61957]: ERROR nova.compute.manager [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 710.915440] env[61957]: ERROR nova.compute.manager [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] return self._exit_event.wait() [ 710.915440] env[61957]: ERROR nova.compute.manager [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 710.915749] env[61957]: ERROR nova.compute.manager [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] result = hub.switch() [ 710.915749] env[61957]: ERROR nova.compute.manager [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 710.915749] env[61957]: ERROR nova.compute.manager [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] return self.greenlet.switch() [ 710.915749] env[61957]: ERROR nova.compute.manager [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 710.915749] env[61957]: ERROR nova.compute.manager [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] result = function(*args, **kwargs) [ 710.915749] env[61957]: ERROR nova.compute.manager [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 710.915749] env[61957]: ERROR nova.compute.manager [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] return func(*args, **kwargs) [ 710.915749] env[61957]: ERROR nova.compute.manager [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 710.915749] env[61957]: ERROR nova.compute.manager [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] raise e [ 710.915749] env[61957]: ERROR nova.compute.manager [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 710.915749] env[61957]: ERROR nova.compute.manager [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] nwinfo = self.network_api.allocate_for_instance( [ 710.915749] env[61957]: ERROR nova.compute.manager [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 710.915749] env[61957]: ERROR nova.compute.manager [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] created_port_ids = self._update_ports_for_instance( [ 710.916069] env[61957]: ERROR nova.compute.manager [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 710.916069] env[61957]: ERROR nova.compute.manager [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] with excutils.save_and_reraise_exception(): [ 710.916069] env[61957]: ERROR nova.compute.manager [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 710.916069] env[61957]: ERROR nova.compute.manager [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] self.force_reraise() [ 710.916069] env[61957]: ERROR nova.compute.manager [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 710.916069] env[61957]: ERROR nova.compute.manager [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] raise self.value [ 710.916069] env[61957]: ERROR nova.compute.manager [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 710.916069] env[61957]: ERROR nova.compute.manager [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] updated_port = self._update_port( [ 710.916069] env[61957]: ERROR nova.compute.manager [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 710.916069] env[61957]: ERROR nova.compute.manager [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] _ensure_no_port_binding_failure(port) [ 710.916069] env[61957]: ERROR nova.compute.manager [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 710.916069] env[61957]: ERROR nova.compute.manager [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] raise exception.PortBindingFailed(port_id=port['id']) [ 710.916363] env[61957]: ERROR nova.compute.manager [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] nova.exception.PortBindingFailed: Binding failed for port 75f87eff-42b6-4098-9846-409a7c48d82f, please check neutron logs for more information. [ 710.916363] env[61957]: ERROR nova.compute.manager [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] [ 710.916363] env[61957]: INFO nova.compute.manager [None req-691c932a-6913-4bb6-98f2-4142c7d131a8 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] Terminating instance [ 710.917873] env[61957]: DEBUG oslo_concurrency.lockutils [None req-691c932a-6913-4bb6-98f2-4142c7d131a8 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Acquiring lock "refresh_cache-c5024e22-6499-4f2e-8d1b-99ca34b33104" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 711.101628] env[61957]: DEBUG nova.compute.utils [None req-456a04ab-263c-4c55-898a-70a189651712 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] Using /dev/sd instead of None {{(pid=61957) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 711.103114] env[61957]: DEBUG nova.compute.manager [None req-456a04ab-263c-4c55-898a-70a189651712 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] Allocating IP information in the background. {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 711.103290] env[61957]: DEBUG nova.network.neutron [None req-456a04ab-263c-4c55-898a-70a189651712 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] allocate_for_instance() {{(pid=61957) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 711.152338] env[61957]: DEBUG nova.policy [None req-456a04ab-263c-4c55-898a-70a189651712 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '02106e108faa446cad15bec05f2cc763', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c33de8fc6166438dabcfb820e0513085', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61957) authorize /opt/stack/nova/nova/policy.py:203}} [ 711.199359] env[61957]: DEBUG nova.network.neutron [req-a54afb02-2307-45f1-b357-e6ea69a9fcd7 req-7ee47b85-adcf-4362-b222-0e0d7e1a417f service nova] [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 711.312489] env[61957]: DEBUG nova.network.neutron [req-a54afb02-2307-45f1-b357-e6ea69a9fcd7 req-7ee47b85-adcf-4362-b222-0e0d7e1a417f service nova] [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 711.540798] env[61957]: DEBUG nova.network.neutron [None req-456a04ab-263c-4c55-898a-70a189651712 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] Successfully created port: a0d948b4-5d2e-4133-8e39-e685c226b80f {{(pid=61957) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 711.607793] env[61957]: DEBUG nova.compute.manager [None req-456a04ab-263c-4c55-898a-70a189651712 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] Start building block device mappings for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 711.815513] env[61957]: DEBUG oslo_concurrency.lockutils [req-a54afb02-2307-45f1-b357-e6ea69a9fcd7 req-7ee47b85-adcf-4362-b222-0e0d7e1a417f service nova] Releasing lock "refresh_cache-c5024e22-6499-4f2e-8d1b-99ca34b33104" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 711.816190] env[61957]: DEBUG oslo_concurrency.lockutils [None req-691c932a-6913-4bb6-98f2-4142c7d131a8 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Acquired lock "refresh_cache-c5024e22-6499-4f2e-8d1b-99ca34b33104" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 711.816352] env[61957]: DEBUG nova.network.neutron [None req-691c932a-6913-4bb6-98f2-4142c7d131a8 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 712.022315] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6714c62d-9098-496d-85ee-48e674a62c08 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.031744] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9666a717-0fbf-4f60-8f13-78338f720aeb {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.066233] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec354afd-0086-4288-93b9-d415fae6c5bf {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.073248] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8a45511-a0c2-4444-9a97-16533d2d330d {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.088898] env[61957]: DEBUG nova.compute.provider_tree [None req-ca1e8d1e-8c18-4652-924a-ab1cdb751a05 tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 712.341279] env[61957]: DEBUG nova.network.neutron [None req-691c932a-6913-4bb6-98f2-4142c7d131a8 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 712.495068] env[61957]: DEBUG nova.network.neutron [None req-691c932a-6913-4bb6-98f2-4142c7d131a8 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 712.594708] env[61957]: DEBUG nova.scheduler.client.report [None req-ca1e8d1e-8c18-4652-924a-ab1cdb751a05 tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 712.621731] env[61957]: DEBUG nova.compute.manager [None req-456a04ab-263c-4c55-898a-70a189651712 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] Start spawning the instance on the hypervisor. {{(pid=61957) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 712.658372] env[61957]: DEBUG nova.virt.hardware [None req-456a04ab-263c-4c55-898a-70a189651712 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T17:22:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T17:21:50Z,direct_url=,disk_format='vmdk',id=11c76a2c-f705-470a-ba9d-4657858bab38,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='81d3096d0f094373913fc3dc8f5c3c6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T17:21:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 712.658636] env[61957]: DEBUG nova.virt.hardware [None req-456a04ab-263c-4c55-898a-70a189651712 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] Flavor limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 712.658791] env[61957]: DEBUG nova.virt.hardware [None req-456a04ab-263c-4c55-898a-70a189651712 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] Image limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 712.658970] env[61957]: DEBUG nova.virt.hardware [None req-456a04ab-263c-4c55-898a-70a189651712 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] Flavor pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 712.659128] env[61957]: DEBUG nova.virt.hardware [None req-456a04ab-263c-4c55-898a-70a189651712 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] Image pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 712.659347] env[61957]: DEBUG nova.virt.hardware [None req-456a04ab-263c-4c55-898a-70a189651712 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 712.659571] env[61957]: DEBUG nova.virt.hardware [None req-456a04ab-263c-4c55-898a-70a189651712 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 712.660478] env[61957]: DEBUG nova.virt.hardware [None req-456a04ab-263c-4c55-898a-70a189651712 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 712.660723] env[61957]: DEBUG nova.virt.hardware [None req-456a04ab-263c-4c55-898a-70a189651712 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] Got 1 possible topologies {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 712.660904] env[61957]: DEBUG nova.virt.hardware [None req-456a04ab-263c-4c55-898a-70a189651712 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 712.661091] env[61957]: DEBUG nova.virt.hardware [None req-456a04ab-263c-4c55-898a-70a189651712 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 712.665470] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab501774-4629-4885-ba35-255592a4bda8 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.675968] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5521ccd-ebbf-4e7d-b306-66123f3b1c6a {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.691619] env[61957]: ERROR nova.compute.manager [None req-456a04ab-263c-4c55-898a-70a189651712 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port a0d948b4-5d2e-4133-8e39-e685c226b80f, please check neutron logs for more information. [ 712.691619] env[61957]: ERROR nova.compute.manager Traceback (most recent call last): [ 712.691619] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 712.691619] env[61957]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 712.691619] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 712.691619] env[61957]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 712.691619] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 712.691619] env[61957]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 712.691619] env[61957]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 712.691619] env[61957]: ERROR nova.compute.manager self.force_reraise() [ 712.691619] env[61957]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 712.691619] env[61957]: ERROR nova.compute.manager raise self.value [ 712.691619] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 712.691619] env[61957]: ERROR nova.compute.manager updated_port = self._update_port( [ 712.691619] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 712.691619] env[61957]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 712.692331] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 712.692331] env[61957]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 712.692331] env[61957]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port a0d948b4-5d2e-4133-8e39-e685c226b80f, please check neutron logs for more information. [ 712.692331] env[61957]: ERROR nova.compute.manager [ 712.692331] env[61957]: Traceback (most recent call last): [ 712.692331] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 712.692331] env[61957]: listener.cb(fileno) [ 712.692331] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 712.692331] env[61957]: result = function(*args, **kwargs) [ 712.692331] env[61957]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 712.692331] env[61957]: return func(*args, **kwargs) [ 712.692331] env[61957]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 712.692331] env[61957]: raise e [ 712.692331] env[61957]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 712.692331] env[61957]: nwinfo = self.network_api.allocate_for_instance( [ 712.692331] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 712.692331] env[61957]: created_port_ids = self._update_ports_for_instance( [ 712.692331] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 712.692331] env[61957]: with excutils.save_and_reraise_exception(): [ 712.692331] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 712.692331] env[61957]: self.force_reraise() [ 712.692331] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 712.692331] env[61957]: raise self.value [ 712.692331] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 712.692331] env[61957]: updated_port = self._update_port( [ 712.692331] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 712.692331] env[61957]: _ensure_no_port_binding_failure(port) [ 712.692331] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 712.692331] env[61957]: raise exception.PortBindingFailed(port_id=port['id']) [ 712.693041] env[61957]: nova.exception.PortBindingFailed: Binding failed for port a0d948b4-5d2e-4133-8e39-e685c226b80f, please check neutron logs for more information. [ 712.693041] env[61957]: Removing descriptor: 16 [ 712.693041] env[61957]: ERROR nova.compute.manager [None req-456a04ab-263c-4c55-898a-70a189651712 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port a0d948b4-5d2e-4133-8e39-e685c226b80f, please check neutron logs for more information. [ 712.693041] env[61957]: ERROR nova.compute.manager [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] Traceback (most recent call last): [ 712.693041] env[61957]: ERROR nova.compute.manager [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 712.693041] env[61957]: ERROR nova.compute.manager [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] yield resources [ 712.693041] env[61957]: ERROR nova.compute.manager [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 712.693041] env[61957]: ERROR nova.compute.manager [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] self.driver.spawn(context, instance, image_meta, [ 712.693041] env[61957]: ERROR nova.compute.manager [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 712.693041] env[61957]: ERROR nova.compute.manager [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] self._vmops.spawn(context, instance, image_meta, injected_files, [ 712.693041] env[61957]: ERROR nova.compute.manager [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 712.693041] env[61957]: ERROR nova.compute.manager [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] vm_ref = self.build_virtual_machine(instance, [ 712.693499] env[61957]: ERROR nova.compute.manager [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 712.693499] env[61957]: ERROR nova.compute.manager [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] vif_infos = vmwarevif.get_vif_info(self._session, [ 712.693499] env[61957]: ERROR nova.compute.manager [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 712.693499] env[61957]: ERROR nova.compute.manager [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] for vif in network_info: [ 712.693499] env[61957]: ERROR nova.compute.manager [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 712.693499] env[61957]: ERROR nova.compute.manager [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] return self._sync_wrapper(fn, *args, **kwargs) [ 712.693499] env[61957]: ERROR nova.compute.manager [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 712.693499] env[61957]: ERROR nova.compute.manager [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] self.wait() [ 712.693499] env[61957]: ERROR nova.compute.manager [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 712.693499] env[61957]: ERROR nova.compute.manager [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] self[:] = self._gt.wait() [ 712.693499] env[61957]: ERROR nova.compute.manager [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 712.693499] env[61957]: ERROR nova.compute.manager [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] return self._exit_event.wait() [ 712.693499] env[61957]: ERROR nova.compute.manager [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 712.694032] env[61957]: ERROR nova.compute.manager [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] current.throw(*self._exc) [ 712.694032] env[61957]: ERROR nova.compute.manager [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 712.694032] env[61957]: ERROR nova.compute.manager [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] result = function(*args, **kwargs) [ 712.694032] env[61957]: ERROR nova.compute.manager [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 712.694032] env[61957]: ERROR nova.compute.manager [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] return func(*args, **kwargs) [ 712.694032] env[61957]: ERROR nova.compute.manager [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 712.694032] env[61957]: ERROR nova.compute.manager [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] raise e [ 712.694032] env[61957]: ERROR nova.compute.manager [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 712.694032] env[61957]: ERROR nova.compute.manager [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] nwinfo = self.network_api.allocate_for_instance( [ 712.694032] env[61957]: ERROR nova.compute.manager [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 712.694032] env[61957]: ERROR nova.compute.manager [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] created_port_ids = self._update_ports_for_instance( [ 712.694032] env[61957]: ERROR nova.compute.manager [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 712.694032] env[61957]: ERROR nova.compute.manager [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] with excutils.save_and_reraise_exception(): [ 712.694528] env[61957]: ERROR nova.compute.manager [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 712.694528] env[61957]: ERROR nova.compute.manager [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] self.force_reraise() [ 712.694528] env[61957]: ERROR nova.compute.manager [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 712.694528] env[61957]: ERROR nova.compute.manager [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] raise self.value [ 712.694528] env[61957]: ERROR nova.compute.manager [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 712.694528] env[61957]: ERROR nova.compute.manager [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] updated_port = self._update_port( [ 712.694528] env[61957]: ERROR nova.compute.manager [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 712.694528] env[61957]: ERROR nova.compute.manager [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] _ensure_no_port_binding_failure(port) [ 712.694528] env[61957]: ERROR nova.compute.manager [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 712.694528] env[61957]: ERROR nova.compute.manager [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] raise exception.PortBindingFailed(port_id=port['id']) [ 712.694528] env[61957]: ERROR nova.compute.manager [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] nova.exception.PortBindingFailed: Binding failed for port a0d948b4-5d2e-4133-8e39-e685c226b80f, please check neutron logs for more information. [ 712.694528] env[61957]: ERROR nova.compute.manager [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] [ 712.694830] env[61957]: INFO nova.compute.manager [None req-456a04ab-263c-4c55-898a-70a189651712 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] Terminating instance [ 712.699115] env[61957]: DEBUG oslo_concurrency.lockutils [None req-456a04ab-263c-4c55-898a-70a189651712 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] Acquiring lock "refresh_cache-6241fb44-ce1c-4881-bc32-7d19ddc83a61" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 712.699244] env[61957]: DEBUG oslo_concurrency.lockutils [None req-456a04ab-263c-4c55-898a-70a189651712 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] Acquired lock "refresh_cache-6241fb44-ce1c-4881-bc32-7d19ddc83a61" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 712.699409] env[61957]: DEBUG nova.network.neutron [None req-456a04ab-263c-4c55-898a-70a189651712 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 712.744245] env[61957]: DEBUG nova.compute.manager [req-f0f061db-6270-4da7-94e4-248c6bf45880 req-9769aa9b-6ba9-4a9d-b263-53d0ddcad45e service nova] [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] Received event network-vif-deleted-75f87eff-42b6-4098-9846-409a7c48d82f {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 712.744245] env[61957]: DEBUG nova.compute.manager [req-f0f061db-6270-4da7-94e4-248c6bf45880 req-9769aa9b-6ba9-4a9d-b263-53d0ddcad45e service nova] [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] Received event network-changed-a0d948b4-5d2e-4133-8e39-e685c226b80f {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 712.744245] env[61957]: DEBUG nova.compute.manager [req-f0f061db-6270-4da7-94e4-248c6bf45880 req-9769aa9b-6ba9-4a9d-b263-53d0ddcad45e service nova] [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] Refreshing instance network info cache due to event network-changed-a0d948b4-5d2e-4133-8e39-e685c226b80f. {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 712.744440] env[61957]: DEBUG oslo_concurrency.lockutils [req-f0f061db-6270-4da7-94e4-248c6bf45880 req-9769aa9b-6ba9-4a9d-b263-53d0ddcad45e service nova] Acquiring lock "refresh_cache-6241fb44-ce1c-4881-bc32-7d19ddc83a61" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 712.914220] env[61957]: DEBUG oslo_concurrency.lockutils [None req-dce171c8-b166-439d-a4ea-36b63ebfc3eb tempest-ServerGroupTestJSON-950728618 tempest-ServerGroupTestJSON-950728618-project-member] Acquiring lock "e9ac118f-08b3-430b-848c-461c2b2e3e02" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 712.915231] env[61957]: DEBUG oslo_concurrency.lockutils [None req-dce171c8-b166-439d-a4ea-36b63ebfc3eb tempest-ServerGroupTestJSON-950728618 tempest-ServerGroupTestJSON-950728618-project-member] Lock "e9ac118f-08b3-430b-848c-461c2b2e3e02" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 712.997963] env[61957]: DEBUG oslo_concurrency.lockutils [None req-691c932a-6913-4bb6-98f2-4142c7d131a8 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Releasing lock "refresh_cache-c5024e22-6499-4f2e-8d1b-99ca34b33104" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 712.998477] env[61957]: DEBUG nova.compute.manager [None req-691c932a-6913-4bb6-98f2-4142c7d131a8 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] Start destroying the instance on the hypervisor. {{(pid=61957) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 712.998674] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-691c932a-6913-4bb6-98f2-4142c7d131a8 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] Destroying instance {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 712.999302] env[61957]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f3f03844-5b68-4b30-8465-a12762ad639c {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.008073] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47feca90-7b61-435f-89b2-23ae5283e079 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.030929] env[61957]: WARNING nova.virt.vmwareapi.vmops [None req-691c932a-6913-4bb6-98f2-4142c7d131a8 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance c5024e22-6499-4f2e-8d1b-99ca34b33104 could not be found. [ 713.030929] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-691c932a-6913-4bb6-98f2-4142c7d131a8 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] Instance destroyed {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 713.030929] env[61957]: INFO nova.compute.manager [None req-691c932a-6913-4bb6-98f2-4142c7d131a8 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] Took 0.03 seconds to destroy the instance on the hypervisor. [ 713.031981] env[61957]: DEBUG oslo.service.loopingcall [None req-691c932a-6913-4bb6-98f2-4142c7d131a8 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 713.031981] env[61957]: DEBUG nova.compute.manager [-] [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 713.031981] env[61957]: DEBUG nova.network.neutron [-] [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 713.048062] env[61957]: DEBUG nova.network.neutron [-] [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 713.100823] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ca1e8d1e-8c18-4652-924a-ab1cdb751a05 tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.501s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 713.101496] env[61957]: DEBUG nova.compute.manager [None req-ca1e8d1e-8c18-4652-924a-ab1cdb751a05 tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] [instance: 825130d4-b1f6-434e-be3e-dca952abe930] Start building networks asynchronously for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 713.104125] env[61957]: DEBUG oslo_concurrency.lockutils [None req-58fe0400-f1d0-4e1d-a459-7541ba7aeeed tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.027s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 713.105823] env[61957]: INFO nova.compute.claims [None req-58fe0400-f1d0-4e1d-a459-7541ba7aeeed tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] [instance: c369d148-2093-4724-82cc-d80a67131fea] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 713.219398] env[61957]: DEBUG nova.network.neutron [None req-456a04ab-263c-4c55-898a-70a189651712 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 713.305845] env[61957]: DEBUG nova.network.neutron [None req-456a04ab-263c-4c55-898a-70a189651712 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 713.550947] env[61957]: DEBUG nova.network.neutron [-] [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 713.616634] env[61957]: DEBUG nova.compute.utils [None req-ca1e8d1e-8c18-4652-924a-ab1cdb751a05 tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] Using /dev/sd instead of None {{(pid=61957) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 713.618509] env[61957]: DEBUG nova.compute.manager [None req-ca1e8d1e-8c18-4652-924a-ab1cdb751a05 tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] [instance: 825130d4-b1f6-434e-be3e-dca952abe930] Allocating IP information in the background. {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 713.619205] env[61957]: DEBUG nova.network.neutron [None req-ca1e8d1e-8c18-4652-924a-ab1cdb751a05 tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] [instance: 825130d4-b1f6-434e-be3e-dca952abe930] allocate_for_instance() {{(pid=61957) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 713.683895] env[61957]: DEBUG nova.policy [None req-ca1e8d1e-8c18-4652-924a-ab1cdb751a05 tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2908dce4d5134e158bbed442a09c1df9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '21351f437ea54c9580753dfddd7f47be', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61957) authorize /opt/stack/nova/nova/policy.py:203}} [ 713.808900] env[61957]: DEBUG oslo_concurrency.lockutils [None req-456a04ab-263c-4c55-898a-70a189651712 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] Releasing lock "refresh_cache-6241fb44-ce1c-4881-bc32-7d19ddc83a61" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 713.809425] env[61957]: DEBUG nova.compute.manager [None req-456a04ab-263c-4c55-898a-70a189651712 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] Start destroying the instance on the hypervisor. {{(pid=61957) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 713.809634] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-456a04ab-263c-4c55-898a-70a189651712 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] Destroying instance {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 713.809943] env[61957]: DEBUG oslo_concurrency.lockutils [req-f0f061db-6270-4da7-94e4-248c6bf45880 req-9769aa9b-6ba9-4a9d-b263-53d0ddcad45e service nova] Acquired lock "refresh_cache-6241fb44-ce1c-4881-bc32-7d19ddc83a61" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 713.810537] env[61957]: DEBUG nova.network.neutron [req-f0f061db-6270-4da7-94e4-248c6bf45880 req-9769aa9b-6ba9-4a9d-b263-53d0ddcad45e service nova] [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] Refreshing network info cache for port a0d948b4-5d2e-4133-8e39-e685c226b80f {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 713.812099] env[61957]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ab531e6b-00ba-48f3-99b5-51c4c0ec3435 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.825758] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3b71e0b-64a7-4674-accd-dd8b749dc020 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.849715] env[61957]: WARNING nova.virt.vmwareapi.vmops [None req-456a04ab-263c-4c55-898a-70a189651712 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 6241fb44-ce1c-4881-bc32-7d19ddc83a61 could not be found. [ 713.849947] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-456a04ab-263c-4c55-898a-70a189651712 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] Instance destroyed {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 713.850147] env[61957]: INFO nova.compute.manager [None req-456a04ab-263c-4c55-898a-70a189651712 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] Took 0.04 seconds to destroy the instance on the hypervisor. [ 713.850395] env[61957]: DEBUG oslo.service.loopingcall [None req-456a04ab-263c-4c55-898a-70a189651712 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 713.850642] env[61957]: DEBUG nova.compute.manager [-] [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 713.850779] env[61957]: DEBUG nova.network.neutron [-] [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 713.870870] env[61957]: DEBUG nova.network.neutron [-] [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 713.980873] env[61957]: DEBUG nova.network.neutron [None req-ca1e8d1e-8c18-4652-924a-ab1cdb751a05 tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] [instance: 825130d4-b1f6-434e-be3e-dca952abe930] Successfully created port: b95b009e-bbe6-4159-bed7-0705d27090ee {{(pid=61957) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 714.054081] env[61957]: INFO nova.compute.manager [-] [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] Took 1.02 seconds to deallocate network for instance. [ 714.056824] env[61957]: DEBUG nova.compute.claims [None req-691c932a-6913-4bb6-98f2-4142c7d131a8 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] Aborting claim: {{(pid=61957) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 714.057013] env[61957]: DEBUG oslo_concurrency.lockutils [None req-691c932a-6913-4bb6-98f2-4142c7d131a8 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 714.124092] env[61957]: DEBUG nova.compute.manager [None req-ca1e8d1e-8c18-4652-924a-ab1cdb751a05 tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] [instance: 825130d4-b1f6-434e-be3e-dca952abe930] Start building block device mappings for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 714.342368] env[61957]: DEBUG nova.network.neutron [req-f0f061db-6270-4da7-94e4-248c6bf45880 req-9769aa9b-6ba9-4a9d-b263-53d0ddcad45e service nova] [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 714.373334] env[61957]: DEBUG nova.network.neutron [-] [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 714.462874] env[61957]: DEBUG nova.network.neutron [req-f0f061db-6270-4da7-94e4-248c6bf45880 req-9769aa9b-6ba9-4a9d-b263-53d0ddcad45e service nova] [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 714.617710] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8cf15e0-640b-4472-9e3c-04f82d77f0cb {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.628929] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a930fb7a-cb2b-4d7c-b196-1c9b0c928525 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.670029] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47681c52-168f-4e6b-9bd5-cb9702da294a {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.678715] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fb82573-4ba4-41ec-a91b-1bf2cba53965 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.693707] env[61957]: DEBUG nova.compute.provider_tree [None req-58fe0400-f1d0-4e1d-a459-7541ba7aeeed tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 714.807996] env[61957]: DEBUG nova.compute.manager [req-3052e092-daf0-47e6-9e3a-4e9da07a6179 req-66112e71-24e0-4f0e-b127-a1cb4528230d service nova] [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] Received event network-vif-deleted-a0d948b4-5d2e-4133-8e39-e685c226b80f {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 714.876043] env[61957]: INFO nova.compute.manager [-] [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] Took 1.03 seconds to deallocate network for instance. [ 714.880467] env[61957]: DEBUG nova.compute.claims [None req-456a04ab-263c-4c55-898a-70a189651712 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] Aborting claim: {{(pid=61957) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 714.880467] env[61957]: DEBUG oslo_concurrency.lockutils [None req-456a04ab-263c-4c55-898a-70a189651712 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 714.965338] env[61957]: DEBUG oslo_concurrency.lockutils [req-f0f061db-6270-4da7-94e4-248c6bf45880 req-9769aa9b-6ba9-4a9d-b263-53d0ddcad45e service nova] Releasing lock "refresh_cache-6241fb44-ce1c-4881-bc32-7d19ddc83a61" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 715.137103] env[61957]: ERROR nova.compute.manager [None req-ca1e8d1e-8c18-4652-924a-ab1cdb751a05 tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port b95b009e-bbe6-4159-bed7-0705d27090ee, please check neutron logs for more information. [ 715.137103] env[61957]: ERROR nova.compute.manager Traceback (most recent call last): [ 715.137103] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 715.137103] env[61957]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 715.137103] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 715.137103] env[61957]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 715.137103] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 715.137103] env[61957]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 715.137103] env[61957]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 715.137103] env[61957]: ERROR nova.compute.manager self.force_reraise() [ 715.137103] env[61957]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 715.137103] env[61957]: ERROR nova.compute.manager raise self.value [ 715.137103] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 715.137103] env[61957]: ERROR nova.compute.manager updated_port = self._update_port( [ 715.137103] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 715.137103] env[61957]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 715.137561] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 715.137561] env[61957]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 715.137561] env[61957]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port b95b009e-bbe6-4159-bed7-0705d27090ee, please check neutron logs for more information. [ 715.137561] env[61957]: ERROR nova.compute.manager [ 715.137561] env[61957]: Traceback (most recent call last): [ 715.137561] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 715.137561] env[61957]: listener.cb(fileno) [ 715.137561] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 715.137561] env[61957]: result = function(*args, **kwargs) [ 715.137561] env[61957]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 715.137561] env[61957]: return func(*args, **kwargs) [ 715.137561] env[61957]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 715.137561] env[61957]: raise e [ 715.137561] env[61957]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 715.137561] env[61957]: nwinfo = self.network_api.allocate_for_instance( [ 715.137561] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 715.137561] env[61957]: created_port_ids = self._update_ports_for_instance( [ 715.137561] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 715.137561] env[61957]: with excutils.save_and_reraise_exception(): [ 715.137561] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 715.137561] env[61957]: self.force_reraise() [ 715.137561] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 715.137561] env[61957]: raise self.value [ 715.137561] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 715.137561] env[61957]: updated_port = self._update_port( [ 715.137561] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 715.137561] env[61957]: _ensure_no_port_binding_failure(port) [ 715.137561] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 715.137561] env[61957]: raise exception.PortBindingFailed(port_id=port['id']) [ 715.138465] env[61957]: nova.exception.PortBindingFailed: Binding failed for port b95b009e-bbe6-4159-bed7-0705d27090ee, please check neutron logs for more information. [ 715.138465] env[61957]: Removing descriptor: 16 [ 715.140052] env[61957]: DEBUG nova.compute.manager [None req-ca1e8d1e-8c18-4652-924a-ab1cdb751a05 tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] [instance: 825130d4-b1f6-434e-be3e-dca952abe930] Start spawning the instance on the hypervisor. {{(pid=61957) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 715.168164] env[61957]: DEBUG nova.virt.hardware [None req-ca1e8d1e-8c18-4652-924a-ab1cdb751a05 tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T17:22:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T17:21:50Z,direct_url=,disk_format='vmdk',id=11c76a2c-f705-470a-ba9d-4657858bab38,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='81d3096d0f094373913fc3dc8f5c3c6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T17:21:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 715.168449] env[61957]: DEBUG nova.virt.hardware [None req-ca1e8d1e-8c18-4652-924a-ab1cdb751a05 tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] Flavor limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 715.168609] env[61957]: DEBUG nova.virt.hardware [None req-ca1e8d1e-8c18-4652-924a-ab1cdb751a05 tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] Image limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 715.168789] env[61957]: DEBUG nova.virt.hardware [None req-ca1e8d1e-8c18-4652-924a-ab1cdb751a05 tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] Flavor pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 715.168931] env[61957]: DEBUG nova.virt.hardware [None req-ca1e8d1e-8c18-4652-924a-ab1cdb751a05 tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] Image pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 715.169085] env[61957]: DEBUG nova.virt.hardware [None req-ca1e8d1e-8c18-4652-924a-ab1cdb751a05 tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 715.169309] env[61957]: DEBUG nova.virt.hardware [None req-ca1e8d1e-8c18-4652-924a-ab1cdb751a05 tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 715.169502] env[61957]: DEBUG nova.virt.hardware [None req-ca1e8d1e-8c18-4652-924a-ab1cdb751a05 tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 715.169679] env[61957]: DEBUG nova.virt.hardware [None req-ca1e8d1e-8c18-4652-924a-ab1cdb751a05 tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] Got 1 possible topologies {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 715.169841] env[61957]: DEBUG nova.virt.hardware [None req-ca1e8d1e-8c18-4652-924a-ab1cdb751a05 tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 715.170016] env[61957]: DEBUG nova.virt.hardware [None req-ca1e8d1e-8c18-4652-924a-ab1cdb751a05 tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 715.170887] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b91b44eb-3101-46eb-85f0-41977f8041e3 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.179613] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6cad717-44f7-44c3-9fb4-174410f29e15 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.193126] env[61957]: ERROR nova.compute.manager [None req-ca1e8d1e-8c18-4652-924a-ab1cdb751a05 tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] [instance: 825130d4-b1f6-434e-be3e-dca952abe930] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port b95b009e-bbe6-4159-bed7-0705d27090ee, please check neutron logs for more information. [ 715.193126] env[61957]: ERROR nova.compute.manager [instance: 825130d4-b1f6-434e-be3e-dca952abe930] Traceback (most recent call last): [ 715.193126] env[61957]: ERROR nova.compute.manager [instance: 825130d4-b1f6-434e-be3e-dca952abe930] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 715.193126] env[61957]: ERROR nova.compute.manager [instance: 825130d4-b1f6-434e-be3e-dca952abe930] yield resources [ 715.193126] env[61957]: ERROR nova.compute.manager [instance: 825130d4-b1f6-434e-be3e-dca952abe930] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 715.193126] env[61957]: ERROR nova.compute.manager [instance: 825130d4-b1f6-434e-be3e-dca952abe930] self.driver.spawn(context, instance, image_meta, [ 715.193126] env[61957]: ERROR nova.compute.manager [instance: 825130d4-b1f6-434e-be3e-dca952abe930] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 715.193126] env[61957]: ERROR nova.compute.manager [instance: 825130d4-b1f6-434e-be3e-dca952abe930] self._vmops.spawn(context, instance, image_meta, injected_files, [ 715.193126] env[61957]: ERROR nova.compute.manager [instance: 825130d4-b1f6-434e-be3e-dca952abe930] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 715.193126] env[61957]: ERROR nova.compute.manager [instance: 825130d4-b1f6-434e-be3e-dca952abe930] vm_ref = self.build_virtual_machine(instance, [ 715.193126] env[61957]: ERROR nova.compute.manager [instance: 825130d4-b1f6-434e-be3e-dca952abe930] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 715.193675] env[61957]: ERROR nova.compute.manager [instance: 825130d4-b1f6-434e-be3e-dca952abe930] vif_infos = vmwarevif.get_vif_info(self._session, [ 715.193675] env[61957]: ERROR nova.compute.manager [instance: 825130d4-b1f6-434e-be3e-dca952abe930] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 715.193675] env[61957]: ERROR nova.compute.manager [instance: 825130d4-b1f6-434e-be3e-dca952abe930] for vif in network_info: [ 715.193675] env[61957]: ERROR nova.compute.manager [instance: 825130d4-b1f6-434e-be3e-dca952abe930] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 715.193675] env[61957]: ERROR nova.compute.manager [instance: 825130d4-b1f6-434e-be3e-dca952abe930] return self._sync_wrapper(fn, *args, **kwargs) [ 715.193675] env[61957]: ERROR nova.compute.manager [instance: 825130d4-b1f6-434e-be3e-dca952abe930] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 715.193675] env[61957]: ERROR nova.compute.manager [instance: 825130d4-b1f6-434e-be3e-dca952abe930] self.wait() [ 715.193675] env[61957]: ERROR nova.compute.manager [instance: 825130d4-b1f6-434e-be3e-dca952abe930] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 715.193675] env[61957]: ERROR nova.compute.manager [instance: 825130d4-b1f6-434e-be3e-dca952abe930] self[:] = self._gt.wait() [ 715.193675] env[61957]: ERROR nova.compute.manager [instance: 825130d4-b1f6-434e-be3e-dca952abe930] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 715.193675] env[61957]: ERROR nova.compute.manager [instance: 825130d4-b1f6-434e-be3e-dca952abe930] return self._exit_event.wait() [ 715.193675] env[61957]: ERROR nova.compute.manager [instance: 825130d4-b1f6-434e-be3e-dca952abe930] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 715.193675] env[61957]: ERROR nova.compute.manager [instance: 825130d4-b1f6-434e-be3e-dca952abe930] current.throw(*self._exc) [ 715.194334] env[61957]: ERROR nova.compute.manager [instance: 825130d4-b1f6-434e-be3e-dca952abe930] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 715.194334] env[61957]: ERROR nova.compute.manager [instance: 825130d4-b1f6-434e-be3e-dca952abe930] result = function(*args, **kwargs) [ 715.194334] env[61957]: ERROR nova.compute.manager [instance: 825130d4-b1f6-434e-be3e-dca952abe930] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 715.194334] env[61957]: ERROR nova.compute.manager [instance: 825130d4-b1f6-434e-be3e-dca952abe930] return func(*args, **kwargs) [ 715.194334] env[61957]: ERROR nova.compute.manager [instance: 825130d4-b1f6-434e-be3e-dca952abe930] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 715.194334] env[61957]: ERROR nova.compute.manager [instance: 825130d4-b1f6-434e-be3e-dca952abe930] raise e [ 715.194334] env[61957]: ERROR nova.compute.manager [instance: 825130d4-b1f6-434e-be3e-dca952abe930] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 715.194334] env[61957]: ERROR nova.compute.manager [instance: 825130d4-b1f6-434e-be3e-dca952abe930] nwinfo = self.network_api.allocate_for_instance( [ 715.194334] env[61957]: ERROR nova.compute.manager [instance: 825130d4-b1f6-434e-be3e-dca952abe930] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 715.194334] env[61957]: ERROR nova.compute.manager [instance: 825130d4-b1f6-434e-be3e-dca952abe930] created_port_ids = self._update_ports_for_instance( [ 715.194334] env[61957]: ERROR nova.compute.manager [instance: 825130d4-b1f6-434e-be3e-dca952abe930] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 715.194334] env[61957]: ERROR nova.compute.manager [instance: 825130d4-b1f6-434e-be3e-dca952abe930] with excutils.save_and_reraise_exception(): [ 715.194334] env[61957]: ERROR nova.compute.manager [instance: 825130d4-b1f6-434e-be3e-dca952abe930] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 715.194919] env[61957]: ERROR nova.compute.manager [instance: 825130d4-b1f6-434e-be3e-dca952abe930] self.force_reraise() [ 715.194919] env[61957]: ERROR nova.compute.manager [instance: 825130d4-b1f6-434e-be3e-dca952abe930] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 715.194919] env[61957]: ERROR nova.compute.manager [instance: 825130d4-b1f6-434e-be3e-dca952abe930] raise self.value [ 715.194919] env[61957]: ERROR nova.compute.manager [instance: 825130d4-b1f6-434e-be3e-dca952abe930] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 715.194919] env[61957]: ERROR nova.compute.manager [instance: 825130d4-b1f6-434e-be3e-dca952abe930] updated_port = self._update_port( [ 715.194919] env[61957]: ERROR nova.compute.manager [instance: 825130d4-b1f6-434e-be3e-dca952abe930] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 715.194919] env[61957]: ERROR nova.compute.manager [instance: 825130d4-b1f6-434e-be3e-dca952abe930] _ensure_no_port_binding_failure(port) [ 715.194919] env[61957]: ERROR nova.compute.manager [instance: 825130d4-b1f6-434e-be3e-dca952abe930] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 715.194919] env[61957]: ERROR nova.compute.manager [instance: 825130d4-b1f6-434e-be3e-dca952abe930] raise exception.PortBindingFailed(port_id=port['id']) [ 715.194919] env[61957]: ERROR nova.compute.manager [instance: 825130d4-b1f6-434e-be3e-dca952abe930] nova.exception.PortBindingFailed: Binding failed for port b95b009e-bbe6-4159-bed7-0705d27090ee, please check neutron logs for more information. [ 715.194919] env[61957]: ERROR nova.compute.manager [instance: 825130d4-b1f6-434e-be3e-dca952abe930] [ 715.194919] env[61957]: INFO nova.compute.manager [None req-ca1e8d1e-8c18-4652-924a-ab1cdb751a05 tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] [instance: 825130d4-b1f6-434e-be3e-dca952abe930] Terminating instance [ 715.196832] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ca1e8d1e-8c18-4652-924a-ab1cdb751a05 tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] Acquiring lock "refresh_cache-825130d4-b1f6-434e-be3e-dca952abe930" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 715.196981] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ca1e8d1e-8c18-4652-924a-ab1cdb751a05 tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] Acquired lock "refresh_cache-825130d4-b1f6-434e-be3e-dca952abe930" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 715.197945] env[61957]: DEBUG nova.network.neutron [None req-ca1e8d1e-8c18-4652-924a-ab1cdb751a05 tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] [instance: 825130d4-b1f6-434e-be3e-dca952abe930] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 715.200756] env[61957]: DEBUG nova.scheduler.client.report [None req-58fe0400-f1d0-4e1d-a459-7541ba7aeeed tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 715.708832] env[61957]: DEBUG oslo_concurrency.lockutils [None req-58fe0400-f1d0-4e1d-a459-7541ba7aeeed tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.605s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 715.709487] env[61957]: DEBUG nova.compute.manager [None req-58fe0400-f1d0-4e1d-a459-7541ba7aeeed tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] [instance: c369d148-2093-4724-82cc-d80a67131fea] Start building networks asynchronously for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 715.712462] env[61957]: DEBUG oslo_concurrency.lockutils [None req-eb81388f-07fa-4a13-ac24-ee048434d451 tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 21.903s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 715.716020] env[61957]: DEBUG nova.objects.instance [None req-eb81388f-07fa-4a13-ac24-ee048434d451 tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] [instance: 188b0fd3-5d71-4feb-aca5-75a2bd28895a] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61957) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 715.730785] env[61957]: DEBUG nova.network.neutron [None req-ca1e8d1e-8c18-4652-924a-ab1cdb751a05 tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] [instance: 825130d4-b1f6-434e-be3e-dca952abe930] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 715.826208] env[61957]: DEBUG nova.network.neutron [None req-ca1e8d1e-8c18-4652-924a-ab1cdb751a05 tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] [instance: 825130d4-b1f6-434e-be3e-dca952abe930] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 716.154276] env[61957]: DEBUG oslo_concurrency.lockutils [None req-36e797a2-06a5-481b-9dd9-66f4ec9dadf3 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Acquiring lock "14ba830d-4e2c-4e9a-a059-3c86209f0127" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 716.154508] env[61957]: DEBUG oslo_concurrency.lockutils [None req-36e797a2-06a5-481b-9dd9-66f4ec9dadf3 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Lock "14ba830d-4e2c-4e9a-a059-3c86209f0127" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 716.217905] env[61957]: DEBUG nova.compute.utils [None req-58fe0400-f1d0-4e1d-a459-7541ba7aeeed tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] Using /dev/sd instead of None {{(pid=61957) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 716.221800] env[61957]: DEBUG nova.compute.manager [None req-58fe0400-f1d0-4e1d-a459-7541ba7aeeed tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] [instance: c369d148-2093-4724-82cc-d80a67131fea] Allocating IP information in the background. {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 716.221962] env[61957]: DEBUG nova.network.neutron [None req-58fe0400-f1d0-4e1d-a459-7541ba7aeeed tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] [instance: c369d148-2093-4724-82cc-d80a67131fea] allocate_for_instance() {{(pid=61957) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 716.262153] env[61957]: DEBUG nova.policy [None req-58fe0400-f1d0-4e1d-a459-7541ba7aeeed tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ed78767457b4494c88b3bd747ff643e9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f602bc62cde24f68931b0fc1948ec17c', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61957) authorize /opt/stack/nova/nova/policy.py:203}} [ 716.328852] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ca1e8d1e-8c18-4652-924a-ab1cdb751a05 tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] Releasing lock "refresh_cache-825130d4-b1f6-434e-be3e-dca952abe930" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 716.329267] env[61957]: DEBUG nova.compute.manager [None req-ca1e8d1e-8c18-4652-924a-ab1cdb751a05 tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] [instance: 825130d4-b1f6-434e-be3e-dca952abe930] Start destroying the instance on the hypervisor. {{(pid=61957) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 716.329458] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-ca1e8d1e-8c18-4652-924a-ab1cdb751a05 tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] [instance: 825130d4-b1f6-434e-be3e-dca952abe930] Destroying instance {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 716.329752] env[61957]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4c6835c7-2aaf-46ff-9bfa-15c93d65b38a {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.339511] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b34a832-48c2-4ea7-ae6d-bc0987663edb {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.360474] env[61957]: WARNING nova.virt.vmwareapi.vmops [None req-ca1e8d1e-8c18-4652-924a-ab1cdb751a05 tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] [instance: 825130d4-b1f6-434e-be3e-dca952abe930] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 825130d4-b1f6-434e-be3e-dca952abe930 could not be found. [ 716.360679] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-ca1e8d1e-8c18-4652-924a-ab1cdb751a05 tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] [instance: 825130d4-b1f6-434e-be3e-dca952abe930] Instance destroyed {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 716.360858] env[61957]: INFO nova.compute.manager [None req-ca1e8d1e-8c18-4652-924a-ab1cdb751a05 tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] [instance: 825130d4-b1f6-434e-be3e-dca952abe930] Took 0.03 seconds to destroy the instance on the hypervisor. [ 716.361105] env[61957]: DEBUG oslo.service.loopingcall [None req-ca1e8d1e-8c18-4652-924a-ab1cdb751a05 tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 716.361624] env[61957]: DEBUG nova.compute.manager [-] [instance: 825130d4-b1f6-434e-be3e-dca952abe930] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 716.361728] env[61957]: DEBUG nova.network.neutron [-] [instance: 825130d4-b1f6-434e-be3e-dca952abe930] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 716.380622] env[61957]: DEBUG nova.network.neutron [-] [instance: 825130d4-b1f6-434e-be3e-dca952abe930] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 716.541224] env[61957]: DEBUG nova.network.neutron [None req-58fe0400-f1d0-4e1d-a459-7541ba7aeeed tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] [instance: c369d148-2093-4724-82cc-d80a67131fea] Successfully created port: 2fe40c60-9f72-4975-a0f4-fe5d0e4b3b72 {{(pid=61957) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 716.726240] env[61957]: DEBUG oslo_concurrency.lockutils [None req-eb81388f-07fa-4a13-ac24-ee048434d451 tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.013s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 716.727631] env[61957]: DEBUG nova.compute.manager [None req-58fe0400-f1d0-4e1d-a459-7541ba7aeeed tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] [instance: c369d148-2093-4724-82cc-d80a67131fea] Start building block device mappings for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 716.730642] env[61957]: DEBUG oslo_concurrency.lockutils [None req-74fa7fa3-fb2c-4064-80f1-1405089fca10 tempest-ServerMetadataNegativeTestJSON-1373515003 tempest-ServerMetadataNegativeTestJSON-1373515003-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.229s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 716.830184] env[61957]: DEBUG nova.compute.manager [req-7dd9c0e6-f8e9-48c4-a558-7a7782787866 req-4cf7c2f6-b6dd-4133-aebf-06bffb7724e5 service nova] [instance: 825130d4-b1f6-434e-be3e-dca952abe930] Received event network-changed-b95b009e-bbe6-4159-bed7-0705d27090ee {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 716.830344] env[61957]: DEBUG nova.compute.manager [req-7dd9c0e6-f8e9-48c4-a558-7a7782787866 req-4cf7c2f6-b6dd-4133-aebf-06bffb7724e5 service nova] [instance: 825130d4-b1f6-434e-be3e-dca952abe930] Refreshing instance network info cache due to event network-changed-b95b009e-bbe6-4159-bed7-0705d27090ee. {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 716.830697] env[61957]: DEBUG oslo_concurrency.lockutils [req-7dd9c0e6-f8e9-48c4-a558-7a7782787866 req-4cf7c2f6-b6dd-4133-aebf-06bffb7724e5 service nova] Acquiring lock "refresh_cache-825130d4-b1f6-434e-be3e-dca952abe930" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 716.830924] env[61957]: DEBUG oslo_concurrency.lockutils [req-7dd9c0e6-f8e9-48c4-a558-7a7782787866 req-4cf7c2f6-b6dd-4133-aebf-06bffb7724e5 service nova] Acquired lock "refresh_cache-825130d4-b1f6-434e-be3e-dca952abe930" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 716.833180] env[61957]: DEBUG nova.network.neutron [req-7dd9c0e6-f8e9-48c4-a558-7a7782787866 req-4cf7c2f6-b6dd-4133-aebf-06bffb7724e5 service nova] [instance: 825130d4-b1f6-434e-be3e-dca952abe930] Refreshing network info cache for port b95b009e-bbe6-4159-bed7-0705d27090ee {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 716.883443] env[61957]: DEBUG nova.network.neutron [-] [instance: 825130d4-b1f6-434e-be3e-dca952abe930] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 717.355155] env[61957]: DEBUG nova.network.neutron [req-7dd9c0e6-f8e9-48c4-a558-7a7782787866 req-4cf7c2f6-b6dd-4133-aebf-06bffb7724e5 service nova] [instance: 825130d4-b1f6-434e-be3e-dca952abe930] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 717.387395] env[61957]: INFO nova.compute.manager [-] [instance: 825130d4-b1f6-434e-be3e-dca952abe930] Took 1.03 seconds to deallocate network for instance. [ 717.390838] env[61957]: DEBUG nova.compute.claims [None req-ca1e8d1e-8c18-4652-924a-ab1cdb751a05 tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] [instance: 825130d4-b1f6-434e-be3e-dca952abe930] Aborting claim: {{(pid=61957) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 717.390982] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ca1e8d1e-8c18-4652-924a-ab1cdb751a05 tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 717.473821] env[61957]: ERROR nova.compute.manager [None req-58fe0400-f1d0-4e1d-a459-7541ba7aeeed tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 2fe40c60-9f72-4975-a0f4-fe5d0e4b3b72, please check neutron logs for more information. [ 717.473821] env[61957]: ERROR nova.compute.manager Traceback (most recent call last): [ 717.473821] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 717.473821] env[61957]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 717.473821] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 717.473821] env[61957]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 717.473821] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 717.473821] env[61957]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 717.473821] env[61957]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 717.473821] env[61957]: ERROR nova.compute.manager self.force_reraise() [ 717.473821] env[61957]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 717.473821] env[61957]: ERROR nova.compute.manager raise self.value [ 717.473821] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 717.473821] env[61957]: ERROR nova.compute.manager updated_port = self._update_port( [ 717.473821] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 717.473821] env[61957]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 717.476784] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 717.476784] env[61957]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 717.476784] env[61957]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 2fe40c60-9f72-4975-a0f4-fe5d0e4b3b72, please check neutron logs for more information. [ 717.476784] env[61957]: ERROR nova.compute.manager [ 717.476784] env[61957]: Traceback (most recent call last): [ 717.476784] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 717.476784] env[61957]: listener.cb(fileno) [ 717.476784] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 717.476784] env[61957]: result = function(*args, **kwargs) [ 717.476784] env[61957]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 717.476784] env[61957]: return func(*args, **kwargs) [ 717.476784] env[61957]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 717.476784] env[61957]: raise e [ 717.476784] env[61957]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 717.476784] env[61957]: nwinfo = self.network_api.allocate_for_instance( [ 717.476784] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 717.476784] env[61957]: created_port_ids = self._update_ports_for_instance( [ 717.476784] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 717.476784] env[61957]: with excutils.save_and_reraise_exception(): [ 717.476784] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 717.476784] env[61957]: self.force_reraise() [ 717.476784] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 717.476784] env[61957]: raise self.value [ 717.476784] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 717.476784] env[61957]: updated_port = self._update_port( [ 717.476784] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 717.476784] env[61957]: _ensure_no_port_binding_failure(port) [ 717.476784] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 717.476784] env[61957]: raise exception.PortBindingFailed(port_id=port['id']) [ 717.477779] env[61957]: nova.exception.PortBindingFailed: Binding failed for port 2fe40c60-9f72-4975-a0f4-fe5d0e4b3b72, please check neutron logs for more information. [ 717.477779] env[61957]: Removing descriptor: 16 [ 717.477832] env[61957]: DEBUG nova.network.neutron [req-7dd9c0e6-f8e9-48c4-a558-7a7782787866 req-4cf7c2f6-b6dd-4133-aebf-06bffb7724e5 service nova] [instance: 825130d4-b1f6-434e-be3e-dca952abe930] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 717.573676] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a96797c8-2e35-4ff6-8809-05f8db23033e {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.581667] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f296c2b-53f0-45eb-81c0-31ff58618b0e {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.610348] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6c52cfb-b027-4ec5-938c-8deef2db83ec {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.617672] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5f904bf-abf9-49a7-85dd-633cb22d3712 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.630852] env[61957]: DEBUG nova.compute.provider_tree [None req-74fa7fa3-fb2c-4064-80f1-1405089fca10 tempest-ServerMetadataNegativeTestJSON-1373515003 tempest-ServerMetadataNegativeTestJSON-1373515003-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 717.746980] env[61957]: DEBUG nova.compute.manager [None req-58fe0400-f1d0-4e1d-a459-7541ba7aeeed tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] [instance: c369d148-2093-4724-82cc-d80a67131fea] Start spawning the instance on the hypervisor. {{(pid=61957) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 717.771945] env[61957]: DEBUG nova.virt.hardware [None req-58fe0400-f1d0-4e1d-a459-7541ba7aeeed tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T17:22:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T17:21:50Z,direct_url=,disk_format='vmdk',id=11c76a2c-f705-470a-ba9d-4657858bab38,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='81d3096d0f094373913fc3dc8f5c3c6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T17:21:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 717.772228] env[61957]: DEBUG nova.virt.hardware [None req-58fe0400-f1d0-4e1d-a459-7541ba7aeeed tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] Flavor limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 717.772386] env[61957]: DEBUG nova.virt.hardware [None req-58fe0400-f1d0-4e1d-a459-7541ba7aeeed tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] Image limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 717.772568] env[61957]: DEBUG nova.virt.hardware [None req-58fe0400-f1d0-4e1d-a459-7541ba7aeeed tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] Flavor pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 717.772748] env[61957]: DEBUG nova.virt.hardware [None req-58fe0400-f1d0-4e1d-a459-7541ba7aeeed tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] Image pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 717.772904] env[61957]: DEBUG nova.virt.hardware [None req-58fe0400-f1d0-4e1d-a459-7541ba7aeeed tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 717.773116] env[61957]: DEBUG nova.virt.hardware [None req-58fe0400-f1d0-4e1d-a459-7541ba7aeeed tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 717.773274] env[61957]: DEBUG nova.virt.hardware [None req-58fe0400-f1d0-4e1d-a459-7541ba7aeeed tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 717.773434] env[61957]: DEBUG nova.virt.hardware [None req-58fe0400-f1d0-4e1d-a459-7541ba7aeeed tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] Got 1 possible topologies {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 717.773590] env[61957]: DEBUG nova.virt.hardware [None req-58fe0400-f1d0-4e1d-a459-7541ba7aeeed tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 717.773755] env[61957]: DEBUG nova.virt.hardware [None req-58fe0400-f1d0-4e1d-a459-7541ba7aeeed tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 717.774686] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7def63fb-8b8e-431e-88e8-634bc066af85 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.782492] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-721d7b19-05db-4e33-9d30-1e2ada9e23c7 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.797591] env[61957]: ERROR nova.compute.manager [None req-58fe0400-f1d0-4e1d-a459-7541ba7aeeed tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] [instance: c369d148-2093-4724-82cc-d80a67131fea] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 2fe40c60-9f72-4975-a0f4-fe5d0e4b3b72, please check neutron logs for more information. [ 717.797591] env[61957]: ERROR nova.compute.manager [instance: c369d148-2093-4724-82cc-d80a67131fea] Traceback (most recent call last): [ 717.797591] env[61957]: ERROR nova.compute.manager [instance: c369d148-2093-4724-82cc-d80a67131fea] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 717.797591] env[61957]: ERROR nova.compute.manager [instance: c369d148-2093-4724-82cc-d80a67131fea] yield resources [ 717.797591] env[61957]: ERROR nova.compute.manager [instance: c369d148-2093-4724-82cc-d80a67131fea] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 717.797591] env[61957]: ERROR nova.compute.manager [instance: c369d148-2093-4724-82cc-d80a67131fea] self.driver.spawn(context, instance, image_meta, [ 717.797591] env[61957]: ERROR nova.compute.manager [instance: c369d148-2093-4724-82cc-d80a67131fea] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 717.797591] env[61957]: ERROR nova.compute.manager [instance: c369d148-2093-4724-82cc-d80a67131fea] self._vmops.spawn(context, instance, image_meta, injected_files, [ 717.797591] env[61957]: ERROR nova.compute.manager [instance: c369d148-2093-4724-82cc-d80a67131fea] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 717.797591] env[61957]: ERROR nova.compute.manager [instance: c369d148-2093-4724-82cc-d80a67131fea] vm_ref = self.build_virtual_machine(instance, [ 717.797591] env[61957]: ERROR nova.compute.manager [instance: c369d148-2093-4724-82cc-d80a67131fea] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 717.797927] env[61957]: ERROR nova.compute.manager [instance: c369d148-2093-4724-82cc-d80a67131fea] vif_infos = vmwarevif.get_vif_info(self._session, [ 717.797927] env[61957]: ERROR nova.compute.manager [instance: c369d148-2093-4724-82cc-d80a67131fea] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 717.797927] env[61957]: ERROR nova.compute.manager [instance: c369d148-2093-4724-82cc-d80a67131fea] for vif in network_info: [ 717.797927] env[61957]: ERROR nova.compute.manager [instance: c369d148-2093-4724-82cc-d80a67131fea] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 717.797927] env[61957]: ERROR nova.compute.manager [instance: c369d148-2093-4724-82cc-d80a67131fea] return self._sync_wrapper(fn, *args, **kwargs) [ 717.797927] env[61957]: ERROR nova.compute.manager [instance: c369d148-2093-4724-82cc-d80a67131fea] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 717.797927] env[61957]: ERROR nova.compute.manager [instance: c369d148-2093-4724-82cc-d80a67131fea] self.wait() [ 717.797927] env[61957]: ERROR nova.compute.manager [instance: c369d148-2093-4724-82cc-d80a67131fea] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 717.797927] env[61957]: ERROR nova.compute.manager [instance: c369d148-2093-4724-82cc-d80a67131fea] self[:] = self._gt.wait() [ 717.797927] env[61957]: ERROR nova.compute.manager [instance: c369d148-2093-4724-82cc-d80a67131fea] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 717.797927] env[61957]: ERROR nova.compute.manager [instance: c369d148-2093-4724-82cc-d80a67131fea] return self._exit_event.wait() [ 717.797927] env[61957]: ERROR nova.compute.manager [instance: c369d148-2093-4724-82cc-d80a67131fea] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 717.797927] env[61957]: ERROR nova.compute.manager [instance: c369d148-2093-4724-82cc-d80a67131fea] current.throw(*self._exc) [ 717.798258] env[61957]: ERROR nova.compute.manager [instance: c369d148-2093-4724-82cc-d80a67131fea] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 717.798258] env[61957]: ERROR nova.compute.manager [instance: c369d148-2093-4724-82cc-d80a67131fea] result = function(*args, **kwargs) [ 717.798258] env[61957]: ERROR nova.compute.manager [instance: c369d148-2093-4724-82cc-d80a67131fea] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 717.798258] env[61957]: ERROR nova.compute.manager [instance: c369d148-2093-4724-82cc-d80a67131fea] return func(*args, **kwargs) [ 717.798258] env[61957]: ERROR nova.compute.manager [instance: c369d148-2093-4724-82cc-d80a67131fea] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 717.798258] env[61957]: ERROR nova.compute.manager [instance: c369d148-2093-4724-82cc-d80a67131fea] raise e [ 717.798258] env[61957]: ERROR nova.compute.manager [instance: c369d148-2093-4724-82cc-d80a67131fea] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 717.798258] env[61957]: ERROR nova.compute.manager [instance: c369d148-2093-4724-82cc-d80a67131fea] nwinfo = self.network_api.allocate_for_instance( [ 717.798258] env[61957]: ERROR nova.compute.manager [instance: c369d148-2093-4724-82cc-d80a67131fea] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 717.798258] env[61957]: ERROR nova.compute.manager [instance: c369d148-2093-4724-82cc-d80a67131fea] created_port_ids = self._update_ports_for_instance( [ 717.798258] env[61957]: ERROR nova.compute.manager [instance: c369d148-2093-4724-82cc-d80a67131fea] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 717.798258] env[61957]: ERROR nova.compute.manager [instance: c369d148-2093-4724-82cc-d80a67131fea] with excutils.save_and_reraise_exception(): [ 717.798258] env[61957]: ERROR nova.compute.manager [instance: c369d148-2093-4724-82cc-d80a67131fea] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 717.798645] env[61957]: ERROR nova.compute.manager [instance: c369d148-2093-4724-82cc-d80a67131fea] self.force_reraise() [ 717.798645] env[61957]: ERROR nova.compute.manager [instance: c369d148-2093-4724-82cc-d80a67131fea] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 717.798645] env[61957]: ERROR nova.compute.manager [instance: c369d148-2093-4724-82cc-d80a67131fea] raise self.value [ 717.798645] env[61957]: ERROR nova.compute.manager [instance: c369d148-2093-4724-82cc-d80a67131fea] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 717.798645] env[61957]: ERROR nova.compute.manager [instance: c369d148-2093-4724-82cc-d80a67131fea] updated_port = self._update_port( [ 717.798645] env[61957]: ERROR nova.compute.manager [instance: c369d148-2093-4724-82cc-d80a67131fea] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 717.798645] env[61957]: ERROR nova.compute.manager [instance: c369d148-2093-4724-82cc-d80a67131fea] _ensure_no_port_binding_failure(port) [ 717.798645] env[61957]: ERROR nova.compute.manager [instance: c369d148-2093-4724-82cc-d80a67131fea] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 717.798645] env[61957]: ERROR nova.compute.manager [instance: c369d148-2093-4724-82cc-d80a67131fea] raise exception.PortBindingFailed(port_id=port['id']) [ 717.798645] env[61957]: ERROR nova.compute.manager [instance: c369d148-2093-4724-82cc-d80a67131fea] nova.exception.PortBindingFailed: Binding failed for port 2fe40c60-9f72-4975-a0f4-fe5d0e4b3b72, please check neutron logs for more information. [ 717.798645] env[61957]: ERROR nova.compute.manager [instance: c369d148-2093-4724-82cc-d80a67131fea] [ 717.798645] env[61957]: INFO nova.compute.manager [None req-58fe0400-f1d0-4e1d-a459-7541ba7aeeed tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] [instance: c369d148-2093-4724-82cc-d80a67131fea] Terminating instance [ 717.799864] env[61957]: DEBUG oslo_concurrency.lockutils [None req-58fe0400-f1d0-4e1d-a459-7541ba7aeeed tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] Acquiring lock "refresh_cache-c369d148-2093-4724-82cc-d80a67131fea" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 717.800026] env[61957]: DEBUG oslo_concurrency.lockutils [None req-58fe0400-f1d0-4e1d-a459-7541ba7aeeed tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] Acquired lock "refresh_cache-c369d148-2093-4724-82cc-d80a67131fea" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 717.800193] env[61957]: DEBUG nova.network.neutron [None req-58fe0400-f1d0-4e1d-a459-7541ba7aeeed tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] [instance: c369d148-2093-4724-82cc-d80a67131fea] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 717.980775] env[61957]: DEBUG oslo_concurrency.lockutils [req-7dd9c0e6-f8e9-48c4-a558-7a7782787866 req-4cf7c2f6-b6dd-4133-aebf-06bffb7724e5 service nova] Releasing lock "refresh_cache-825130d4-b1f6-434e-be3e-dca952abe930" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 717.981269] env[61957]: DEBUG nova.compute.manager [req-7dd9c0e6-f8e9-48c4-a558-7a7782787866 req-4cf7c2f6-b6dd-4133-aebf-06bffb7724e5 service nova] [instance: 825130d4-b1f6-434e-be3e-dca952abe930] Received event network-vif-deleted-b95b009e-bbe6-4159-bed7-0705d27090ee {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 718.134320] env[61957]: DEBUG nova.scheduler.client.report [None req-74fa7fa3-fb2c-4064-80f1-1405089fca10 tempest-ServerMetadataNegativeTestJSON-1373515003 tempest-ServerMetadataNegativeTestJSON-1373515003-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 718.317194] env[61957]: DEBUG nova.network.neutron [None req-58fe0400-f1d0-4e1d-a459-7541ba7aeeed tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] [instance: c369d148-2093-4724-82cc-d80a67131fea] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 718.400892] env[61957]: DEBUG nova.network.neutron [None req-58fe0400-f1d0-4e1d-a459-7541ba7aeeed tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] [instance: c369d148-2093-4724-82cc-d80a67131fea] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 718.641058] env[61957]: DEBUG oslo_concurrency.lockutils [None req-74fa7fa3-fb2c-4064-80f1-1405089fca10 tempest-ServerMetadataNegativeTestJSON-1373515003 tempest-ServerMetadataNegativeTestJSON-1373515003-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.909s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 718.641058] env[61957]: ERROR nova.compute.manager [None req-74fa7fa3-fb2c-4064-80f1-1405089fca10 tempest-ServerMetadataNegativeTestJSON-1373515003 tempest-ServerMetadataNegativeTestJSON-1373515003-project-member] [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port ddfc7e7f-b45c-426a-8485-f28a208b52e8, please check neutron logs for more information. [ 718.641058] env[61957]: ERROR nova.compute.manager [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] Traceback (most recent call last): [ 718.641058] env[61957]: ERROR nova.compute.manager [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 718.641058] env[61957]: ERROR nova.compute.manager [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] self.driver.spawn(context, instance, image_meta, [ 718.641058] env[61957]: ERROR nova.compute.manager [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 718.641058] env[61957]: ERROR nova.compute.manager [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] self._vmops.spawn(context, instance, image_meta, injected_files, [ 718.641058] env[61957]: ERROR nova.compute.manager [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 718.641058] env[61957]: ERROR nova.compute.manager [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] vm_ref = self.build_virtual_machine(instance, [ 718.641368] env[61957]: ERROR nova.compute.manager [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 718.641368] env[61957]: ERROR nova.compute.manager [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] vif_infos = vmwarevif.get_vif_info(self._session, [ 718.641368] env[61957]: ERROR nova.compute.manager [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 718.641368] env[61957]: ERROR nova.compute.manager [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] for vif in network_info: [ 718.641368] env[61957]: ERROR nova.compute.manager [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 718.641368] env[61957]: ERROR nova.compute.manager [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] return self._sync_wrapper(fn, *args, **kwargs) [ 718.641368] env[61957]: ERROR nova.compute.manager [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 718.641368] env[61957]: ERROR nova.compute.manager [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] self.wait() [ 718.641368] env[61957]: ERROR nova.compute.manager [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 718.641368] env[61957]: ERROR nova.compute.manager [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] self[:] = self._gt.wait() [ 718.641368] env[61957]: ERROR nova.compute.manager [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 718.641368] env[61957]: ERROR nova.compute.manager [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] return self._exit_event.wait() [ 718.641368] env[61957]: ERROR nova.compute.manager [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 718.641679] env[61957]: ERROR nova.compute.manager [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] current.throw(*self._exc) [ 718.641679] env[61957]: ERROR nova.compute.manager [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 718.641679] env[61957]: ERROR nova.compute.manager [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] result = function(*args, **kwargs) [ 718.641679] env[61957]: ERROR nova.compute.manager [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 718.641679] env[61957]: ERROR nova.compute.manager [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] return func(*args, **kwargs) [ 718.641679] env[61957]: ERROR nova.compute.manager [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 718.641679] env[61957]: ERROR nova.compute.manager [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] raise e [ 718.641679] env[61957]: ERROR nova.compute.manager [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 718.641679] env[61957]: ERROR nova.compute.manager [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] nwinfo = self.network_api.allocate_for_instance( [ 718.641679] env[61957]: ERROR nova.compute.manager [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 718.641679] env[61957]: ERROR nova.compute.manager [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] created_port_ids = self._update_ports_for_instance( [ 718.641679] env[61957]: ERROR nova.compute.manager [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 718.641679] env[61957]: ERROR nova.compute.manager [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] with excutils.save_and_reraise_exception(): [ 718.641993] env[61957]: ERROR nova.compute.manager [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 718.641993] env[61957]: ERROR nova.compute.manager [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] self.force_reraise() [ 718.641993] env[61957]: ERROR nova.compute.manager [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 718.641993] env[61957]: ERROR nova.compute.manager [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] raise self.value [ 718.641993] env[61957]: ERROR nova.compute.manager [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 718.641993] env[61957]: ERROR nova.compute.manager [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] updated_port = self._update_port( [ 718.641993] env[61957]: ERROR nova.compute.manager [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 718.641993] env[61957]: ERROR nova.compute.manager [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] _ensure_no_port_binding_failure(port) [ 718.641993] env[61957]: ERROR nova.compute.manager [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 718.641993] env[61957]: ERROR nova.compute.manager [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] raise exception.PortBindingFailed(port_id=port['id']) [ 718.641993] env[61957]: ERROR nova.compute.manager [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] nova.exception.PortBindingFailed: Binding failed for port ddfc7e7f-b45c-426a-8485-f28a208b52e8, please check neutron logs for more information. [ 718.641993] env[61957]: ERROR nova.compute.manager [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] [ 718.642296] env[61957]: DEBUG nova.compute.utils [None req-74fa7fa3-fb2c-4064-80f1-1405089fca10 tempest-ServerMetadataNegativeTestJSON-1373515003 tempest-ServerMetadataNegativeTestJSON-1373515003-project-member] [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] Binding failed for port ddfc7e7f-b45c-426a-8485-f28a208b52e8, please check neutron logs for more information. {{(pid=61957) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 718.642873] env[61957]: DEBUG oslo_concurrency.lockutils [None req-707bfa11-3857-45e5-a4be-581ffade0e04 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.119s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 718.644442] env[61957]: INFO nova.compute.claims [None req-707bfa11-3857-45e5-a4be-581ffade0e04 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 718.647154] env[61957]: DEBUG nova.compute.manager [None req-74fa7fa3-fb2c-4064-80f1-1405089fca10 tempest-ServerMetadataNegativeTestJSON-1373515003 tempest-ServerMetadataNegativeTestJSON-1373515003-project-member] [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] Build of instance 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1 was re-scheduled: Binding failed for port ddfc7e7f-b45c-426a-8485-f28a208b52e8, please check neutron logs for more information. {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 718.647609] env[61957]: DEBUG nova.compute.manager [None req-74fa7fa3-fb2c-4064-80f1-1405089fca10 tempest-ServerMetadataNegativeTestJSON-1373515003 tempest-ServerMetadataNegativeTestJSON-1373515003-project-member] [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] Unplugging VIFs for instance {{(pid=61957) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 718.647860] env[61957]: DEBUG oslo_concurrency.lockutils [None req-74fa7fa3-fb2c-4064-80f1-1405089fca10 tempest-ServerMetadataNegativeTestJSON-1373515003 tempest-ServerMetadataNegativeTestJSON-1373515003-project-member] Acquiring lock "refresh_cache-3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 718.648038] env[61957]: DEBUG oslo_concurrency.lockutils [None req-74fa7fa3-fb2c-4064-80f1-1405089fca10 tempest-ServerMetadataNegativeTestJSON-1373515003 tempest-ServerMetadataNegativeTestJSON-1373515003-project-member] Acquired lock "refresh_cache-3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 718.648230] env[61957]: DEBUG nova.network.neutron [None req-74fa7fa3-fb2c-4064-80f1-1405089fca10 tempest-ServerMetadataNegativeTestJSON-1373515003 tempest-ServerMetadataNegativeTestJSON-1373515003-project-member] [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 718.856655] env[61957]: DEBUG nova.compute.manager [req-7cbe4396-2c30-4991-8f92-ffd90885cef1 req-9c78c36e-9742-49b3-b483-52e24e51fcaf service nova] [instance: c369d148-2093-4724-82cc-d80a67131fea] Received event network-changed-2fe40c60-9f72-4975-a0f4-fe5d0e4b3b72 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 718.857021] env[61957]: DEBUG nova.compute.manager [req-7cbe4396-2c30-4991-8f92-ffd90885cef1 req-9c78c36e-9742-49b3-b483-52e24e51fcaf service nova] [instance: c369d148-2093-4724-82cc-d80a67131fea] Refreshing instance network info cache due to event network-changed-2fe40c60-9f72-4975-a0f4-fe5d0e4b3b72. {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 718.857021] env[61957]: DEBUG oslo_concurrency.lockutils [req-7cbe4396-2c30-4991-8f92-ffd90885cef1 req-9c78c36e-9742-49b3-b483-52e24e51fcaf service nova] Acquiring lock "refresh_cache-c369d148-2093-4724-82cc-d80a67131fea" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 718.902644] env[61957]: DEBUG oslo_concurrency.lockutils [None req-58fe0400-f1d0-4e1d-a459-7541ba7aeeed tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] Releasing lock "refresh_cache-c369d148-2093-4724-82cc-d80a67131fea" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 718.903081] env[61957]: DEBUG nova.compute.manager [None req-58fe0400-f1d0-4e1d-a459-7541ba7aeeed tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] [instance: c369d148-2093-4724-82cc-d80a67131fea] Start destroying the instance on the hypervisor. {{(pid=61957) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 718.903272] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-58fe0400-f1d0-4e1d-a459-7541ba7aeeed tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] [instance: c369d148-2093-4724-82cc-d80a67131fea] Destroying instance {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 718.903566] env[61957]: DEBUG oslo_concurrency.lockutils [req-7cbe4396-2c30-4991-8f92-ffd90885cef1 req-9c78c36e-9742-49b3-b483-52e24e51fcaf service nova] Acquired lock "refresh_cache-c369d148-2093-4724-82cc-d80a67131fea" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 718.903730] env[61957]: DEBUG nova.network.neutron [req-7cbe4396-2c30-4991-8f92-ffd90885cef1 req-9c78c36e-9742-49b3-b483-52e24e51fcaf service nova] [instance: c369d148-2093-4724-82cc-d80a67131fea] Refreshing network info cache for port 2fe40c60-9f72-4975-a0f4-fe5d0e4b3b72 {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 718.904741] env[61957]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-df29fb1b-7e5f-4411-b518-cbed9f9892cc {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.914260] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6da587a3-8129-4ddf-a964-5e842f913e11 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.935509] env[61957]: WARNING nova.virt.vmwareapi.vmops [None req-58fe0400-f1d0-4e1d-a459-7541ba7aeeed tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] [instance: c369d148-2093-4724-82cc-d80a67131fea] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance c369d148-2093-4724-82cc-d80a67131fea could not be found. [ 718.935688] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-58fe0400-f1d0-4e1d-a459-7541ba7aeeed tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] [instance: c369d148-2093-4724-82cc-d80a67131fea] Instance destroyed {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 718.935858] env[61957]: INFO nova.compute.manager [None req-58fe0400-f1d0-4e1d-a459-7541ba7aeeed tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] [instance: c369d148-2093-4724-82cc-d80a67131fea] Took 0.03 seconds to destroy the instance on the hypervisor. [ 718.936096] env[61957]: DEBUG oslo.service.loopingcall [None req-58fe0400-f1d0-4e1d-a459-7541ba7aeeed tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 718.936336] env[61957]: DEBUG nova.compute.manager [-] [instance: c369d148-2093-4724-82cc-d80a67131fea] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 718.936437] env[61957]: DEBUG nova.network.neutron [-] [instance: c369d148-2093-4724-82cc-d80a67131fea] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 718.952793] env[61957]: DEBUG nova.network.neutron [-] [instance: c369d148-2093-4724-82cc-d80a67131fea] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 719.167927] env[61957]: DEBUG nova.network.neutron [None req-74fa7fa3-fb2c-4064-80f1-1405089fca10 tempest-ServerMetadataNegativeTestJSON-1373515003 tempest-ServerMetadataNegativeTestJSON-1373515003-project-member] [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 719.240596] env[61957]: DEBUG nova.network.neutron [None req-74fa7fa3-fb2c-4064-80f1-1405089fca10 tempest-ServerMetadataNegativeTestJSON-1373515003 tempest-ServerMetadataNegativeTestJSON-1373515003-project-member] [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 719.424183] env[61957]: DEBUG nova.network.neutron [req-7cbe4396-2c30-4991-8f92-ffd90885cef1 req-9c78c36e-9742-49b3-b483-52e24e51fcaf service nova] [instance: c369d148-2093-4724-82cc-d80a67131fea] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 719.455719] env[61957]: DEBUG nova.network.neutron [-] [instance: c369d148-2093-4724-82cc-d80a67131fea] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 719.505661] env[61957]: DEBUG nova.network.neutron [req-7cbe4396-2c30-4991-8f92-ffd90885cef1 req-9c78c36e-9742-49b3-b483-52e24e51fcaf service nova] [instance: c369d148-2093-4724-82cc-d80a67131fea] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 719.744856] env[61957]: DEBUG oslo_concurrency.lockutils [None req-74fa7fa3-fb2c-4064-80f1-1405089fca10 tempest-ServerMetadataNegativeTestJSON-1373515003 tempest-ServerMetadataNegativeTestJSON-1373515003-project-member] Releasing lock "refresh_cache-3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 719.745112] env[61957]: DEBUG nova.compute.manager [None req-74fa7fa3-fb2c-4064-80f1-1405089fca10 tempest-ServerMetadataNegativeTestJSON-1373515003 tempest-ServerMetadataNegativeTestJSON-1373515003-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61957) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 719.745296] env[61957]: DEBUG nova.compute.manager [None req-74fa7fa3-fb2c-4064-80f1-1405089fca10 tempest-ServerMetadataNegativeTestJSON-1373515003 tempest-ServerMetadataNegativeTestJSON-1373515003-project-member] [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 719.745476] env[61957]: DEBUG nova.network.neutron [None req-74fa7fa3-fb2c-4064-80f1-1405089fca10 tempest-ServerMetadataNegativeTestJSON-1373515003 tempest-ServerMetadataNegativeTestJSON-1373515003-project-member] [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 719.762849] env[61957]: DEBUG nova.network.neutron [None req-74fa7fa3-fb2c-4064-80f1-1405089fca10 tempest-ServerMetadataNegativeTestJSON-1373515003 tempest-ServerMetadataNegativeTestJSON-1373515003-project-member] [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 719.959017] env[61957]: INFO nova.compute.manager [-] [instance: c369d148-2093-4724-82cc-d80a67131fea] Took 1.02 seconds to deallocate network for instance. [ 719.964082] env[61957]: DEBUG nova.compute.claims [None req-58fe0400-f1d0-4e1d-a459-7541ba7aeeed tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] [instance: c369d148-2093-4724-82cc-d80a67131fea] Aborting claim: {{(pid=61957) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 719.964271] env[61957]: DEBUG oslo_concurrency.lockutils [None req-58fe0400-f1d0-4e1d-a459-7541ba7aeeed tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 720.009083] env[61957]: DEBUG oslo_concurrency.lockutils [req-7cbe4396-2c30-4991-8f92-ffd90885cef1 req-9c78c36e-9742-49b3-b483-52e24e51fcaf service nova] Releasing lock "refresh_cache-c369d148-2093-4724-82cc-d80a67131fea" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 720.009371] env[61957]: DEBUG nova.compute.manager [req-7cbe4396-2c30-4991-8f92-ffd90885cef1 req-9c78c36e-9742-49b3-b483-52e24e51fcaf service nova] [instance: c369d148-2093-4724-82cc-d80a67131fea] Received event network-vif-deleted-2fe40c60-9f72-4975-a0f4-fe5d0e4b3b72 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 720.033398] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e42b232-0cdc-4c36-bf60-cc1e8c827c4d {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.042164] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4897f7b4-c6f0-4f68-a893-e4f4b6379ad3 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.072628] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcf2ec9e-93ac-432f-a1d6-004f6a3fe717 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.080240] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3c51eae-6731-4ccc-b4db-a6957b8bc6d1 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.093804] env[61957]: DEBUG nova.compute.provider_tree [None req-707bfa11-3857-45e5-a4be-581ffade0e04 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 720.265698] env[61957]: DEBUG nova.network.neutron [None req-74fa7fa3-fb2c-4064-80f1-1405089fca10 tempest-ServerMetadataNegativeTestJSON-1373515003 tempest-ServerMetadataNegativeTestJSON-1373515003-project-member] [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 720.598054] env[61957]: DEBUG nova.scheduler.client.report [None req-707bfa11-3857-45e5-a4be-581ffade0e04 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 720.770592] env[61957]: INFO nova.compute.manager [None req-74fa7fa3-fb2c-4064-80f1-1405089fca10 tempest-ServerMetadataNegativeTestJSON-1373515003 tempest-ServerMetadataNegativeTestJSON-1373515003-project-member] [instance: 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1] Took 1.02 seconds to deallocate network for instance. [ 721.102398] env[61957]: DEBUG oslo_concurrency.lockutils [None req-707bfa11-3857-45e5-a4be-581ffade0e04 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.459s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 721.103072] env[61957]: DEBUG nova.compute.manager [None req-707bfa11-3857-45e5-a4be-581ffade0e04 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] Start building networks asynchronously for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 721.105669] env[61957]: DEBUG oslo_concurrency.lockutils [None req-2d3c8bae-3f1e-41b1-82a9-b293bdd893bd tempest-ServersAdmin275Test-2089451214 tempest-ServersAdmin275Test-2089451214-project-admin] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 17.736s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 721.105850] env[61957]: DEBUG nova.objects.instance [None req-2d3c8bae-3f1e-41b1-82a9-b293bdd893bd tempest-ServersAdmin275Test-2089451214 tempest-ServersAdmin275Test-2089451214-project-admin] [instance: 188b0fd3-5d71-4feb-aca5-75a2bd28895a] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61957) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 721.612855] env[61957]: DEBUG nova.compute.utils [None req-707bfa11-3857-45e5-a4be-581ffade0e04 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Using /dev/sd instead of None {{(pid=61957) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 721.612855] env[61957]: DEBUG nova.compute.manager [None req-707bfa11-3857-45e5-a4be-581ffade0e04 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] Allocating IP information in the background. {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 721.612855] env[61957]: DEBUG nova.network.neutron [None req-707bfa11-3857-45e5-a4be-581ffade0e04 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] allocate_for_instance() {{(pid=61957) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 721.656183] env[61957]: DEBUG nova.policy [None req-707bfa11-3857-45e5-a4be-581ffade0e04 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '76e1754f96ed438ea0ed91ad337419f1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '975189ebb3cc4cdb9391880f0c9ba6ba', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61957) authorize /opt/stack/nova/nova/policy.py:203}} [ 721.797205] env[61957]: INFO nova.scheduler.client.report [None req-74fa7fa3-fb2c-4064-80f1-1405089fca10 tempest-ServerMetadataNegativeTestJSON-1373515003 tempest-ServerMetadataNegativeTestJSON-1373515003-project-member] Deleted allocations for instance 3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1 [ 721.900052] env[61957]: DEBUG nova.network.neutron [None req-707bfa11-3857-45e5-a4be-581ffade0e04 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] Successfully created port: 9392bc30-2697-41c7-83dd-b8963c4dc6c3 {{(pid=61957) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 722.119963] env[61957]: DEBUG oslo_concurrency.lockutils [None req-2d3c8bae-3f1e-41b1-82a9-b293bdd893bd tempest-ServersAdmin275Test-2089451214 tempest-ServersAdmin275Test-2089451214-project-admin] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.014s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 722.121241] env[61957]: DEBUG nova.compute.manager [None req-707bfa11-3857-45e5-a4be-581ffade0e04 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] Start building block device mappings for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 722.130017] env[61957]: DEBUG oslo_concurrency.lockutils [None req-dbfaab8a-0784-4441-ad61-79e56d9c3c60 tempest-FloatingIPsAssociationTestJSON-1786153562 tempest-FloatingIPsAssociationTestJSON-1786153562-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.794s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 722.305335] env[61957]: DEBUG oslo_concurrency.lockutils [None req-74fa7fa3-fb2c-4064-80f1-1405089fca10 tempest-ServerMetadataNegativeTestJSON-1373515003 tempest-ServerMetadataNegativeTestJSON-1373515003-project-member] Lock "3ee7a6f2-aa1c-46a4-9e1b-8af55bdf94d1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 115.165s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 722.741191] env[61957]: DEBUG nova.compute.manager [req-08af5ca9-bccb-4340-b4a5-65051c05011d req-5bc310f5-3c38-49ab-8edd-7e8a47a6ea69 service nova] [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] Received event network-changed-9392bc30-2697-41c7-83dd-b8963c4dc6c3 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 722.741395] env[61957]: DEBUG nova.compute.manager [req-08af5ca9-bccb-4340-b4a5-65051c05011d req-5bc310f5-3c38-49ab-8edd-7e8a47a6ea69 service nova] [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] Refreshing instance network info cache due to event network-changed-9392bc30-2697-41c7-83dd-b8963c4dc6c3. {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 722.741605] env[61957]: DEBUG oslo_concurrency.lockutils [req-08af5ca9-bccb-4340-b4a5-65051c05011d req-5bc310f5-3c38-49ab-8edd-7e8a47a6ea69 service nova] Acquiring lock "refresh_cache-2252ef6e-19ec-4356-8221-33168c01bef0" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 722.741744] env[61957]: DEBUG oslo_concurrency.lockutils [req-08af5ca9-bccb-4340-b4a5-65051c05011d req-5bc310f5-3c38-49ab-8edd-7e8a47a6ea69 service nova] Acquired lock "refresh_cache-2252ef6e-19ec-4356-8221-33168c01bef0" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 722.742122] env[61957]: DEBUG nova.network.neutron [req-08af5ca9-bccb-4340-b4a5-65051c05011d req-5bc310f5-3c38-49ab-8edd-7e8a47a6ea69 service nova] [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] Refreshing network info cache for port 9392bc30-2697-41c7-83dd-b8963c4dc6c3 {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 722.773528] env[61957]: ERROR nova.compute.manager [None req-707bfa11-3857-45e5-a4be-581ffade0e04 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 9392bc30-2697-41c7-83dd-b8963c4dc6c3, please check neutron logs for more information. [ 722.773528] env[61957]: ERROR nova.compute.manager Traceback (most recent call last): [ 722.773528] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 722.773528] env[61957]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 722.773528] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 722.773528] env[61957]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 722.773528] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 722.773528] env[61957]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 722.773528] env[61957]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 722.773528] env[61957]: ERROR nova.compute.manager self.force_reraise() [ 722.773528] env[61957]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 722.773528] env[61957]: ERROR nova.compute.manager raise self.value [ 722.773528] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 722.773528] env[61957]: ERROR nova.compute.manager updated_port = self._update_port( [ 722.773528] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 722.773528] env[61957]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 722.774010] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 722.774010] env[61957]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 722.774010] env[61957]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 9392bc30-2697-41c7-83dd-b8963c4dc6c3, please check neutron logs for more information. [ 722.774010] env[61957]: ERROR nova.compute.manager [ 722.774010] env[61957]: Traceback (most recent call last): [ 722.774010] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 722.774010] env[61957]: listener.cb(fileno) [ 722.774010] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 722.774010] env[61957]: result = function(*args, **kwargs) [ 722.774010] env[61957]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 722.774010] env[61957]: return func(*args, **kwargs) [ 722.774010] env[61957]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 722.774010] env[61957]: raise e [ 722.774010] env[61957]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 722.774010] env[61957]: nwinfo = self.network_api.allocate_for_instance( [ 722.774010] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 722.774010] env[61957]: created_port_ids = self._update_ports_for_instance( [ 722.774010] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 722.774010] env[61957]: with excutils.save_and_reraise_exception(): [ 722.774010] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 722.774010] env[61957]: self.force_reraise() [ 722.774010] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 722.774010] env[61957]: raise self.value [ 722.774010] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 722.774010] env[61957]: updated_port = self._update_port( [ 722.774010] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 722.774010] env[61957]: _ensure_no_port_binding_failure(port) [ 722.774010] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 722.774010] env[61957]: raise exception.PortBindingFailed(port_id=port['id']) [ 722.774887] env[61957]: nova.exception.PortBindingFailed: Binding failed for port 9392bc30-2697-41c7-83dd-b8963c4dc6c3, please check neutron logs for more information. [ 722.774887] env[61957]: Removing descriptor: 17 [ 722.807349] env[61957]: DEBUG nova.compute.manager [None req-5ca5fdda-dbef-4c48-b38f-9d2aa78e9841 tempest-FloatingIPsAssociationNegativeTestJSON-572957506 tempest-FloatingIPsAssociationNegativeTestJSON-572957506-project-member] [instance: 348ea580-2268-4082-a5f1-33c954063e4d] Starting instance... {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 722.984754] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-342bf800-f30b-4874-8fec-d2d4abb0efa2 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.993403] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7049a1ca-f9f9-4a8f-bc75-20f6a3e1b3d9 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.026366] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c2a1744-0485-4632-8d8f-0cc71036d923 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.034265] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec848a9f-23f4-4549-a145-a28697082c56 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.047826] env[61957]: DEBUG nova.compute.provider_tree [None req-dbfaab8a-0784-4441-ad61-79e56d9c3c60 tempest-FloatingIPsAssociationTestJSON-1786153562 tempest-FloatingIPsAssociationTestJSON-1786153562-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 723.140022] env[61957]: DEBUG nova.compute.manager [None req-707bfa11-3857-45e5-a4be-581ffade0e04 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] Start spawning the instance on the hypervisor. {{(pid=61957) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 723.165925] env[61957]: DEBUG nova.virt.hardware [None req-707bfa11-3857-45e5-a4be-581ffade0e04 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T17:22:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T17:21:50Z,direct_url=,disk_format='vmdk',id=11c76a2c-f705-470a-ba9d-4657858bab38,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='81d3096d0f094373913fc3dc8f5c3c6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T17:21:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 723.166222] env[61957]: DEBUG nova.virt.hardware [None req-707bfa11-3857-45e5-a4be-581ffade0e04 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Flavor limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 723.166357] env[61957]: DEBUG nova.virt.hardware [None req-707bfa11-3857-45e5-a4be-581ffade0e04 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Image limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 723.166558] env[61957]: DEBUG nova.virt.hardware [None req-707bfa11-3857-45e5-a4be-581ffade0e04 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Flavor pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 723.166704] env[61957]: DEBUG nova.virt.hardware [None req-707bfa11-3857-45e5-a4be-581ffade0e04 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Image pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 723.166846] env[61957]: DEBUG nova.virt.hardware [None req-707bfa11-3857-45e5-a4be-581ffade0e04 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 723.167076] env[61957]: DEBUG nova.virt.hardware [None req-707bfa11-3857-45e5-a4be-581ffade0e04 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 723.167228] env[61957]: DEBUG nova.virt.hardware [None req-707bfa11-3857-45e5-a4be-581ffade0e04 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 723.167390] env[61957]: DEBUG nova.virt.hardware [None req-707bfa11-3857-45e5-a4be-581ffade0e04 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Got 1 possible topologies {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 723.167557] env[61957]: DEBUG nova.virt.hardware [None req-707bfa11-3857-45e5-a4be-581ffade0e04 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 723.167720] env[61957]: DEBUG nova.virt.hardware [None req-707bfa11-3857-45e5-a4be-581ffade0e04 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 723.168661] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91fddfe6-df86-4f6b-b80a-f259883cb5c0 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.176386] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3cd15358-02c7-422f-907c-7e0251da63d4 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.190731] env[61957]: ERROR nova.compute.manager [None req-707bfa11-3857-45e5-a4be-581ffade0e04 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 9392bc30-2697-41c7-83dd-b8963c4dc6c3, please check neutron logs for more information. [ 723.190731] env[61957]: ERROR nova.compute.manager [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] Traceback (most recent call last): [ 723.190731] env[61957]: ERROR nova.compute.manager [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 723.190731] env[61957]: ERROR nova.compute.manager [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] yield resources [ 723.190731] env[61957]: ERROR nova.compute.manager [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 723.190731] env[61957]: ERROR nova.compute.manager [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] self.driver.spawn(context, instance, image_meta, [ 723.190731] env[61957]: ERROR nova.compute.manager [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 723.190731] env[61957]: ERROR nova.compute.manager [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] self._vmops.spawn(context, instance, image_meta, injected_files, [ 723.190731] env[61957]: ERROR nova.compute.manager [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 723.190731] env[61957]: ERROR nova.compute.manager [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] vm_ref = self.build_virtual_machine(instance, [ 723.190731] env[61957]: ERROR nova.compute.manager [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 723.191732] env[61957]: ERROR nova.compute.manager [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] vif_infos = vmwarevif.get_vif_info(self._session, [ 723.191732] env[61957]: ERROR nova.compute.manager [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 723.191732] env[61957]: ERROR nova.compute.manager [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] for vif in network_info: [ 723.191732] env[61957]: ERROR nova.compute.manager [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 723.191732] env[61957]: ERROR nova.compute.manager [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] return self._sync_wrapper(fn, *args, **kwargs) [ 723.191732] env[61957]: ERROR nova.compute.manager [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 723.191732] env[61957]: ERROR nova.compute.manager [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] self.wait() [ 723.191732] env[61957]: ERROR nova.compute.manager [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 723.191732] env[61957]: ERROR nova.compute.manager [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] self[:] = self._gt.wait() [ 723.191732] env[61957]: ERROR nova.compute.manager [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 723.191732] env[61957]: ERROR nova.compute.manager [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] return self._exit_event.wait() [ 723.191732] env[61957]: ERROR nova.compute.manager [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 723.191732] env[61957]: ERROR nova.compute.manager [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] current.throw(*self._exc) [ 723.192606] env[61957]: ERROR nova.compute.manager [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 723.192606] env[61957]: ERROR nova.compute.manager [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] result = function(*args, **kwargs) [ 723.192606] env[61957]: ERROR nova.compute.manager [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 723.192606] env[61957]: ERROR nova.compute.manager [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] return func(*args, **kwargs) [ 723.192606] env[61957]: ERROR nova.compute.manager [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 723.192606] env[61957]: ERROR nova.compute.manager [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] raise e [ 723.192606] env[61957]: ERROR nova.compute.manager [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 723.192606] env[61957]: ERROR nova.compute.manager [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] nwinfo = self.network_api.allocate_for_instance( [ 723.192606] env[61957]: ERROR nova.compute.manager [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 723.192606] env[61957]: ERROR nova.compute.manager [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] created_port_ids = self._update_ports_for_instance( [ 723.192606] env[61957]: ERROR nova.compute.manager [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 723.192606] env[61957]: ERROR nova.compute.manager [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] with excutils.save_and_reraise_exception(): [ 723.192606] env[61957]: ERROR nova.compute.manager [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 723.193657] env[61957]: ERROR nova.compute.manager [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] self.force_reraise() [ 723.193657] env[61957]: ERROR nova.compute.manager [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 723.193657] env[61957]: ERROR nova.compute.manager [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] raise self.value [ 723.193657] env[61957]: ERROR nova.compute.manager [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 723.193657] env[61957]: ERROR nova.compute.manager [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] updated_port = self._update_port( [ 723.193657] env[61957]: ERROR nova.compute.manager [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 723.193657] env[61957]: ERROR nova.compute.manager [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] _ensure_no_port_binding_failure(port) [ 723.193657] env[61957]: ERROR nova.compute.manager [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 723.193657] env[61957]: ERROR nova.compute.manager [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] raise exception.PortBindingFailed(port_id=port['id']) [ 723.193657] env[61957]: ERROR nova.compute.manager [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] nova.exception.PortBindingFailed: Binding failed for port 9392bc30-2697-41c7-83dd-b8963c4dc6c3, please check neutron logs for more information. [ 723.193657] env[61957]: ERROR nova.compute.manager [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] [ 723.193657] env[61957]: INFO nova.compute.manager [None req-707bfa11-3857-45e5-a4be-581ffade0e04 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] Terminating instance [ 723.194608] env[61957]: DEBUG oslo_concurrency.lockutils [None req-707bfa11-3857-45e5-a4be-581ffade0e04 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Acquiring lock "refresh_cache-2252ef6e-19ec-4356-8221-33168c01bef0" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 723.266962] env[61957]: DEBUG nova.network.neutron [req-08af5ca9-bccb-4340-b4a5-65051c05011d req-5bc310f5-3c38-49ab-8edd-7e8a47a6ea69 service nova] [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 723.330259] env[61957]: DEBUG oslo_concurrency.lockutils [None req-5ca5fdda-dbef-4c48-b38f-9d2aa78e9841 tempest-FloatingIPsAssociationNegativeTestJSON-572957506 tempest-FloatingIPsAssociationNegativeTestJSON-572957506-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 723.409569] env[61957]: DEBUG nova.network.neutron [req-08af5ca9-bccb-4340-b4a5-65051c05011d req-5bc310f5-3c38-49ab-8edd-7e8a47a6ea69 service nova] [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 723.551260] env[61957]: DEBUG nova.scheduler.client.report [None req-dbfaab8a-0784-4441-ad61-79e56d9c3c60 tempest-FloatingIPsAssociationTestJSON-1786153562 tempest-FloatingIPsAssociationTestJSON-1786153562-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 723.913660] env[61957]: DEBUG oslo_concurrency.lockutils [req-08af5ca9-bccb-4340-b4a5-65051c05011d req-5bc310f5-3c38-49ab-8edd-7e8a47a6ea69 service nova] Releasing lock "refresh_cache-2252ef6e-19ec-4356-8221-33168c01bef0" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 723.913660] env[61957]: DEBUG oslo_concurrency.lockutils [None req-707bfa11-3857-45e5-a4be-581ffade0e04 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Acquired lock "refresh_cache-2252ef6e-19ec-4356-8221-33168c01bef0" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 723.913660] env[61957]: DEBUG nova.network.neutron [None req-707bfa11-3857-45e5-a4be-581ffade0e04 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 724.056164] env[61957]: DEBUG oslo_concurrency.lockutils [None req-dbfaab8a-0784-4441-ad61-79e56d9c3c60 tempest-FloatingIPsAssociationTestJSON-1786153562 tempest-FloatingIPsAssociationTestJSON-1786153562-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.929s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 724.056794] env[61957]: ERROR nova.compute.manager [None req-dbfaab8a-0784-4441-ad61-79e56d9c3c60 tempest-FloatingIPsAssociationTestJSON-1786153562 tempest-FloatingIPsAssociationTestJSON-1786153562-project-member] [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port b04626cb-e752-4764-b985-7ae70cc01fd7, please check neutron logs for more information. [ 724.056794] env[61957]: ERROR nova.compute.manager [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] Traceback (most recent call last): [ 724.056794] env[61957]: ERROR nova.compute.manager [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 724.056794] env[61957]: ERROR nova.compute.manager [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] self.driver.spawn(context, instance, image_meta, [ 724.056794] env[61957]: ERROR nova.compute.manager [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 724.056794] env[61957]: ERROR nova.compute.manager [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] self._vmops.spawn(context, instance, image_meta, injected_files, [ 724.056794] env[61957]: ERROR nova.compute.manager [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 724.056794] env[61957]: ERROR nova.compute.manager [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] vm_ref = self.build_virtual_machine(instance, [ 724.056794] env[61957]: ERROR nova.compute.manager [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 724.056794] env[61957]: ERROR nova.compute.manager [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] vif_infos = vmwarevif.get_vif_info(self._session, [ 724.056794] env[61957]: ERROR nova.compute.manager [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 724.057204] env[61957]: ERROR nova.compute.manager [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] for vif in network_info: [ 724.057204] env[61957]: ERROR nova.compute.manager [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 724.057204] env[61957]: ERROR nova.compute.manager [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] return self._sync_wrapper(fn, *args, **kwargs) [ 724.057204] env[61957]: ERROR nova.compute.manager [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 724.057204] env[61957]: ERROR nova.compute.manager [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] self.wait() [ 724.057204] env[61957]: ERROR nova.compute.manager [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 724.057204] env[61957]: ERROR nova.compute.manager [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] self[:] = self._gt.wait() [ 724.057204] env[61957]: ERROR nova.compute.manager [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 724.057204] env[61957]: ERROR nova.compute.manager [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] return self._exit_event.wait() [ 724.057204] env[61957]: ERROR nova.compute.manager [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 724.057204] env[61957]: ERROR nova.compute.manager [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] current.throw(*self._exc) [ 724.057204] env[61957]: ERROR nova.compute.manager [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 724.057204] env[61957]: ERROR nova.compute.manager [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] result = function(*args, **kwargs) [ 724.057568] env[61957]: ERROR nova.compute.manager [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 724.057568] env[61957]: ERROR nova.compute.manager [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] return func(*args, **kwargs) [ 724.057568] env[61957]: ERROR nova.compute.manager [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 724.057568] env[61957]: ERROR nova.compute.manager [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] raise e [ 724.057568] env[61957]: ERROR nova.compute.manager [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 724.057568] env[61957]: ERROR nova.compute.manager [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] nwinfo = self.network_api.allocate_for_instance( [ 724.057568] env[61957]: ERROR nova.compute.manager [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 724.057568] env[61957]: ERROR nova.compute.manager [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] created_port_ids = self._update_ports_for_instance( [ 724.057568] env[61957]: ERROR nova.compute.manager [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 724.057568] env[61957]: ERROR nova.compute.manager [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] with excutils.save_and_reraise_exception(): [ 724.057568] env[61957]: ERROR nova.compute.manager [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 724.057568] env[61957]: ERROR nova.compute.manager [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] self.force_reraise() [ 724.057568] env[61957]: ERROR nova.compute.manager [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 724.057880] env[61957]: ERROR nova.compute.manager [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] raise self.value [ 724.057880] env[61957]: ERROR nova.compute.manager [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 724.057880] env[61957]: ERROR nova.compute.manager [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] updated_port = self._update_port( [ 724.057880] env[61957]: ERROR nova.compute.manager [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 724.057880] env[61957]: ERROR nova.compute.manager [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] _ensure_no_port_binding_failure(port) [ 724.057880] env[61957]: ERROR nova.compute.manager [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 724.057880] env[61957]: ERROR nova.compute.manager [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] raise exception.PortBindingFailed(port_id=port['id']) [ 724.057880] env[61957]: ERROR nova.compute.manager [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] nova.exception.PortBindingFailed: Binding failed for port b04626cb-e752-4764-b985-7ae70cc01fd7, please check neutron logs for more information. [ 724.057880] env[61957]: ERROR nova.compute.manager [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] [ 724.057880] env[61957]: DEBUG nova.compute.utils [None req-dbfaab8a-0784-4441-ad61-79e56d9c3c60 tempest-FloatingIPsAssociationTestJSON-1786153562 tempest-FloatingIPsAssociationTestJSON-1786153562-project-member] [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] Binding failed for port b04626cb-e752-4764-b985-7ae70cc01fd7, please check neutron logs for more information. {{(pid=61957) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 724.058673] env[61957]: DEBUG oslo_concurrency.lockutils [None req-2b9492ff-1cb8-4ba9-ba5e-b513c4b108ad tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.163s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 724.060863] env[61957]: INFO nova.compute.claims [None req-2b9492ff-1cb8-4ba9-ba5e-b513c4b108ad tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 724.063459] env[61957]: DEBUG nova.compute.manager [None req-dbfaab8a-0784-4441-ad61-79e56d9c3c60 tempest-FloatingIPsAssociationTestJSON-1786153562 tempest-FloatingIPsAssociationTestJSON-1786153562-project-member] [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] Build of instance 8257a521-2c25-45a2-a2e5-c735ece03da2 was re-scheduled: Binding failed for port b04626cb-e752-4764-b985-7ae70cc01fd7, please check neutron logs for more information. {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 724.063868] env[61957]: DEBUG nova.compute.manager [None req-dbfaab8a-0784-4441-ad61-79e56d9c3c60 tempest-FloatingIPsAssociationTestJSON-1786153562 tempest-FloatingIPsAssociationTestJSON-1786153562-project-member] [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] Unplugging VIFs for instance {{(pid=61957) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 724.064092] env[61957]: DEBUG oslo_concurrency.lockutils [None req-dbfaab8a-0784-4441-ad61-79e56d9c3c60 tempest-FloatingIPsAssociationTestJSON-1786153562 tempest-FloatingIPsAssociationTestJSON-1786153562-project-member] Acquiring lock "refresh_cache-8257a521-2c25-45a2-a2e5-c735ece03da2" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 724.064229] env[61957]: DEBUG oslo_concurrency.lockutils [None req-dbfaab8a-0784-4441-ad61-79e56d9c3c60 tempest-FloatingIPsAssociationTestJSON-1786153562 tempest-FloatingIPsAssociationTestJSON-1786153562-project-member] Acquired lock "refresh_cache-8257a521-2c25-45a2-a2e5-c735ece03da2" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 724.064381] env[61957]: DEBUG nova.network.neutron [None req-dbfaab8a-0784-4441-ad61-79e56d9c3c60 tempest-FloatingIPsAssociationTestJSON-1786153562 tempest-FloatingIPsAssociationTestJSON-1786153562-project-member] [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 724.440902] env[61957]: DEBUG nova.network.neutron [None req-707bfa11-3857-45e5-a4be-581ffade0e04 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 724.559422] env[61957]: DEBUG nova.network.neutron [None req-707bfa11-3857-45e5-a4be-581ffade0e04 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 724.590165] env[61957]: DEBUG nova.network.neutron [None req-dbfaab8a-0784-4441-ad61-79e56d9c3c60 tempest-FloatingIPsAssociationTestJSON-1786153562 tempest-FloatingIPsAssociationTestJSON-1786153562-project-member] [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 724.671534] env[61957]: DEBUG nova.network.neutron [None req-dbfaab8a-0784-4441-ad61-79e56d9c3c60 tempest-FloatingIPsAssociationTestJSON-1786153562 tempest-FloatingIPsAssociationTestJSON-1786153562-project-member] [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 724.769997] env[61957]: DEBUG nova.compute.manager [req-4d64718c-2581-4c95-8db4-3d88476f4d5e req-4d9ec609-6516-40d8-b60d-bf2067f11ca9 service nova] [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] Received event network-vif-deleted-9392bc30-2697-41c7-83dd-b8963c4dc6c3 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 725.062316] env[61957]: DEBUG oslo_concurrency.lockutils [None req-707bfa11-3857-45e5-a4be-581ffade0e04 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Releasing lock "refresh_cache-2252ef6e-19ec-4356-8221-33168c01bef0" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 725.062772] env[61957]: DEBUG nova.compute.manager [None req-707bfa11-3857-45e5-a4be-581ffade0e04 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] Start destroying the instance on the hypervisor. {{(pid=61957) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 725.062974] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-707bfa11-3857-45e5-a4be-581ffade0e04 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] Destroying instance {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 725.064595] env[61957]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-79469cee-e2f9-4a50-9cba-2f4d251ff141 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.076082] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f1c27ad-ae18-4f48-bf84-8a08864acce8 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.098094] env[61957]: WARNING nova.virt.vmwareapi.vmops [None req-707bfa11-3857-45e5-a4be-581ffade0e04 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 2252ef6e-19ec-4356-8221-33168c01bef0 could not be found. [ 725.098332] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-707bfa11-3857-45e5-a4be-581ffade0e04 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] Instance destroyed {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 725.098522] env[61957]: INFO nova.compute.manager [None req-707bfa11-3857-45e5-a4be-581ffade0e04 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] Took 0.04 seconds to destroy the instance on the hypervisor. [ 725.098816] env[61957]: DEBUG oslo.service.loopingcall [None req-707bfa11-3857-45e5-a4be-581ffade0e04 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 725.099085] env[61957]: DEBUG nova.compute.manager [-] [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 725.099184] env[61957]: DEBUG nova.network.neutron [-] [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 725.117175] env[61957]: DEBUG nova.network.neutron [-] [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 725.175817] env[61957]: DEBUG oslo_concurrency.lockutils [None req-dbfaab8a-0784-4441-ad61-79e56d9c3c60 tempest-FloatingIPsAssociationTestJSON-1786153562 tempest-FloatingIPsAssociationTestJSON-1786153562-project-member] Releasing lock "refresh_cache-8257a521-2c25-45a2-a2e5-c735ece03da2" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 725.175817] env[61957]: DEBUG nova.compute.manager [None req-dbfaab8a-0784-4441-ad61-79e56d9c3c60 tempest-FloatingIPsAssociationTestJSON-1786153562 tempest-FloatingIPsAssociationTestJSON-1786153562-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61957) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 725.175817] env[61957]: DEBUG nova.compute.manager [None req-dbfaab8a-0784-4441-ad61-79e56d9c3c60 tempest-FloatingIPsAssociationTestJSON-1786153562 tempest-FloatingIPsAssociationTestJSON-1786153562-project-member] [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 725.175817] env[61957]: DEBUG nova.network.neutron [None req-dbfaab8a-0784-4441-ad61-79e56d9c3c60 tempest-FloatingIPsAssociationTestJSON-1786153562 tempest-FloatingIPsAssociationTestJSON-1786153562-project-member] [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 725.193441] env[61957]: DEBUG nova.network.neutron [None req-dbfaab8a-0784-4441-ad61-79e56d9c3c60 tempest-FloatingIPsAssociationTestJSON-1786153562 tempest-FloatingIPsAssociationTestJSON-1786153562-project-member] [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 725.410974] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a73b50f-3d86-4530-87d5-1079a53273e9 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.419589] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7023416-a9fb-450f-a874-dbc03c8d7500 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.448666] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f20d786-308d-4e08-9ee8-15a15f7f3dee {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.456091] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d48ba5a-95c6-4056-9a6b-40c5ac0eed92 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.471948] env[61957]: DEBUG nova.compute.provider_tree [None req-2b9492ff-1cb8-4ba9-ba5e-b513c4b108ad tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 725.622297] env[61957]: DEBUG nova.network.neutron [-] [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 725.695796] env[61957]: DEBUG nova.network.neutron [None req-dbfaab8a-0784-4441-ad61-79e56d9c3c60 tempest-FloatingIPsAssociationTestJSON-1786153562 tempest-FloatingIPsAssociationTestJSON-1786153562-project-member] [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 725.891971] env[61957]: DEBUG oslo_concurrency.lockutils [None req-9387ee6a-989e-4086-be03-e20e994ba789 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Acquiring lock "17f00b20-9d3b-45e6-919d-6fab9999ec77" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 725.892244] env[61957]: DEBUG oslo_concurrency.lockutils [None req-9387ee6a-989e-4086-be03-e20e994ba789 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Lock "17f00b20-9d3b-45e6-919d-6fab9999ec77" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 725.975144] env[61957]: DEBUG nova.scheduler.client.report [None req-2b9492ff-1cb8-4ba9-ba5e-b513c4b108ad tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 726.124954] env[61957]: INFO nova.compute.manager [-] [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] Took 1.03 seconds to deallocate network for instance. [ 726.127220] env[61957]: DEBUG nova.compute.claims [None req-707bfa11-3857-45e5-a4be-581ffade0e04 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] Aborting claim: {{(pid=61957) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 726.127399] env[61957]: DEBUG oslo_concurrency.lockutils [None req-707bfa11-3857-45e5-a4be-581ffade0e04 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 726.198122] env[61957]: INFO nova.compute.manager [None req-dbfaab8a-0784-4441-ad61-79e56d9c3c60 tempest-FloatingIPsAssociationTestJSON-1786153562 tempest-FloatingIPsAssociationTestJSON-1786153562-project-member] [instance: 8257a521-2c25-45a2-a2e5-c735ece03da2] Took 1.02 seconds to deallocate network for instance. [ 726.480611] env[61957]: DEBUG oslo_concurrency.lockutils [None req-2b9492ff-1cb8-4ba9-ba5e-b513c4b108ad tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.422s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 726.481250] env[61957]: DEBUG nova.compute.manager [None req-2b9492ff-1cb8-4ba9-ba5e-b513c4b108ad tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] Start building networks asynchronously for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 726.483971] env[61957]: DEBUG oslo_concurrency.lockutils [None req-4e657274-3c21-45ad-b242-d1090e30b7c2 tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 18.705s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 726.485796] env[61957]: DEBUG nova.objects.instance [None req-4e657274-3c21-45ad-b242-d1090e30b7c2 tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Lazy-loading 'resources' on Instance uuid 188b0fd3-5d71-4feb-aca5-75a2bd28895a {{(pid=61957) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 726.990719] env[61957]: DEBUG nova.compute.utils [None req-2b9492ff-1cb8-4ba9-ba5e-b513c4b108ad tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] Using /dev/sd instead of None {{(pid=61957) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 726.993055] env[61957]: DEBUG nova.compute.manager [None req-2b9492ff-1cb8-4ba9-ba5e-b513c4b108ad tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] Allocating IP information in the background. {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 726.994048] env[61957]: DEBUG nova.network.neutron [None req-2b9492ff-1cb8-4ba9-ba5e-b513c4b108ad tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] allocate_for_instance() {{(pid=61957) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 727.034402] env[61957]: DEBUG nova.policy [None req-2b9492ff-1cb8-4ba9-ba5e-b513c4b108ad tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ed78767457b4494c88b3bd747ff643e9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f602bc62cde24f68931b0fc1948ec17c', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61957) authorize /opt/stack/nova/nova/policy.py:203}} [ 727.228912] env[61957]: INFO nova.scheduler.client.report [None req-dbfaab8a-0784-4441-ad61-79e56d9c3c60 tempest-FloatingIPsAssociationTestJSON-1786153562 tempest-FloatingIPsAssociationTestJSON-1786153562-project-member] Deleted allocations for instance 8257a521-2c25-45a2-a2e5-c735ece03da2 [ 727.320604] env[61957]: DEBUG nova.network.neutron [None req-2b9492ff-1cb8-4ba9-ba5e-b513c4b108ad tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] Successfully created port: d5344278-a5e5-4fe0-ab8d-80293d0e444b {{(pid=61957) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 727.376773] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c57bbf9-232f-4794-90f3-9e0a4399acce {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.383742] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32c714ab-658f-42f7-a31d-caa1aba97c67 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.417053] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cced9708-4333-4bba-958a-1fba2b25eab1 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.425286] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fef0c090-d11c-4996-8bd9-d0b98ab00d97 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.441758] env[61957]: DEBUG nova.compute.provider_tree [None req-4e657274-3c21-45ad-b242-d1090e30b7c2 tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 727.495989] env[61957]: DEBUG nova.compute.manager [None req-2b9492ff-1cb8-4ba9-ba5e-b513c4b108ad tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] Start building block device mappings for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 727.749273] env[61957]: DEBUG oslo_concurrency.lockutils [None req-dbfaab8a-0784-4441-ad61-79e56d9c3c60 tempest-FloatingIPsAssociationTestJSON-1786153562 tempest-FloatingIPsAssociationTestJSON-1786153562-project-member] Lock "8257a521-2c25-45a2-a2e5-c735ece03da2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 117.439s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 727.945208] env[61957]: DEBUG nova.scheduler.client.report [None req-4e657274-3c21-45ad-b242-d1090e30b7c2 tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 727.995331] env[61957]: DEBUG nova.compute.manager [req-833348f1-0926-41b6-b121-e3c3fca44581 req-70a4d399-b2f3-40dc-b602-e50aaf7e1c11 service nova] [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] Received event network-changed-d5344278-a5e5-4fe0-ab8d-80293d0e444b {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 727.995535] env[61957]: DEBUG nova.compute.manager [req-833348f1-0926-41b6-b121-e3c3fca44581 req-70a4d399-b2f3-40dc-b602-e50aaf7e1c11 service nova] [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] Refreshing instance network info cache due to event network-changed-d5344278-a5e5-4fe0-ab8d-80293d0e444b. {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 727.995747] env[61957]: DEBUG oslo_concurrency.lockutils [req-833348f1-0926-41b6-b121-e3c3fca44581 req-70a4d399-b2f3-40dc-b602-e50aaf7e1c11 service nova] Acquiring lock "refresh_cache-f8235178-7ae0-4733-bb05-3515d5489a5b" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 727.995941] env[61957]: DEBUG oslo_concurrency.lockutils [req-833348f1-0926-41b6-b121-e3c3fca44581 req-70a4d399-b2f3-40dc-b602-e50aaf7e1c11 service nova] Acquired lock "refresh_cache-f8235178-7ae0-4733-bb05-3515d5489a5b" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 727.996068] env[61957]: DEBUG nova.network.neutron [req-833348f1-0926-41b6-b121-e3c3fca44581 req-70a4d399-b2f3-40dc-b602-e50aaf7e1c11 service nova] [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] Refreshing network info cache for port d5344278-a5e5-4fe0-ab8d-80293d0e444b {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 728.250677] env[61957]: DEBUG nova.compute.manager [None req-838e8ad7-4b6c-4dbc-8d4b-8380b652e559 tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] [instance: 042425a0-5398-47a6-9842-380d5b3299c0] Starting instance... {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 728.286798] env[61957]: ERROR nova.compute.manager [None req-2b9492ff-1cb8-4ba9-ba5e-b513c4b108ad tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port d5344278-a5e5-4fe0-ab8d-80293d0e444b, please check neutron logs for more information. [ 728.286798] env[61957]: ERROR nova.compute.manager Traceback (most recent call last): [ 728.286798] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 728.286798] env[61957]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 728.286798] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 728.286798] env[61957]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 728.286798] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 728.286798] env[61957]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 728.286798] env[61957]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 728.286798] env[61957]: ERROR nova.compute.manager self.force_reraise() [ 728.286798] env[61957]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 728.286798] env[61957]: ERROR nova.compute.manager raise self.value [ 728.286798] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 728.286798] env[61957]: ERROR nova.compute.manager updated_port = self._update_port( [ 728.286798] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 728.286798] env[61957]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 728.287336] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 728.287336] env[61957]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 728.287336] env[61957]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port d5344278-a5e5-4fe0-ab8d-80293d0e444b, please check neutron logs for more information. [ 728.287336] env[61957]: ERROR nova.compute.manager [ 728.287336] env[61957]: Traceback (most recent call last): [ 728.287336] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 728.287336] env[61957]: listener.cb(fileno) [ 728.287336] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 728.287336] env[61957]: result = function(*args, **kwargs) [ 728.287336] env[61957]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 728.287336] env[61957]: return func(*args, **kwargs) [ 728.287336] env[61957]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 728.287336] env[61957]: raise e [ 728.287336] env[61957]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 728.287336] env[61957]: nwinfo = self.network_api.allocate_for_instance( [ 728.287336] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 728.287336] env[61957]: created_port_ids = self._update_ports_for_instance( [ 728.287336] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 728.287336] env[61957]: with excutils.save_and_reraise_exception(): [ 728.287336] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 728.287336] env[61957]: self.force_reraise() [ 728.287336] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 728.287336] env[61957]: raise self.value [ 728.287336] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 728.287336] env[61957]: updated_port = self._update_port( [ 728.287336] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 728.287336] env[61957]: _ensure_no_port_binding_failure(port) [ 728.287336] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 728.287336] env[61957]: raise exception.PortBindingFailed(port_id=port['id']) [ 728.288218] env[61957]: nova.exception.PortBindingFailed: Binding failed for port d5344278-a5e5-4fe0-ab8d-80293d0e444b, please check neutron logs for more information. [ 728.288218] env[61957]: Removing descriptor: 17 [ 728.452795] env[61957]: DEBUG oslo_concurrency.lockutils [None req-4e657274-3c21-45ad-b242-d1090e30b7c2 tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.969s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 728.455764] env[61957]: DEBUG oslo_concurrency.lockutils [None req-c2f4fc75-b8bd-417b-8703-3b9048472a92 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 19.887s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 728.472855] env[61957]: INFO nova.scheduler.client.report [None req-4e657274-3c21-45ad-b242-d1090e30b7c2 tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Deleted allocations for instance 188b0fd3-5d71-4feb-aca5-75a2bd28895a [ 728.505029] env[61957]: DEBUG nova.compute.manager [None req-2b9492ff-1cb8-4ba9-ba5e-b513c4b108ad tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] Start spawning the instance on the hypervisor. {{(pid=61957) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 728.516260] env[61957]: DEBUG nova.network.neutron [req-833348f1-0926-41b6-b121-e3c3fca44581 req-70a4d399-b2f3-40dc-b602-e50aaf7e1c11 service nova] [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 728.531553] env[61957]: DEBUG nova.virt.hardware [None req-2b9492ff-1cb8-4ba9-ba5e-b513c4b108ad tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T17:22:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T17:21:50Z,direct_url=,disk_format='vmdk',id=11c76a2c-f705-470a-ba9d-4657858bab38,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='81d3096d0f094373913fc3dc8f5c3c6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T17:21:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 728.531798] env[61957]: DEBUG nova.virt.hardware [None req-2b9492ff-1cb8-4ba9-ba5e-b513c4b108ad tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] Flavor limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 728.531949] env[61957]: DEBUG nova.virt.hardware [None req-2b9492ff-1cb8-4ba9-ba5e-b513c4b108ad tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] Image limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 728.532148] env[61957]: DEBUG nova.virt.hardware [None req-2b9492ff-1cb8-4ba9-ba5e-b513c4b108ad tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] Flavor pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 728.532322] env[61957]: DEBUG nova.virt.hardware [None req-2b9492ff-1cb8-4ba9-ba5e-b513c4b108ad tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] Image pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 728.532515] env[61957]: DEBUG nova.virt.hardware [None req-2b9492ff-1cb8-4ba9-ba5e-b513c4b108ad tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 728.532774] env[61957]: DEBUG nova.virt.hardware [None req-2b9492ff-1cb8-4ba9-ba5e-b513c4b108ad tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 728.532975] env[61957]: DEBUG nova.virt.hardware [None req-2b9492ff-1cb8-4ba9-ba5e-b513c4b108ad tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 728.533280] env[61957]: DEBUG nova.virt.hardware [None req-2b9492ff-1cb8-4ba9-ba5e-b513c4b108ad tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] Got 1 possible topologies {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 728.533592] env[61957]: DEBUG nova.virt.hardware [None req-2b9492ff-1cb8-4ba9-ba5e-b513c4b108ad tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 728.533677] env[61957]: DEBUG nova.virt.hardware [None req-2b9492ff-1cb8-4ba9-ba5e-b513c4b108ad tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 728.534718] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79b065f7-ed3e-4e3f-a4c0-1a9a10495a80 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.543396] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb50bd9e-21af-407d-b883-335017e404aa {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.557063] env[61957]: ERROR nova.compute.manager [None req-2b9492ff-1cb8-4ba9-ba5e-b513c4b108ad tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port d5344278-a5e5-4fe0-ab8d-80293d0e444b, please check neutron logs for more information. [ 728.557063] env[61957]: ERROR nova.compute.manager [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] Traceback (most recent call last): [ 728.557063] env[61957]: ERROR nova.compute.manager [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 728.557063] env[61957]: ERROR nova.compute.manager [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] yield resources [ 728.557063] env[61957]: ERROR nova.compute.manager [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 728.557063] env[61957]: ERROR nova.compute.manager [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] self.driver.spawn(context, instance, image_meta, [ 728.557063] env[61957]: ERROR nova.compute.manager [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 728.557063] env[61957]: ERROR nova.compute.manager [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 728.557063] env[61957]: ERROR nova.compute.manager [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 728.557063] env[61957]: ERROR nova.compute.manager [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] vm_ref = self.build_virtual_machine(instance, [ 728.557063] env[61957]: ERROR nova.compute.manager [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 728.557425] env[61957]: ERROR nova.compute.manager [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] vif_infos = vmwarevif.get_vif_info(self._session, [ 728.557425] env[61957]: ERROR nova.compute.manager [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 728.557425] env[61957]: ERROR nova.compute.manager [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] for vif in network_info: [ 728.557425] env[61957]: ERROR nova.compute.manager [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 728.557425] env[61957]: ERROR nova.compute.manager [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] return self._sync_wrapper(fn, *args, **kwargs) [ 728.557425] env[61957]: ERROR nova.compute.manager [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 728.557425] env[61957]: ERROR nova.compute.manager [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] self.wait() [ 728.557425] env[61957]: ERROR nova.compute.manager [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 728.557425] env[61957]: ERROR nova.compute.manager [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] self[:] = self._gt.wait() [ 728.557425] env[61957]: ERROR nova.compute.manager [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 728.557425] env[61957]: ERROR nova.compute.manager [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] return self._exit_event.wait() [ 728.557425] env[61957]: ERROR nova.compute.manager [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 728.557425] env[61957]: ERROR nova.compute.manager [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] current.throw(*self._exc) [ 728.557848] env[61957]: ERROR nova.compute.manager [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 728.557848] env[61957]: ERROR nova.compute.manager [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] result = function(*args, **kwargs) [ 728.557848] env[61957]: ERROR nova.compute.manager [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 728.557848] env[61957]: ERROR nova.compute.manager [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] return func(*args, **kwargs) [ 728.557848] env[61957]: ERROR nova.compute.manager [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 728.557848] env[61957]: ERROR nova.compute.manager [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] raise e [ 728.557848] env[61957]: ERROR nova.compute.manager [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 728.557848] env[61957]: ERROR nova.compute.manager [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] nwinfo = self.network_api.allocate_for_instance( [ 728.557848] env[61957]: ERROR nova.compute.manager [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 728.557848] env[61957]: ERROR nova.compute.manager [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] created_port_ids = self._update_ports_for_instance( [ 728.557848] env[61957]: ERROR nova.compute.manager [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 728.557848] env[61957]: ERROR nova.compute.manager [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] with excutils.save_and_reraise_exception(): [ 728.557848] env[61957]: ERROR nova.compute.manager [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 728.558211] env[61957]: ERROR nova.compute.manager [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] self.force_reraise() [ 728.558211] env[61957]: ERROR nova.compute.manager [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 728.558211] env[61957]: ERROR nova.compute.manager [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] raise self.value [ 728.558211] env[61957]: ERROR nova.compute.manager [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 728.558211] env[61957]: ERROR nova.compute.manager [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] updated_port = self._update_port( [ 728.558211] env[61957]: ERROR nova.compute.manager [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 728.558211] env[61957]: ERROR nova.compute.manager [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] _ensure_no_port_binding_failure(port) [ 728.558211] env[61957]: ERROR nova.compute.manager [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 728.558211] env[61957]: ERROR nova.compute.manager [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] raise exception.PortBindingFailed(port_id=port['id']) [ 728.558211] env[61957]: ERROR nova.compute.manager [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] nova.exception.PortBindingFailed: Binding failed for port d5344278-a5e5-4fe0-ab8d-80293d0e444b, please check neutron logs for more information. [ 728.558211] env[61957]: ERROR nova.compute.manager [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] [ 728.558211] env[61957]: INFO nova.compute.manager [None req-2b9492ff-1cb8-4ba9-ba5e-b513c4b108ad tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] Terminating instance [ 728.560400] env[61957]: DEBUG oslo_concurrency.lockutils [None req-2b9492ff-1cb8-4ba9-ba5e-b513c4b108ad tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] Acquiring lock "refresh_cache-f8235178-7ae0-4733-bb05-3515d5489a5b" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 728.605341] env[61957]: DEBUG nova.network.neutron [req-833348f1-0926-41b6-b121-e3c3fca44581 req-70a4d399-b2f3-40dc-b602-e50aaf7e1c11 service nova] [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 728.776929] env[61957]: DEBUG oslo_concurrency.lockutils [None req-838e8ad7-4b6c-4dbc-8d4b-8380b652e559 tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 728.980405] env[61957]: DEBUG oslo_concurrency.lockutils [None req-4e657274-3c21-45ad-b242-d1090e30b7c2 tempest-ServersAdmin275Test-980003697 tempest-ServersAdmin275Test-980003697-project-member] Lock "188b0fd3-5d71-4feb-aca5-75a2bd28895a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 24.952s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 729.110066] env[61957]: DEBUG oslo_concurrency.lockutils [req-833348f1-0926-41b6-b121-e3c3fca44581 req-70a4d399-b2f3-40dc-b602-e50aaf7e1c11 service nova] Releasing lock "refresh_cache-f8235178-7ae0-4733-bb05-3515d5489a5b" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 729.110066] env[61957]: DEBUG oslo_concurrency.lockutils [None req-2b9492ff-1cb8-4ba9-ba5e-b513c4b108ad tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] Acquired lock "refresh_cache-f8235178-7ae0-4733-bb05-3515d5489a5b" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 729.110066] env[61957]: DEBUG nova.network.neutron [None req-2b9492ff-1cb8-4ba9-ba5e-b513c4b108ad tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 729.322627] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64ac05dd-5477-4043-8c47-e586bc954daf {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.330711] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b18fe7e-387b-40d6-895e-4ec7801ebd27 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.364501] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb88d84a-4f32-423a-9297-621177183c45 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.377295] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a6586a0-f694-44ae-b129-8c61e6282f46 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.393710] env[61957]: DEBUG nova.compute.provider_tree [None req-c2f4fc75-b8bd-417b-8703-3b9048472a92 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 729.634784] env[61957]: DEBUG nova.network.neutron [None req-2b9492ff-1cb8-4ba9-ba5e-b513c4b108ad tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 729.727362] env[61957]: DEBUG nova.network.neutron [None req-2b9492ff-1cb8-4ba9-ba5e-b513c4b108ad tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 729.896533] env[61957]: DEBUG nova.scheduler.client.report [None req-c2f4fc75-b8bd-417b-8703-3b9048472a92 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 730.043747] env[61957]: DEBUG nova.compute.manager [req-3805be0d-e157-422c-b31b-58ed4126bdbe req-90486ec8-73e7-4cb1-a2be-75807f9c2c29 service nova] [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] Received event network-vif-deleted-d5344278-a5e5-4fe0-ab8d-80293d0e444b {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 730.229719] env[61957]: DEBUG oslo_concurrency.lockutils [None req-2b9492ff-1cb8-4ba9-ba5e-b513c4b108ad tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] Releasing lock "refresh_cache-f8235178-7ae0-4733-bb05-3515d5489a5b" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 730.232387] env[61957]: DEBUG nova.compute.manager [None req-2b9492ff-1cb8-4ba9-ba5e-b513c4b108ad tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] Start destroying the instance on the hypervisor. {{(pid=61957) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 730.232387] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-2b9492ff-1cb8-4ba9-ba5e-b513c4b108ad tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] Destroying instance {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 730.232387] env[61957]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2ee11cba-a745-4ee6-af2a-ffb22c6e1071 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.246525] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0cf94d81-f3d2-48b6-b101-630e777a6933 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.272658] env[61957]: WARNING nova.virt.vmwareapi.vmops [None req-2b9492ff-1cb8-4ba9-ba5e-b513c4b108ad tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance f8235178-7ae0-4733-bb05-3515d5489a5b could not be found. [ 730.272922] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-2b9492ff-1cb8-4ba9-ba5e-b513c4b108ad tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] Instance destroyed {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 730.273073] env[61957]: INFO nova.compute.manager [None req-2b9492ff-1cb8-4ba9-ba5e-b513c4b108ad tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] Took 0.04 seconds to destroy the instance on the hypervisor. [ 730.273966] env[61957]: DEBUG oslo.service.loopingcall [None req-2b9492ff-1cb8-4ba9-ba5e-b513c4b108ad tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 730.273966] env[61957]: DEBUG nova.compute.manager [-] [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 730.274086] env[61957]: DEBUG nova.network.neutron [-] [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 730.296271] env[61957]: DEBUG nova.network.neutron [-] [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 730.403675] env[61957]: DEBUG oslo_concurrency.lockutils [None req-c2f4fc75-b8bd-417b-8703-3b9048472a92 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.948s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 730.404344] env[61957]: ERROR nova.compute.manager [None req-c2f4fc75-b8bd-417b-8703-3b9048472a92 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port b8d0e37e-0f54-4ef7-90b0-fb5903f71e0f, please check neutron logs for more information. [ 730.404344] env[61957]: ERROR nova.compute.manager [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] Traceback (most recent call last): [ 730.404344] env[61957]: ERROR nova.compute.manager [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 730.404344] env[61957]: ERROR nova.compute.manager [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] self.driver.spawn(context, instance, image_meta, [ 730.404344] env[61957]: ERROR nova.compute.manager [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 730.404344] env[61957]: ERROR nova.compute.manager [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] self._vmops.spawn(context, instance, image_meta, injected_files, [ 730.404344] env[61957]: ERROR nova.compute.manager [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 730.404344] env[61957]: ERROR nova.compute.manager [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] vm_ref = self.build_virtual_machine(instance, [ 730.404344] env[61957]: ERROR nova.compute.manager [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 730.404344] env[61957]: ERROR nova.compute.manager [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] vif_infos = vmwarevif.get_vif_info(self._session, [ 730.404344] env[61957]: ERROR nova.compute.manager [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 730.404737] env[61957]: ERROR nova.compute.manager [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] for vif in network_info: [ 730.404737] env[61957]: ERROR nova.compute.manager [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 730.404737] env[61957]: ERROR nova.compute.manager [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] return self._sync_wrapper(fn, *args, **kwargs) [ 730.404737] env[61957]: ERROR nova.compute.manager [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 730.404737] env[61957]: ERROR nova.compute.manager [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] self.wait() [ 730.404737] env[61957]: ERROR nova.compute.manager [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 730.404737] env[61957]: ERROR nova.compute.manager [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] self[:] = self._gt.wait() [ 730.404737] env[61957]: ERROR nova.compute.manager [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 730.404737] env[61957]: ERROR nova.compute.manager [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] return self._exit_event.wait() [ 730.404737] env[61957]: ERROR nova.compute.manager [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 730.404737] env[61957]: ERROR nova.compute.manager [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] current.throw(*self._exc) [ 730.404737] env[61957]: ERROR nova.compute.manager [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 730.404737] env[61957]: ERROR nova.compute.manager [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] result = function(*args, **kwargs) [ 730.405137] env[61957]: ERROR nova.compute.manager [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 730.405137] env[61957]: ERROR nova.compute.manager [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] return func(*args, **kwargs) [ 730.405137] env[61957]: ERROR nova.compute.manager [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 730.405137] env[61957]: ERROR nova.compute.manager [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] raise e [ 730.405137] env[61957]: ERROR nova.compute.manager [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 730.405137] env[61957]: ERROR nova.compute.manager [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] nwinfo = self.network_api.allocate_for_instance( [ 730.405137] env[61957]: ERROR nova.compute.manager [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 730.405137] env[61957]: ERROR nova.compute.manager [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] created_port_ids = self._update_ports_for_instance( [ 730.405137] env[61957]: ERROR nova.compute.manager [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 730.405137] env[61957]: ERROR nova.compute.manager [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] with excutils.save_and_reraise_exception(): [ 730.405137] env[61957]: ERROR nova.compute.manager [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 730.405137] env[61957]: ERROR nova.compute.manager [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] self.force_reraise() [ 730.405137] env[61957]: ERROR nova.compute.manager [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 730.405732] env[61957]: ERROR nova.compute.manager [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] raise self.value [ 730.405732] env[61957]: ERROR nova.compute.manager [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 730.405732] env[61957]: ERROR nova.compute.manager [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] updated_port = self._update_port( [ 730.405732] env[61957]: ERROR nova.compute.manager [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 730.405732] env[61957]: ERROR nova.compute.manager [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] _ensure_no_port_binding_failure(port) [ 730.405732] env[61957]: ERROR nova.compute.manager [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 730.405732] env[61957]: ERROR nova.compute.manager [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] raise exception.PortBindingFailed(port_id=port['id']) [ 730.405732] env[61957]: ERROR nova.compute.manager [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] nova.exception.PortBindingFailed: Binding failed for port b8d0e37e-0f54-4ef7-90b0-fb5903f71e0f, please check neutron logs for more information. [ 730.405732] env[61957]: ERROR nova.compute.manager [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] [ 730.405732] env[61957]: DEBUG nova.compute.utils [None req-c2f4fc75-b8bd-417b-8703-3b9048472a92 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] Binding failed for port b8d0e37e-0f54-4ef7-90b0-fb5903f71e0f, please check neutron logs for more information. {{(pid=61957) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 730.406437] env[61957]: DEBUG oslo_concurrency.lockutils [None req-f462e606-4c03-40e0-ac91-60efbd401559 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.966s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 730.408202] env[61957]: INFO nova.compute.claims [None req-f462e606-4c03-40e0-ac91-60efbd401559 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 730.410971] env[61957]: DEBUG nova.compute.manager [None req-c2f4fc75-b8bd-417b-8703-3b9048472a92 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] Build of instance d0bde8ec-da55-42b5-8d2c-9df2b90b88a7 was re-scheduled: Binding failed for port b8d0e37e-0f54-4ef7-90b0-fb5903f71e0f, please check neutron logs for more information. {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 730.411399] env[61957]: DEBUG nova.compute.manager [None req-c2f4fc75-b8bd-417b-8703-3b9048472a92 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] Unplugging VIFs for instance {{(pid=61957) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 730.411618] env[61957]: DEBUG oslo_concurrency.lockutils [None req-c2f4fc75-b8bd-417b-8703-3b9048472a92 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] Acquiring lock "refresh_cache-d0bde8ec-da55-42b5-8d2c-9df2b90b88a7" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 730.411763] env[61957]: DEBUG oslo_concurrency.lockutils [None req-c2f4fc75-b8bd-417b-8703-3b9048472a92 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] Acquired lock "refresh_cache-d0bde8ec-da55-42b5-8d2c-9df2b90b88a7" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 730.411955] env[61957]: DEBUG nova.network.neutron [None req-c2f4fc75-b8bd-417b-8703-3b9048472a92 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 730.802446] env[61957]: DEBUG nova.network.neutron [-] [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 730.933807] env[61957]: DEBUG nova.network.neutron [None req-c2f4fc75-b8bd-417b-8703-3b9048472a92 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 731.016053] env[61957]: DEBUG nova.network.neutron [None req-c2f4fc75-b8bd-417b-8703-3b9048472a92 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 731.305740] env[61957]: INFO nova.compute.manager [-] [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] Took 1.03 seconds to deallocate network for instance. [ 731.311879] env[61957]: DEBUG nova.compute.claims [None req-2b9492ff-1cb8-4ba9-ba5e-b513c4b108ad tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] Aborting claim: {{(pid=61957) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 731.312114] env[61957]: DEBUG oslo_concurrency.lockutils [None req-2b9492ff-1cb8-4ba9-ba5e-b513c4b108ad tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 731.519859] env[61957]: DEBUG oslo_concurrency.lockutils [None req-c2f4fc75-b8bd-417b-8703-3b9048472a92 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] Releasing lock "refresh_cache-d0bde8ec-da55-42b5-8d2c-9df2b90b88a7" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 731.519859] env[61957]: DEBUG nova.compute.manager [None req-c2f4fc75-b8bd-417b-8703-3b9048472a92 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61957) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 731.519859] env[61957]: DEBUG nova.compute.manager [None req-c2f4fc75-b8bd-417b-8703-3b9048472a92 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 731.519859] env[61957]: DEBUG nova.network.neutron [None req-c2f4fc75-b8bd-417b-8703-3b9048472a92 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 731.543207] env[61957]: DEBUG nova.network.neutron [None req-c2f4fc75-b8bd-417b-8703-3b9048472a92 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 731.778594] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a49e3e7-b71a-4e38-8064-962caf054e84 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.787131] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f212893a-d541-4a56-9b6a-b1402654948c {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.821673] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ff94d0d-8ba5-4846-bd0a-a50a7a4301bf {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.830024] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24183850-99f4-4888-b9b7-fb85a853a5b4 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.843388] env[61957]: DEBUG nova.compute.provider_tree [None req-f462e606-4c03-40e0-ac91-60efbd401559 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 732.046141] env[61957]: DEBUG nova.network.neutron [None req-c2f4fc75-b8bd-417b-8703-3b9048472a92 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 732.346871] env[61957]: DEBUG nova.scheduler.client.report [None req-f462e606-4c03-40e0-ac91-60efbd401559 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 732.553739] env[61957]: INFO nova.compute.manager [None req-c2f4fc75-b8bd-417b-8703-3b9048472a92 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] [instance: d0bde8ec-da55-42b5-8d2c-9df2b90b88a7] Took 1.03 seconds to deallocate network for instance. [ 732.853402] env[61957]: DEBUG oslo_concurrency.lockutils [None req-f462e606-4c03-40e0-ac91-60efbd401559 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.447s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 732.854166] env[61957]: DEBUG nova.compute.manager [None req-f462e606-4c03-40e0-ac91-60efbd401559 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] Start building networks asynchronously for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 732.858399] env[61957]: DEBUG oslo_concurrency.lockutils [None req-691c932a-6913-4bb6-98f2-4142c7d131a8 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 18.801s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 733.363974] env[61957]: DEBUG nova.compute.utils [None req-f462e606-4c03-40e0-ac91-60efbd401559 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Using /dev/sd instead of None {{(pid=61957) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 733.370996] env[61957]: DEBUG nova.compute.manager [None req-f462e606-4c03-40e0-ac91-60efbd401559 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] Allocating IP information in the background. {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 733.370996] env[61957]: DEBUG nova.network.neutron [None req-f462e606-4c03-40e0-ac91-60efbd401559 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] allocate_for_instance() {{(pid=61957) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 733.433889] env[61957]: DEBUG nova.policy [None req-f462e606-4c03-40e0-ac91-60efbd401559 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd0405e9c38cb4012a0212284be7e9aed', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1673c854272f4b14a623a73b8f4b687b', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61957) authorize /opt/stack/nova/nova/policy.py:203}} [ 733.593741] env[61957]: INFO nova.scheduler.client.report [None req-c2f4fc75-b8bd-417b-8703-3b9048472a92 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] Deleted allocations for instance d0bde8ec-da55-42b5-8d2c-9df2b90b88a7 [ 733.813049] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6c46bad-206c-4521-968b-69af6e0b3758 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.821366] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5915a11e-c06e-45bf-9b96-49a5846992b6 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.862819] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bba34374-c843-42e0-a93d-28e1c0b0d3c9 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.872824] env[61957]: DEBUG nova.network.neutron [None req-f462e606-4c03-40e0-ac91-60efbd401559 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] Successfully created port: e0a84501-e615-4647-8dff-5c863b104dc5 {{(pid=61957) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 733.875612] env[61957]: DEBUG nova.compute.manager [None req-f462e606-4c03-40e0-ac91-60efbd401559 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] Start building block device mappings for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 733.884370] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c780e51b-b1e1-43b4-9618-ebf5e88b59d6 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.898401] env[61957]: DEBUG nova.compute.provider_tree [None req-691c932a-6913-4bb6-98f2-4142c7d131a8 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 734.122326] env[61957]: DEBUG oslo_concurrency.lockutils [None req-c2f4fc75-b8bd-417b-8703-3b9048472a92 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] Lock "d0bde8ec-da55-42b5-8d2c-9df2b90b88a7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 122.613s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 734.402121] env[61957]: DEBUG nova.scheduler.client.report [None req-691c932a-6913-4bb6-98f2-4142c7d131a8 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 734.625903] env[61957]: DEBUG nova.compute.manager [None req-47718518-520f-4177-b41f-8f0cae78b6f6 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] [instance: 7f3850c0-a455-4e5b-a329-fb3b02f9725f] Starting instance... {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 734.805615] env[61957]: ERROR nova.compute.manager [None req-f462e606-4c03-40e0-ac91-60efbd401559 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port e0a84501-e615-4647-8dff-5c863b104dc5, please check neutron logs for more information. [ 734.805615] env[61957]: ERROR nova.compute.manager Traceback (most recent call last): [ 734.805615] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 734.805615] env[61957]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 734.805615] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 734.805615] env[61957]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 734.805615] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 734.805615] env[61957]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 734.805615] env[61957]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 734.805615] env[61957]: ERROR nova.compute.manager self.force_reraise() [ 734.805615] env[61957]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 734.805615] env[61957]: ERROR nova.compute.manager raise self.value [ 734.805615] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 734.805615] env[61957]: ERROR nova.compute.manager updated_port = self._update_port( [ 734.805615] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 734.805615] env[61957]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 734.806097] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 734.806097] env[61957]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 734.806097] env[61957]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port e0a84501-e615-4647-8dff-5c863b104dc5, please check neutron logs for more information. [ 734.806097] env[61957]: ERROR nova.compute.manager [ 734.806097] env[61957]: Traceback (most recent call last): [ 734.806097] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 734.806097] env[61957]: listener.cb(fileno) [ 734.806097] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 734.806097] env[61957]: result = function(*args, **kwargs) [ 734.806097] env[61957]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 734.806097] env[61957]: return func(*args, **kwargs) [ 734.806097] env[61957]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 734.806097] env[61957]: raise e [ 734.806097] env[61957]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 734.806097] env[61957]: nwinfo = self.network_api.allocate_for_instance( [ 734.806097] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 734.806097] env[61957]: created_port_ids = self._update_ports_for_instance( [ 734.806097] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 734.806097] env[61957]: with excutils.save_and_reraise_exception(): [ 734.806097] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 734.806097] env[61957]: self.force_reraise() [ 734.806097] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 734.806097] env[61957]: raise self.value [ 734.806097] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 734.806097] env[61957]: updated_port = self._update_port( [ 734.806097] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 734.806097] env[61957]: _ensure_no_port_binding_failure(port) [ 734.806097] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 734.806097] env[61957]: raise exception.PortBindingFailed(port_id=port['id']) [ 734.806826] env[61957]: nova.exception.PortBindingFailed: Binding failed for port e0a84501-e615-4647-8dff-5c863b104dc5, please check neutron logs for more information. [ 734.806826] env[61957]: Removing descriptor: 17 [ 734.885337] env[61957]: DEBUG nova.compute.manager [None req-f462e606-4c03-40e0-ac91-60efbd401559 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] Start spawning the instance on the hypervisor. {{(pid=61957) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 734.917033] env[61957]: DEBUG nova.virt.hardware [None req-f462e606-4c03-40e0-ac91-60efbd401559 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T17:22:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T17:21:50Z,direct_url=,disk_format='vmdk',id=11c76a2c-f705-470a-ba9d-4657858bab38,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='81d3096d0f094373913fc3dc8f5c3c6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T17:21:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 734.917033] env[61957]: DEBUG nova.virt.hardware [None req-f462e606-4c03-40e0-ac91-60efbd401559 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Flavor limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 734.917033] env[61957]: DEBUG nova.virt.hardware [None req-f462e606-4c03-40e0-ac91-60efbd401559 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Image limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 734.917312] env[61957]: DEBUG nova.virt.hardware [None req-f462e606-4c03-40e0-ac91-60efbd401559 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Flavor pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 734.917312] env[61957]: DEBUG nova.virt.hardware [None req-f462e606-4c03-40e0-ac91-60efbd401559 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Image pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 734.917312] env[61957]: DEBUG nova.virt.hardware [None req-f462e606-4c03-40e0-ac91-60efbd401559 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 734.917312] env[61957]: DEBUG nova.virt.hardware [None req-f462e606-4c03-40e0-ac91-60efbd401559 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 734.917312] env[61957]: DEBUG nova.virt.hardware [None req-f462e606-4c03-40e0-ac91-60efbd401559 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 734.917473] env[61957]: DEBUG nova.virt.hardware [None req-f462e606-4c03-40e0-ac91-60efbd401559 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Got 1 possible topologies {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 734.917473] env[61957]: DEBUG nova.virt.hardware [None req-f462e606-4c03-40e0-ac91-60efbd401559 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 734.917473] env[61957]: DEBUG nova.virt.hardware [None req-f462e606-4c03-40e0-ac91-60efbd401559 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 734.917473] env[61957]: DEBUG oslo_concurrency.lockutils [None req-691c932a-6913-4bb6-98f2-4142c7d131a8 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.059s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 734.917886] env[61957]: ERROR nova.compute.manager [None req-691c932a-6913-4bb6-98f2-4142c7d131a8 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 75f87eff-42b6-4098-9846-409a7c48d82f, please check neutron logs for more information. [ 734.917886] env[61957]: ERROR nova.compute.manager [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] Traceback (most recent call last): [ 734.917886] env[61957]: ERROR nova.compute.manager [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 734.917886] env[61957]: ERROR nova.compute.manager [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] self.driver.spawn(context, instance, image_meta, [ 734.917886] env[61957]: ERROR nova.compute.manager [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 734.917886] env[61957]: ERROR nova.compute.manager [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] self._vmops.spawn(context, instance, image_meta, injected_files, [ 734.917886] env[61957]: ERROR nova.compute.manager [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 734.917886] env[61957]: ERROR nova.compute.manager [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] vm_ref = self.build_virtual_machine(instance, [ 734.917886] env[61957]: ERROR nova.compute.manager [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 734.917886] env[61957]: ERROR nova.compute.manager [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] vif_infos = vmwarevif.get_vif_info(self._session, [ 734.917886] env[61957]: ERROR nova.compute.manager [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 734.918267] env[61957]: ERROR nova.compute.manager [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] for vif in network_info: [ 734.918267] env[61957]: ERROR nova.compute.manager [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 734.918267] env[61957]: ERROR nova.compute.manager [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] return self._sync_wrapper(fn, *args, **kwargs) [ 734.918267] env[61957]: ERROR nova.compute.manager [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 734.918267] env[61957]: ERROR nova.compute.manager [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] self.wait() [ 734.918267] env[61957]: ERROR nova.compute.manager [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 734.918267] env[61957]: ERROR nova.compute.manager [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] self[:] = self._gt.wait() [ 734.918267] env[61957]: ERROR nova.compute.manager [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 734.918267] env[61957]: ERROR nova.compute.manager [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] return self._exit_event.wait() [ 734.918267] env[61957]: ERROR nova.compute.manager [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 734.918267] env[61957]: ERROR nova.compute.manager [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] result = hub.switch() [ 734.918267] env[61957]: ERROR nova.compute.manager [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 734.918267] env[61957]: ERROR nova.compute.manager [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] return self.greenlet.switch() [ 734.918757] env[61957]: ERROR nova.compute.manager [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 734.918757] env[61957]: ERROR nova.compute.manager [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] result = function(*args, **kwargs) [ 734.918757] env[61957]: ERROR nova.compute.manager [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 734.918757] env[61957]: ERROR nova.compute.manager [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] return func(*args, **kwargs) [ 734.918757] env[61957]: ERROR nova.compute.manager [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 734.918757] env[61957]: ERROR nova.compute.manager [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] raise e [ 734.918757] env[61957]: ERROR nova.compute.manager [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 734.918757] env[61957]: ERROR nova.compute.manager [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] nwinfo = self.network_api.allocate_for_instance( [ 734.918757] env[61957]: ERROR nova.compute.manager [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 734.918757] env[61957]: ERROR nova.compute.manager [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] created_port_ids = self._update_ports_for_instance( [ 734.918757] env[61957]: ERROR nova.compute.manager [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 734.918757] env[61957]: ERROR nova.compute.manager [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] with excutils.save_and_reraise_exception(): [ 734.918757] env[61957]: ERROR nova.compute.manager [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 734.919133] env[61957]: ERROR nova.compute.manager [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] self.force_reraise() [ 734.919133] env[61957]: ERROR nova.compute.manager [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 734.919133] env[61957]: ERROR nova.compute.manager [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] raise self.value [ 734.919133] env[61957]: ERROR nova.compute.manager [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 734.919133] env[61957]: ERROR nova.compute.manager [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] updated_port = self._update_port( [ 734.919133] env[61957]: ERROR nova.compute.manager [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 734.919133] env[61957]: ERROR nova.compute.manager [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] _ensure_no_port_binding_failure(port) [ 734.919133] env[61957]: ERROR nova.compute.manager [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 734.919133] env[61957]: ERROR nova.compute.manager [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] raise exception.PortBindingFailed(port_id=port['id']) [ 734.919133] env[61957]: ERROR nova.compute.manager [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] nova.exception.PortBindingFailed: Binding failed for port 75f87eff-42b6-4098-9846-409a7c48d82f, please check neutron logs for more information. [ 734.919133] env[61957]: ERROR nova.compute.manager [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] [ 734.922201] env[61957]: DEBUG nova.compute.utils [None req-691c932a-6913-4bb6-98f2-4142c7d131a8 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] Binding failed for port 75f87eff-42b6-4098-9846-409a7c48d82f, please check neutron logs for more information. {{(pid=61957) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 734.925709] env[61957]: DEBUG nova.compute.manager [None req-691c932a-6913-4bb6-98f2-4142c7d131a8 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] Build of instance c5024e22-6499-4f2e-8d1b-99ca34b33104 was re-scheduled: Binding failed for port 75f87eff-42b6-4098-9846-409a7c48d82f, please check neutron logs for more information. {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 734.926712] env[61957]: DEBUG nova.compute.manager [None req-691c932a-6913-4bb6-98f2-4142c7d131a8 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] Unplugging VIFs for instance {{(pid=61957) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 734.927604] env[61957]: DEBUG oslo_concurrency.lockutils [None req-691c932a-6913-4bb6-98f2-4142c7d131a8 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Acquiring lock "refresh_cache-c5024e22-6499-4f2e-8d1b-99ca34b33104" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 734.927794] env[61957]: DEBUG oslo_concurrency.lockutils [None req-691c932a-6913-4bb6-98f2-4142c7d131a8 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Acquired lock "refresh_cache-c5024e22-6499-4f2e-8d1b-99ca34b33104" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 734.927977] env[61957]: DEBUG nova.network.neutron [None req-691c932a-6913-4bb6-98f2-4142c7d131a8 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 734.931473] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4fff390-e6df-4b6f-9eec-c7fbd29a8e2f {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.935123] env[61957]: DEBUG oslo_concurrency.lockutils [None req-456a04ab-263c-4c55-898a-70a189651712 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 20.056s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 734.941120] env[61957]: DEBUG nova.compute.manager [req-e2eb5463-d531-44e3-864d-f6b96b53a695 req-3249ca8a-7117-484b-93f4-3692c0f32110 service nova] [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] Received event network-changed-e0a84501-e615-4647-8dff-5c863b104dc5 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 734.941359] env[61957]: DEBUG nova.compute.manager [req-e2eb5463-d531-44e3-864d-f6b96b53a695 req-3249ca8a-7117-484b-93f4-3692c0f32110 service nova] [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] Refreshing instance network info cache due to event network-changed-e0a84501-e615-4647-8dff-5c863b104dc5. {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 734.941570] env[61957]: DEBUG oslo_concurrency.lockutils [req-e2eb5463-d531-44e3-864d-f6b96b53a695 req-3249ca8a-7117-484b-93f4-3692c0f32110 service nova] Acquiring lock "refresh_cache-115c1bb4-0e6b-4c6c-9d38-9584debea01f" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 734.941772] env[61957]: DEBUG oslo_concurrency.lockutils [req-e2eb5463-d531-44e3-864d-f6b96b53a695 req-3249ca8a-7117-484b-93f4-3692c0f32110 service nova] Acquired lock "refresh_cache-115c1bb4-0e6b-4c6c-9d38-9584debea01f" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 734.942128] env[61957]: DEBUG nova.network.neutron [req-e2eb5463-d531-44e3-864d-f6b96b53a695 req-3249ca8a-7117-484b-93f4-3692c0f32110 service nova] [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] Refreshing network info cache for port e0a84501-e615-4647-8dff-5c863b104dc5 {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 734.951131] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d60147b-758e-49b7-bed8-f23fec0aa096 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.966828] env[61957]: ERROR nova.compute.manager [None req-f462e606-4c03-40e0-ac91-60efbd401559 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port e0a84501-e615-4647-8dff-5c863b104dc5, please check neutron logs for more information. [ 734.966828] env[61957]: ERROR nova.compute.manager [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] Traceback (most recent call last): [ 734.966828] env[61957]: ERROR nova.compute.manager [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 734.966828] env[61957]: ERROR nova.compute.manager [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] yield resources [ 734.966828] env[61957]: ERROR nova.compute.manager [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 734.966828] env[61957]: ERROR nova.compute.manager [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] self.driver.spawn(context, instance, image_meta, [ 734.966828] env[61957]: ERROR nova.compute.manager [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 734.966828] env[61957]: ERROR nova.compute.manager [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 734.966828] env[61957]: ERROR nova.compute.manager [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 734.966828] env[61957]: ERROR nova.compute.manager [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] vm_ref = self.build_virtual_machine(instance, [ 734.966828] env[61957]: ERROR nova.compute.manager [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 734.967251] env[61957]: ERROR nova.compute.manager [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] vif_infos = vmwarevif.get_vif_info(self._session, [ 734.967251] env[61957]: ERROR nova.compute.manager [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 734.967251] env[61957]: ERROR nova.compute.manager [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] for vif in network_info: [ 734.967251] env[61957]: ERROR nova.compute.manager [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 734.967251] env[61957]: ERROR nova.compute.manager [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] return self._sync_wrapper(fn, *args, **kwargs) [ 734.967251] env[61957]: ERROR nova.compute.manager [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 734.967251] env[61957]: ERROR nova.compute.manager [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] self.wait() [ 734.967251] env[61957]: ERROR nova.compute.manager [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 734.967251] env[61957]: ERROR nova.compute.manager [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] self[:] = self._gt.wait() [ 734.967251] env[61957]: ERROR nova.compute.manager [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 734.967251] env[61957]: ERROR nova.compute.manager [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] return self._exit_event.wait() [ 734.967251] env[61957]: ERROR nova.compute.manager [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 734.967251] env[61957]: ERROR nova.compute.manager [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] current.throw(*self._exc) [ 734.968722] env[61957]: ERROR nova.compute.manager [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 734.968722] env[61957]: ERROR nova.compute.manager [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] result = function(*args, **kwargs) [ 734.968722] env[61957]: ERROR nova.compute.manager [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 734.968722] env[61957]: ERROR nova.compute.manager [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] return func(*args, **kwargs) [ 734.968722] env[61957]: ERROR nova.compute.manager [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 734.968722] env[61957]: ERROR nova.compute.manager [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] raise e [ 734.968722] env[61957]: ERROR nova.compute.manager [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 734.968722] env[61957]: ERROR nova.compute.manager [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] nwinfo = self.network_api.allocate_for_instance( [ 734.968722] env[61957]: ERROR nova.compute.manager [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 734.968722] env[61957]: ERROR nova.compute.manager [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] created_port_ids = self._update_ports_for_instance( [ 734.968722] env[61957]: ERROR nova.compute.manager [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 734.968722] env[61957]: ERROR nova.compute.manager [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] with excutils.save_and_reraise_exception(): [ 734.968722] env[61957]: ERROR nova.compute.manager [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 734.969121] env[61957]: ERROR nova.compute.manager [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] self.force_reraise() [ 734.969121] env[61957]: ERROR nova.compute.manager [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 734.969121] env[61957]: ERROR nova.compute.manager [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] raise self.value [ 734.969121] env[61957]: ERROR nova.compute.manager [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 734.969121] env[61957]: ERROR nova.compute.manager [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] updated_port = self._update_port( [ 734.969121] env[61957]: ERROR nova.compute.manager [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 734.969121] env[61957]: ERROR nova.compute.manager [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] _ensure_no_port_binding_failure(port) [ 734.969121] env[61957]: ERROR nova.compute.manager [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 734.969121] env[61957]: ERROR nova.compute.manager [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] raise exception.PortBindingFailed(port_id=port['id']) [ 734.969121] env[61957]: ERROR nova.compute.manager [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] nova.exception.PortBindingFailed: Binding failed for port e0a84501-e615-4647-8dff-5c863b104dc5, please check neutron logs for more information. [ 734.969121] env[61957]: ERROR nova.compute.manager [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] [ 734.969121] env[61957]: INFO nova.compute.manager [None req-f462e606-4c03-40e0-ac91-60efbd401559 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] Terminating instance [ 734.969449] env[61957]: DEBUG oslo_concurrency.lockutils [None req-f462e606-4c03-40e0-ac91-60efbd401559 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Acquiring lock "refresh_cache-115c1bb4-0e6b-4c6c-9d38-9584debea01f" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 735.025870] env[61957]: DEBUG oslo_concurrency.lockutils [None req-7bbc9489-a498-401d-a2af-2f28d461499e tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Acquiring lock "4951132e-7247-4772-8f88-3664c6a7e61e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 735.026115] env[61957]: DEBUG oslo_concurrency.lockutils [None req-7bbc9489-a498-401d-a2af-2f28d461499e tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Lock "4951132e-7247-4772-8f88-3664c6a7e61e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 735.150963] env[61957]: DEBUG oslo_concurrency.lockutils [None req-47718518-520f-4177-b41f-8f0cae78b6f6 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 735.459338] env[61957]: DEBUG nova.network.neutron [None req-691c932a-6913-4bb6-98f2-4142c7d131a8 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 735.465137] env[61957]: DEBUG nova.network.neutron [req-e2eb5463-d531-44e3-864d-f6b96b53a695 req-3249ca8a-7117-484b-93f4-3692c0f32110 service nova] [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 735.585703] env[61957]: DEBUG nova.network.neutron [None req-691c932a-6913-4bb6-98f2-4142c7d131a8 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 735.600113] env[61957]: DEBUG nova.network.neutron [req-e2eb5463-d531-44e3-864d-f6b96b53a695 req-3249ca8a-7117-484b-93f4-3692c0f32110 service nova] [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 735.666080] env[61957]: DEBUG oslo_concurrency.lockutils [None req-3c3df299-fd1f-460a-8cc2-979af1db30d1 tempest-ServersTestJSON-1093633677 tempest-ServersTestJSON-1093633677-project-member] Acquiring lock "ead6aae1-36b5-4f57-9129-3bb02cf103ce" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 735.666311] env[61957]: DEBUG oslo_concurrency.lockutils [None req-3c3df299-fd1f-460a-8cc2-979af1db30d1 tempest-ServersTestJSON-1093633677 tempest-ServersTestJSON-1093633677-project-member] Lock "ead6aae1-36b5-4f57-9129-3bb02cf103ce" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 735.812218] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02adf7b8-778b-4da5-98f6-5f28c5db2b56 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.819999] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13f1bb9f-cc82-4a68-82f0-7d24645782d9 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.850995] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac15a3a1-037c-4038-b15b-c465906f7464 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.859009] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a4c65db-5d6e-441d-b3c5-5bdb2b1bdee1 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.872421] env[61957]: DEBUG nova.compute.provider_tree [None req-456a04ab-263c-4c55-898a-70a189651712 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 736.091873] env[61957]: DEBUG oslo_concurrency.lockutils [None req-691c932a-6913-4bb6-98f2-4142c7d131a8 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Releasing lock "refresh_cache-c5024e22-6499-4f2e-8d1b-99ca34b33104" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 736.092238] env[61957]: DEBUG nova.compute.manager [None req-691c932a-6913-4bb6-98f2-4142c7d131a8 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61957) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 736.092392] env[61957]: DEBUG nova.compute.manager [None req-691c932a-6913-4bb6-98f2-4142c7d131a8 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 736.092564] env[61957]: DEBUG nova.network.neutron [None req-691c932a-6913-4bb6-98f2-4142c7d131a8 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 736.104090] env[61957]: DEBUG oslo_concurrency.lockutils [req-e2eb5463-d531-44e3-864d-f6b96b53a695 req-3249ca8a-7117-484b-93f4-3692c0f32110 service nova] Releasing lock "refresh_cache-115c1bb4-0e6b-4c6c-9d38-9584debea01f" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 736.104536] env[61957]: DEBUG oslo_concurrency.lockutils [None req-f462e606-4c03-40e0-ac91-60efbd401559 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Acquired lock "refresh_cache-115c1bb4-0e6b-4c6c-9d38-9584debea01f" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 736.104719] env[61957]: DEBUG nova.network.neutron [None req-f462e606-4c03-40e0-ac91-60efbd401559 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 736.109640] env[61957]: DEBUG nova.network.neutron [None req-691c932a-6913-4bb6-98f2-4142c7d131a8 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 736.376162] env[61957]: DEBUG nova.scheduler.client.report [None req-456a04ab-263c-4c55-898a-70a189651712 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 736.611697] env[61957]: DEBUG nova.network.neutron [None req-691c932a-6913-4bb6-98f2-4142c7d131a8 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 736.626072] env[61957]: DEBUG nova.network.neutron [None req-f462e606-4c03-40e0-ac91-60efbd401559 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 736.712027] env[61957]: DEBUG nova.network.neutron [None req-f462e606-4c03-40e0-ac91-60efbd401559 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 736.881940] env[61957]: DEBUG oslo_concurrency.lockutils [None req-456a04ab-263c-4c55-898a-70a189651712 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.946s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 736.881940] env[61957]: ERROR nova.compute.manager [None req-456a04ab-263c-4c55-898a-70a189651712 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port a0d948b4-5d2e-4133-8e39-e685c226b80f, please check neutron logs for more information. [ 736.881940] env[61957]: ERROR nova.compute.manager [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] Traceback (most recent call last): [ 736.881940] env[61957]: ERROR nova.compute.manager [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 736.881940] env[61957]: ERROR nova.compute.manager [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] self.driver.spawn(context, instance, image_meta, [ 736.881940] env[61957]: ERROR nova.compute.manager [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 736.881940] env[61957]: ERROR nova.compute.manager [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] self._vmops.spawn(context, instance, image_meta, injected_files, [ 736.881940] env[61957]: ERROR nova.compute.manager [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 736.881940] env[61957]: ERROR nova.compute.manager [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] vm_ref = self.build_virtual_machine(instance, [ 736.882326] env[61957]: ERROR nova.compute.manager [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 736.882326] env[61957]: ERROR nova.compute.manager [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] vif_infos = vmwarevif.get_vif_info(self._session, [ 736.882326] env[61957]: ERROR nova.compute.manager [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 736.882326] env[61957]: ERROR nova.compute.manager [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] for vif in network_info: [ 736.882326] env[61957]: ERROR nova.compute.manager [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 736.882326] env[61957]: ERROR nova.compute.manager [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] return self._sync_wrapper(fn, *args, **kwargs) [ 736.882326] env[61957]: ERROR nova.compute.manager [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 736.882326] env[61957]: ERROR nova.compute.manager [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] self.wait() [ 736.882326] env[61957]: ERROR nova.compute.manager [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 736.882326] env[61957]: ERROR nova.compute.manager [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] self[:] = self._gt.wait() [ 736.882326] env[61957]: ERROR nova.compute.manager [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 736.882326] env[61957]: ERROR nova.compute.manager [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] return self._exit_event.wait() [ 736.882326] env[61957]: ERROR nova.compute.manager [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 736.882668] env[61957]: ERROR nova.compute.manager [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] current.throw(*self._exc) [ 736.882668] env[61957]: ERROR nova.compute.manager [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 736.882668] env[61957]: ERROR nova.compute.manager [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] result = function(*args, **kwargs) [ 736.882668] env[61957]: ERROR nova.compute.manager [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 736.882668] env[61957]: ERROR nova.compute.manager [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] return func(*args, **kwargs) [ 736.882668] env[61957]: ERROR nova.compute.manager [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 736.882668] env[61957]: ERROR nova.compute.manager [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] raise e [ 736.882668] env[61957]: ERROR nova.compute.manager [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 736.882668] env[61957]: ERROR nova.compute.manager [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] nwinfo = self.network_api.allocate_for_instance( [ 736.882668] env[61957]: ERROR nova.compute.manager [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 736.882668] env[61957]: ERROR nova.compute.manager [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] created_port_ids = self._update_ports_for_instance( [ 736.882668] env[61957]: ERROR nova.compute.manager [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 736.882668] env[61957]: ERROR nova.compute.manager [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] with excutils.save_and_reraise_exception(): [ 736.883061] env[61957]: ERROR nova.compute.manager [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 736.883061] env[61957]: ERROR nova.compute.manager [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] self.force_reraise() [ 736.883061] env[61957]: ERROR nova.compute.manager [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 736.883061] env[61957]: ERROR nova.compute.manager [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] raise self.value [ 736.883061] env[61957]: ERROR nova.compute.manager [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 736.883061] env[61957]: ERROR nova.compute.manager [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] updated_port = self._update_port( [ 736.883061] env[61957]: ERROR nova.compute.manager [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 736.883061] env[61957]: ERROR nova.compute.manager [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] _ensure_no_port_binding_failure(port) [ 736.883061] env[61957]: ERROR nova.compute.manager [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 736.883061] env[61957]: ERROR nova.compute.manager [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] raise exception.PortBindingFailed(port_id=port['id']) [ 736.883061] env[61957]: ERROR nova.compute.manager [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] nova.exception.PortBindingFailed: Binding failed for port a0d948b4-5d2e-4133-8e39-e685c226b80f, please check neutron logs for more information. [ 736.883061] env[61957]: ERROR nova.compute.manager [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] [ 736.883425] env[61957]: DEBUG nova.compute.utils [None req-456a04ab-263c-4c55-898a-70a189651712 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] Binding failed for port a0d948b4-5d2e-4133-8e39-e685c226b80f, please check neutron logs for more information. {{(pid=61957) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 736.883795] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ca1e8d1e-8c18-4652-924a-ab1cdb751a05 tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 19.493s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 736.887920] env[61957]: DEBUG nova.compute.manager [None req-456a04ab-263c-4c55-898a-70a189651712 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] Build of instance 6241fb44-ce1c-4881-bc32-7d19ddc83a61 was re-scheduled: Binding failed for port a0d948b4-5d2e-4133-8e39-e685c226b80f, please check neutron logs for more information. {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 736.888461] env[61957]: DEBUG nova.compute.manager [None req-456a04ab-263c-4c55-898a-70a189651712 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] Unplugging VIFs for instance {{(pid=61957) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 736.888861] env[61957]: DEBUG oslo_concurrency.lockutils [None req-456a04ab-263c-4c55-898a-70a189651712 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] Acquiring lock "refresh_cache-6241fb44-ce1c-4881-bc32-7d19ddc83a61" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 736.889027] env[61957]: DEBUG oslo_concurrency.lockutils [None req-456a04ab-263c-4c55-898a-70a189651712 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] Acquired lock "refresh_cache-6241fb44-ce1c-4881-bc32-7d19ddc83a61" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 736.889870] env[61957]: DEBUG nova.network.neutron [None req-456a04ab-263c-4c55-898a-70a189651712 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 736.972933] env[61957]: DEBUG nova.compute.manager [req-3b09aaa2-61f7-4ecd-b670-7aa686871f8c req-a2dd006d-b359-4600-9597-d22b415d0408 service nova] [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] Received event network-vif-deleted-e0a84501-e615-4647-8dff-5c863b104dc5 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 737.114089] env[61957]: INFO nova.compute.manager [None req-691c932a-6913-4bb6-98f2-4142c7d131a8 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: c5024e22-6499-4f2e-8d1b-99ca34b33104] Took 1.02 seconds to deallocate network for instance. [ 737.214141] env[61957]: DEBUG oslo_concurrency.lockutils [None req-f462e606-4c03-40e0-ac91-60efbd401559 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Releasing lock "refresh_cache-115c1bb4-0e6b-4c6c-9d38-9584debea01f" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 737.214248] env[61957]: DEBUG nova.compute.manager [None req-f462e606-4c03-40e0-ac91-60efbd401559 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] Start destroying the instance on the hypervisor. {{(pid=61957) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 737.214443] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-f462e606-4c03-40e0-ac91-60efbd401559 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] Destroying instance {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 737.214745] env[61957]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-42d19fcd-fd1a-4007-8a06-78494110261d {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.224304] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d224bff-e651-4b87-8ea3-b66607e3db6c {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.246544] env[61957]: WARNING nova.virt.vmwareapi.vmops [None req-f462e606-4c03-40e0-ac91-60efbd401559 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 115c1bb4-0e6b-4c6c-9d38-9584debea01f could not be found. [ 737.246544] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-f462e606-4c03-40e0-ac91-60efbd401559 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] Instance destroyed {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 737.246655] env[61957]: INFO nova.compute.manager [None req-f462e606-4c03-40e0-ac91-60efbd401559 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] Took 0.03 seconds to destroy the instance on the hypervisor. [ 737.246851] env[61957]: DEBUG oslo.service.loopingcall [None req-f462e606-4c03-40e0-ac91-60efbd401559 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 737.247097] env[61957]: DEBUG nova.compute.manager [-] [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 737.247809] env[61957]: DEBUG nova.network.neutron [-] [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 737.263287] env[61957]: DEBUG nova.network.neutron [-] [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 737.408849] env[61957]: DEBUG nova.network.neutron [None req-456a04ab-263c-4c55-898a-70a189651712 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 737.483932] env[61957]: DEBUG nova.network.neutron [None req-456a04ab-263c-4c55-898a-70a189651712 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 737.683199] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79c6b471-3240-4b46-b169-df499d0ba08c {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.691328] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35db92ab-639a-466c-a912-5be44339ba64 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.727787] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c451767a-4f05-4a34-8723-cdb903b22694 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.735349] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd407415-beca-452a-b30e-c9a65c17618d {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.749727] env[61957]: DEBUG nova.compute.provider_tree [None req-ca1e8d1e-8c18-4652-924a-ab1cdb751a05 tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 737.765471] env[61957]: DEBUG nova.network.neutron [-] [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 737.986205] env[61957]: DEBUG oslo_concurrency.lockutils [None req-456a04ab-263c-4c55-898a-70a189651712 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] Releasing lock "refresh_cache-6241fb44-ce1c-4881-bc32-7d19ddc83a61" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 737.986477] env[61957]: DEBUG nova.compute.manager [None req-456a04ab-263c-4c55-898a-70a189651712 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61957) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 737.986776] env[61957]: DEBUG nova.compute.manager [None req-456a04ab-263c-4c55-898a-70a189651712 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 737.986976] env[61957]: DEBUG nova.network.neutron [None req-456a04ab-263c-4c55-898a-70a189651712 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 738.002009] env[61957]: DEBUG nova.network.neutron [None req-456a04ab-263c-4c55-898a-70a189651712 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 738.142061] env[61957]: INFO nova.scheduler.client.report [None req-691c932a-6913-4bb6-98f2-4142c7d131a8 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Deleted allocations for instance c5024e22-6499-4f2e-8d1b-99ca34b33104 [ 738.252853] env[61957]: DEBUG nova.scheduler.client.report [None req-ca1e8d1e-8c18-4652-924a-ab1cdb751a05 tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 738.267780] env[61957]: INFO nova.compute.manager [-] [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] Took 1.02 seconds to deallocate network for instance. [ 738.270116] env[61957]: DEBUG nova.compute.claims [None req-f462e606-4c03-40e0-ac91-60efbd401559 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] Aborting claim: {{(pid=61957) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 738.270285] env[61957]: DEBUG oslo_concurrency.lockutils [None req-f462e606-4c03-40e0-ac91-60efbd401559 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 738.505540] env[61957]: DEBUG nova.network.neutron [None req-456a04ab-263c-4c55-898a-70a189651712 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 738.648474] env[61957]: DEBUG oslo_concurrency.lockutils [None req-691c932a-6913-4bb6-98f2-4142c7d131a8 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Lock "c5024e22-6499-4f2e-8d1b-99ca34b33104" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 125.273s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 738.758140] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ca1e8d1e-8c18-4652-924a-ab1cdb751a05 tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.873s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 738.758140] env[61957]: ERROR nova.compute.manager [None req-ca1e8d1e-8c18-4652-924a-ab1cdb751a05 tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] [instance: 825130d4-b1f6-434e-be3e-dca952abe930] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port b95b009e-bbe6-4159-bed7-0705d27090ee, please check neutron logs for more information. [ 738.758140] env[61957]: ERROR nova.compute.manager [instance: 825130d4-b1f6-434e-be3e-dca952abe930] Traceback (most recent call last): [ 738.758140] env[61957]: ERROR nova.compute.manager [instance: 825130d4-b1f6-434e-be3e-dca952abe930] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 738.758140] env[61957]: ERROR nova.compute.manager [instance: 825130d4-b1f6-434e-be3e-dca952abe930] self.driver.spawn(context, instance, image_meta, [ 738.758140] env[61957]: ERROR nova.compute.manager [instance: 825130d4-b1f6-434e-be3e-dca952abe930] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 738.758140] env[61957]: ERROR nova.compute.manager [instance: 825130d4-b1f6-434e-be3e-dca952abe930] self._vmops.spawn(context, instance, image_meta, injected_files, [ 738.758140] env[61957]: ERROR nova.compute.manager [instance: 825130d4-b1f6-434e-be3e-dca952abe930] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 738.758140] env[61957]: ERROR nova.compute.manager [instance: 825130d4-b1f6-434e-be3e-dca952abe930] vm_ref = self.build_virtual_machine(instance, [ 738.758435] env[61957]: ERROR nova.compute.manager [instance: 825130d4-b1f6-434e-be3e-dca952abe930] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 738.758435] env[61957]: ERROR nova.compute.manager [instance: 825130d4-b1f6-434e-be3e-dca952abe930] vif_infos = vmwarevif.get_vif_info(self._session, [ 738.758435] env[61957]: ERROR nova.compute.manager [instance: 825130d4-b1f6-434e-be3e-dca952abe930] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 738.758435] env[61957]: ERROR nova.compute.manager [instance: 825130d4-b1f6-434e-be3e-dca952abe930] for vif in network_info: [ 738.758435] env[61957]: ERROR nova.compute.manager [instance: 825130d4-b1f6-434e-be3e-dca952abe930] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 738.758435] env[61957]: ERROR nova.compute.manager [instance: 825130d4-b1f6-434e-be3e-dca952abe930] return self._sync_wrapper(fn, *args, **kwargs) [ 738.758435] env[61957]: ERROR nova.compute.manager [instance: 825130d4-b1f6-434e-be3e-dca952abe930] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 738.758435] env[61957]: ERROR nova.compute.manager [instance: 825130d4-b1f6-434e-be3e-dca952abe930] self.wait() [ 738.758435] env[61957]: ERROR nova.compute.manager [instance: 825130d4-b1f6-434e-be3e-dca952abe930] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 738.758435] env[61957]: ERROR nova.compute.manager [instance: 825130d4-b1f6-434e-be3e-dca952abe930] self[:] = self._gt.wait() [ 738.758435] env[61957]: ERROR nova.compute.manager [instance: 825130d4-b1f6-434e-be3e-dca952abe930] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 738.758435] env[61957]: ERROR nova.compute.manager [instance: 825130d4-b1f6-434e-be3e-dca952abe930] return self._exit_event.wait() [ 738.758435] env[61957]: ERROR nova.compute.manager [instance: 825130d4-b1f6-434e-be3e-dca952abe930] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 738.758910] env[61957]: ERROR nova.compute.manager [instance: 825130d4-b1f6-434e-be3e-dca952abe930] current.throw(*self._exc) [ 738.758910] env[61957]: ERROR nova.compute.manager [instance: 825130d4-b1f6-434e-be3e-dca952abe930] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 738.758910] env[61957]: ERROR nova.compute.manager [instance: 825130d4-b1f6-434e-be3e-dca952abe930] result = function(*args, **kwargs) [ 738.758910] env[61957]: ERROR nova.compute.manager [instance: 825130d4-b1f6-434e-be3e-dca952abe930] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 738.758910] env[61957]: ERROR nova.compute.manager [instance: 825130d4-b1f6-434e-be3e-dca952abe930] return func(*args, **kwargs) [ 738.758910] env[61957]: ERROR nova.compute.manager [instance: 825130d4-b1f6-434e-be3e-dca952abe930] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 738.758910] env[61957]: ERROR nova.compute.manager [instance: 825130d4-b1f6-434e-be3e-dca952abe930] raise e [ 738.758910] env[61957]: ERROR nova.compute.manager [instance: 825130d4-b1f6-434e-be3e-dca952abe930] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 738.758910] env[61957]: ERROR nova.compute.manager [instance: 825130d4-b1f6-434e-be3e-dca952abe930] nwinfo = self.network_api.allocate_for_instance( [ 738.758910] env[61957]: ERROR nova.compute.manager [instance: 825130d4-b1f6-434e-be3e-dca952abe930] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 738.758910] env[61957]: ERROR nova.compute.manager [instance: 825130d4-b1f6-434e-be3e-dca952abe930] created_port_ids = self._update_ports_for_instance( [ 738.758910] env[61957]: ERROR nova.compute.manager [instance: 825130d4-b1f6-434e-be3e-dca952abe930] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 738.758910] env[61957]: ERROR nova.compute.manager [instance: 825130d4-b1f6-434e-be3e-dca952abe930] with excutils.save_and_reraise_exception(): [ 738.759281] env[61957]: ERROR nova.compute.manager [instance: 825130d4-b1f6-434e-be3e-dca952abe930] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 738.759281] env[61957]: ERROR nova.compute.manager [instance: 825130d4-b1f6-434e-be3e-dca952abe930] self.force_reraise() [ 738.759281] env[61957]: ERROR nova.compute.manager [instance: 825130d4-b1f6-434e-be3e-dca952abe930] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 738.759281] env[61957]: ERROR nova.compute.manager [instance: 825130d4-b1f6-434e-be3e-dca952abe930] raise self.value [ 738.759281] env[61957]: ERROR nova.compute.manager [instance: 825130d4-b1f6-434e-be3e-dca952abe930] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 738.759281] env[61957]: ERROR nova.compute.manager [instance: 825130d4-b1f6-434e-be3e-dca952abe930] updated_port = self._update_port( [ 738.759281] env[61957]: ERROR nova.compute.manager [instance: 825130d4-b1f6-434e-be3e-dca952abe930] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 738.759281] env[61957]: ERROR nova.compute.manager [instance: 825130d4-b1f6-434e-be3e-dca952abe930] _ensure_no_port_binding_failure(port) [ 738.759281] env[61957]: ERROR nova.compute.manager [instance: 825130d4-b1f6-434e-be3e-dca952abe930] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 738.759281] env[61957]: ERROR nova.compute.manager [instance: 825130d4-b1f6-434e-be3e-dca952abe930] raise exception.PortBindingFailed(port_id=port['id']) [ 738.759281] env[61957]: ERROR nova.compute.manager [instance: 825130d4-b1f6-434e-be3e-dca952abe930] nova.exception.PortBindingFailed: Binding failed for port b95b009e-bbe6-4159-bed7-0705d27090ee, please check neutron logs for more information. [ 738.759281] env[61957]: ERROR nova.compute.manager [instance: 825130d4-b1f6-434e-be3e-dca952abe930] [ 738.759603] env[61957]: DEBUG nova.compute.utils [None req-ca1e8d1e-8c18-4652-924a-ab1cdb751a05 tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] [instance: 825130d4-b1f6-434e-be3e-dca952abe930] Binding failed for port b95b009e-bbe6-4159-bed7-0705d27090ee, please check neutron logs for more information. {{(pid=61957) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 738.759842] env[61957]: DEBUG oslo_concurrency.lockutils [None req-58fe0400-f1d0-4e1d-a459-7541ba7aeeed tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 18.796s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 738.764621] env[61957]: DEBUG nova.compute.manager [None req-ca1e8d1e-8c18-4652-924a-ab1cdb751a05 tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] [instance: 825130d4-b1f6-434e-be3e-dca952abe930] Build of instance 825130d4-b1f6-434e-be3e-dca952abe930 was re-scheduled: Binding failed for port b95b009e-bbe6-4159-bed7-0705d27090ee, please check neutron logs for more information. {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 738.764621] env[61957]: DEBUG nova.compute.manager [None req-ca1e8d1e-8c18-4652-924a-ab1cdb751a05 tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] [instance: 825130d4-b1f6-434e-be3e-dca952abe930] Unplugging VIFs for instance {{(pid=61957) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 738.764621] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ca1e8d1e-8c18-4652-924a-ab1cdb751a05 tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] Acquiring lock "refresh_cache-825130d4-b1f6-434e-be3e-dca952abe930" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 738.764621] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ca1e8d1e-8c18-4652-924a-ab1cdb751a05 tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] Acquired lock "refresh_cache-825130d4-b1f6-434e-be3e-dca952abe930" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 738.764897] env[61957]: DEBUG nova.network.neutron [None req-ca1e8d1e-8c18-4652-924a-ab1cdb751a05 tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] [instance: 825130d4-b1f6-434e-be3e-dca952abe930] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 739.008103] env[61957]: INFO nova.compute.manager [None req-456a04ab-263c-4c55-898a-70a189651712 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] [instance: 6241fb44-ce1c-4881-bc32-7d19ddc83a61] Took 1.02 seconds to deallocate network for instance. [ 739.150826] env[61957]: DEBUG nova.compute.manager [None req-09f1eb26-3ef8-431d-92bc-9c62b712a196 tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] [instance: 60ec4271-2614-470d-bf40-47b9955f544c] Starting instance... {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 739.287071] env[61957]: DEBUG nova.network.neutron [None req-ca1e8d1e-8c18-4652-924a-ab1cdb751a05 tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] [instance: 825130d4-b1f6-434e-be3e-dca952abe930] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 739.373969] env[61957]: DEBUG nova.network.neutron [None req-ca1e8d1e-8c18-4652-924a-ab1cdb751a05 tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] [instance: 825130d4-b1f6-434e-be3e-dca952abe930] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 739.590946] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c808b926-03ca-4e5f-8904-57b01b0f3c4c {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.599235] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9caf1f06-aa00-42a9-b662-18634e26103c {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.632126] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-682e00b1-753f-4644-8720-9afdd209c86b {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.638589] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec7c4d77-2d10-41c7-a928-7e23c2304bc2 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.652404] env[61957]: DEBUG nova.compute.provider_tree [None req-58fe0400-f1d0-4e1d-a459-7541ba7aeeed tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 739.676321] env[61957]: DEBUG oslo_concurrency.lockutils [None req-09f1eb26-3ef8-431d-92bc-9c62b712a196 tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 739.877196] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ca1e8d1e-8c18-4652-924a-ab1cdb751a05 tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] Releasing lock "refresh_cache-825130d4-b1f6-434e-be3e-dca952abe930" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 739.877546] env[61957]: DEBUG nova.compute.manager [None req-ca1e8d1e-8c18-4652-924a-ab1cdb751a05 tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61957) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 739.877805] env[61957]: DEBUG nova.compute.manager [None req-ca1e8d1e-8c18-4652-924a-ab1cdb751a05 tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] [instance: 825130d4-b1f6-434e-be3e-dca952abe930] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 739.878055] env[61957]: DEBUG nova.network.neutron [None req-ca1e8d1e-8c18-4652-924a-ab1cdb751a05 tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] [instance: 825130d4-b1f6-434e-be3e-dca952abe930] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 739.894713] env[61957]: DEBUG nova.network.neutron [None req-ca1e8d1e-8c18-4652-924a-ab1cdb751a05 tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] [instance: 825130d4-b1f6-434e-be3e-dca952abe930] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 740.039685] env[61957]: INFO nova.scheduler.client.report [None req-456a04ab-263c-4c55-898a-70a189651712 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] Deleted allocations for instance 6241fb44-ce1c-4881-bc32-7d19ddc83a61 [ 740.096756] env[61957]: DEBUG oslo_concurrency.lockutils [None req-be584a65-2286-47ec-9ff6-29a7835a7d50 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Acquiring lock "37ce46a1-8f9a-4d15-bd81-e40845a0e48a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 740.096990] env[61957]: DEBUG oslo_concurrency.lockutils [None req-be584a65-2286-47ec-9ff6-29a7835a7d50 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Lock "37ce46a1-8f9a-4d15-bd81-e40845a0e48a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 740.157944] env[61957]: DEBUG nova.scheduler.client.report [None req-58fe0400-f1d0-4e1d-a459-7541ba7aeeed tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 740.396933] env[61957]: DEBUG nova.network.neutron [None req-ca1e8d1e-8c18-4652-924a-ab1cdb751a05 tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] [instance: 825130d4-b1f6-434e-be3e-dca952abe930] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 740.547397] env[61957]: DEBUG oslo_concurrency.lockutils [None req-456a04ab-263c-4c55-898a-70a189651712 tempest-ServersAdminTestJSON-1995235418 tempest-ServersAdminTestJSON-1995235418-project-member] Lock "6241fb44-ce1c-4881-bc32-7d19ddc83a61" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 127.016s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 740.663197] env[61957]: DEBUG oslo_concurrency.lockutils [None req-58fe0400-f1d0-4e1d-a459-7541ba7aeeed tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.903s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 740.663843] env[61957]: ERROR nova.compute.manager [None req-58fe0400-f1d0-4e1d-a459-7541ba7aeeed tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] [instance: c369d148-2093-4724-82cc-d80a67131fea] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 2fe40c60-9f72-4975-a0f4-fe5d0e4b3b72, please check neutron logs for more information. [ 740.663843] env[61957]: ERROR nova.compute.manager [instance: c369d148-2093-4724-82cc-d80a67131fea] Traceback (most recent call last): [ 740.663843] env[61957]: ERROR nova.compute.manager [instance: c369d148-2093-4724-82cc-d80a67131fea] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 740.663843] env[61957]: ERROR nova.compute.manager [instance: c369d148-2093-4724-82cc-d80a67131fea] self.driver.spawn(context, instance, image_meta, [ 740.663843] env[61957]: ERROR nova.compute.manager [instance: c369d148-2093-4724-82cc-d80a67131fea] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 740.663843] env[61957]: ERROR nova.compute.manager [instance: c369d148-2093-4724-82cc-d80a67131fea] self._vmops.spawn(context, instance, image_meta, injected_files, [ 740.663843] env[61957]: ERROR nova.compute.manager [instance: c369d148-2093-4724-82cc-d80a67131fea] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 740.663843] env[61957]: ERROR nova.compute.manager [instance: c369d148-2093-4724-82cc-d80a67131fea] vm_ref = self.build_virtual_machine(instance, [ 740.663843] env[61957]: ERROR nova.compute.manager [instance: c369d148-2093-4724-82cc-d80a67131fea] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 740.663843] env[61957]: ERROR nova.compute.manager [instance: c369d148-2093-4724-82cc-d80a67131fea] vif_infos = vmwarevif.get_vif_info(self._session, [ 740.663843] env[61957]: ERROR nova.compute.manager [instance: c369d148-2093-4724-82cc-d80a67131fea] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 740.664231] env[61957]: ERROR nova.compute.manager [instance: c369d148-2093-4724-82cc-d80a67131fea] for vif in network_info: [ 740.664231] env[61957]: ERROR nova.compute.manager [instance: c369d148-2093-4724-82cc-d80a67131fea] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 740.664231] env[61957]: ERROR nova.compute.manager [instance: c369d148-2093-4724-82cc-d80a67131fea] return self._sync_wrapper(fn, *args, **kwargs) [ 740.664231] env[61957]: ERROR nova.compute.manager [instance: c369d148-2093-4724-82cc-d80a67131fea] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 740.664231] env[61957]: ERROR nova.compute.manager [instance: c369d148-2093-4724-82cc-d80a67131fea] self.wait() [ 740.664231] env[61957]: ERROR nova.compute.manager [instance: c369d148-2093-4724-82cc-d80a67131fea] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 740.664231] env[61957]: ERROR nova.compute.manager [instance: c369d148-2093-4724-82cc-d80a67131fea] self[:] = self._gt.wait() [ 740.664231] env[61957]: ERROR nova.compute.manager [instance: c369d148-2093-4724-82cc-d80a67131fea] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 740.664231] env[61957]: ERROR nova.compute.manager [instance: c369d148-2093-4724-82cc-d80a67131fea] return self._exit_event.wait() [ 740.664231] env[61957]: ERROR nova.compute.manager [instance: c369d148-2093-4724-82cc-d80a67131fea] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 740.664231] env[61957]: ERROR nova.compute.manager [instance: c369d148-2093-4724-82cc-d80a67131fea] current.throw(*self._exc) [ 740.664231] env[61957]: ERROR nova.compute.manager [instance: c369d148-2093-4724-82cc-d80a67131fea] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 740.664231] env[61957]: ERROR nova.compute.manager [instance: c369d148-2093-4724-82cc-d80a67131fea] result = function(*args, **kwargs) [ 740.664625] env[61957]: ERROR nova.compute.manager [instance: c369d148-2093-4724-82cc-d80a67131fea] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 740.664625] env[61957]: ERROR nova.compute.manager [instance: c369d148-2093-4724-82cc-d80a67131fea] return func(*args, **kwargs) [ 740.664625] env[61957]: ERROR nova.compute.manager [instance: c369d148-2093-4724-82cc-d80a67131fea] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 740.664625] env[61957]: ERROR nova.compute.manager [instance: c369d148-2093-4724-82cc-d80a67131fea] raise e [ 740.664625] env[61957]: ERROR nova.compute.manager [instance: c369d148-2093-4724-82cc-d80a67131fea] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 740.664625] env[61957]: ERROR nova.compute.manager [instance: c369d148-2093-4724-82cc-d80a67131fea] nwinfo = self.network_api.allocate_for_instance( [ 740.664625] env[61957]: ERROR nova.compute.manager [instance: c369d148-2093-4724-82cc-d80a67131fea] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 740.664625] env[61957]: ERROR nova.compute.manager [instance: c369d148-2093-4724-82cc-d80a67131fea] created_port_ids = self._update_ports_for_instance( [ 740.664625] env[61957]: ERROR nova.compute.manager [instance: c369d148-2093-4724-82cc-d80a67131fea] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 740.664625] env[61957]: ERROR nova.compute.manager [instance: c369d148-2093-4724-82cc-d80a67131fea] with excutils.save_and_reraise_exception(): [ 740.664625] env[61957]: ERROR nova.compute.manager [instance: c369d148-2093-4724-82cc-d80a67131fea] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 740.664625] env[61957]: ERROR nova.compute.manager [instance: c369d148-2093-4724-82cc-d80a67131fea] self.force_reraise() [ 740.664625] env[61957]: ERROR nova.compute.manager [instance: c369d148-2093-4724-82cc-d80a67131fea] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 740.665027] env[61957]: ERROR nova.compute.manager [instance: c369d148-2093-4724-82cc-d80a67131fea] raise self.value [ 740.665027] env[61957]: ERROR nova.compute.manager [instance: c369d148-2093-4724-82cc-d80a67131fea] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 740.665027] env[61957]: ERROR nova.compute.manager [instance: c369d148-2093-4724-82cc-d80a67131fea] updated_port = self._update_port( [ 740.665027] env[61957]: ERROR nova.compute.manager [instance: c369d148-2093-4724-82cc-d80a67131fea] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 740.665027] env[61957]: ERROR nova.compute.manager [instance: c369d148-2093-4724-82cc-d80a67131fea] _ensure_no_port_binding_failure(port) [ 740.665027] env[61957]: ERROR nova.compute.manager [instance: c369d148-2093-4724-82cc-d80a67131fea] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 740.665027] env[61957]: ERROR nova.compute.manager [instance: c369d148-2093-4724-82cc-d80a67131fea] raise exception.PortBindingFailed(port_id=port['id']) [ 740.665027] env[61957]: ERROR nova.compute.manager [instance: c369d148-2093-4724-82cc-d80a67131fea] nova.exception.PortBindingFailed: Binding failed for port 2fe40c60-9f72-4975-a0f4-fe5d0e4b3b72, please check neutron logs for more information. [ 740.665027] env[61957]: ERROR nova.compute.manager [instance: c369d148-2093-4724-82cc-d80a67131fea] [ 740.665027] env[61957]: DEBUG nova.compute.utils [None req-58fe0400-f1d0-4e1d-a459-7541ba7aeeed tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] [instance: c369d148-2093-4724-82cc-d80a67131fea] Binding failed for port 2fe40c60-9f72-4975-a0f4-fe5d0e4b3b72, please check neutron logs for more information. {{(pid=61957) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 740.666151] env[61957]: DEBUG nova.compute.manager [None req-58fe0400-f1d0-4e1d-a459-7541ba7aeeed tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] [instance: c369d148-2093-4724-82cc-d80a67131fea] Build of instance c369d148-2093-4724-82cc-d80a67131fea was re-scheduled: Binding failed for port 2fe40c60-9f72-4975-a0f4-fe5d0e4b3b72, please check neutron logs for more information. {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 740.666600] env[61957]: DEBUG nova.compute.manager [None req-58fe0400-f1d0-4e1d-a459-7541ba7aeeed tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] [instance: c369d148-2093-4724-82cc-d80a67131fea] Unplugging VIFs for instance {{(pid=61957) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 740.666834] env[61957]: DEBUG oslo_concurrency.lockutils [None req-58fe0400-f1d0-4e1d-a459-7541ba7aeeed tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] Acquiring lock "refresh_cache-c369d148-2093-4724-82cc-d80a67131fea" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 740.666982] env[61957]: DEBUG oslo_concurrency.lockutils [None req-58fe0400-f1d0-4e1d-a459-7541ba7aeeed tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] Acquired lock "refresh_cache-c369d148-2093-4724-82cc-d80a67131fea" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 740.667152] env[61957]: DEBUG nova.network.neutron [None req-58fe0400-f1d0-4e1d-a459-7541ba7aeeed tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] [instance: c369d148-2093-4724-82cc-d80a67131fea] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 740.668152] env[61957]: DEBUG oslo_concurrency.lockutils [None req-5ca5fdda-dbef-4c48-b38f-9d2aa78e9841 tempest-FloatingIPsAssociationNegativeTestJSON-572957506 tempest-FloatingIPsAssociationNegativeTestJSON-572957506-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.338s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 740.671627] env[61957]: INFO nova.compute.claims [None req-5ca5fdda-dbef-4c48-b38f-9d2aa78e9841 tempest-FloatingIPsAssociationNegativeTestJSON-572957506 tempest-FloatingIPsAssociationNegativeTestJSON-572957506-project-member] [instance: 348ea580-2268-4082-a5f1-33c954063e4d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 740.899794] env[61957]: INFO nova.compute.manager [None req-ca1e8d1e-8c18-4652-924a-ab1cdb751a05 tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] [instance: 825130d4-b1f6-434e-be3e-dca952abe930] Took 1.02 seconds to deallocate network for instance. [ 741.050983] env[61957]: DEBUG nova.compute.manager [None req-6b91bce8-d0f8-4c40-855e-ed34f3eb7e64 tempest-ServerActionsTestJSON-1567241756 tempest-ServerActionsTestJSON-1567241756-project-member] [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] Starting instance... {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 741.193590] env[61957]: DEBUG nova.network.neutron [None req-58fe0400-f1d0-4e1d-a459-7541ba7aeeed tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] [instance: c369d148-2093-4724-82cc-d80a67131fea] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 741.270058] env[61957]: DEBUG nova.network.neutron [None req-58fe0400-f1d0-4e1d-a459-7541ba7aeeed tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] [instance: c369d148-2093-4724-82cc-d80a67131fea] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 741.573771] env[61957]: DEBUG oslo_concurrency.lockutils [None req-6b91bce8-d0f8-4c40-855e-ed34f3eb7e64 tempest-ServerActionsTestJSON-1567241756 tempest-ServerActionsTestJSON-1567241756-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 741.773061] env[61957]: DEBUG oslo_concurrency.lockutils [None req-58fe0400-f1d0-4e1d-a459-7541ba7aeeed tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] Releasing lock "refresh_cache-c369d148-2093-4724-82cc-d80a67131fea" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 741.773474] env[61957]: DEBUG nova.compute.manager [None req-58fe0400-f1d0-4e1d-a459-7541ba7aeeed tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61957) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 741.773748] env[61957]: DEBUG nova.compute.manager [None req-58fe0400-f1d0-4e1d-a459-7541ba7aeeed tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] [instance: c369d148-2093-4724-82cc-d80a67131fea] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 741.773959] env[61957]: DEBUG nova.network.neutron [None req-58fe0400-f1d0-4e1d-a459-7541ba7aeeed tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] [instance: c369d148-2093-4724-82cc-d80a67131fea] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 741.790823] env[61957]: DEBUG nova.network.neutron [None req-58fe0400-f1d0-4e1d-a459-7541ba7aeeed tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] [instance: c369d148-2093-4724-82cc-d80a67131fea] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 741.925958] env[61957]: INFO nova.scheduler.client.report [None req-ca1e8d1e-8c18-4652-924a-ab1cdb751a05 tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] Deleted allocations for instance 825130d4-b1f6-434e-be3e-dca952abe930 [ 742.002960] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c67f838-49fa-4877-ab8c-b6ab23ac1d01 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.010574] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b86ab552-315a-4033-92ce-d56a184fbca8 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.040823] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4fcbc8e-1597-4343-b7f2-1b893cd416f1 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.048019] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-898fe4f8-9521-40cb-8591-7b5ab23af2d0 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.060807] env[61957]: DEBUG nova.compute.provider_tree [None req-5ca5fdda-dbef-4c48-b38f-9d2aa78e9841 tempest-FloatingIPsAssociationNegativeTestJSON-572957506 tempest-FloatingIPsAssociationNegativeTestJSON-572957506-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 742.293272] env[61957]: DEBUG nova.network.neutron [None req-58fe0400-f1d0-4e1d-a459-7541ba7aeeed tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] [instance: c369d148-2093-4724-82cc-d80a67131fea] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 742.440842] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ca1e8d1e-8c18-4652-924a-ab1cdb751a05 tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] Lock "825130d4-b1f6-434e-be3e-dca952abe930" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 116.486s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 742.564518] env[61957]: DEBUG nova.scheduler.client.report [None req-5ca5fdda-dbef-4c48-b38f-9d2aa78e9841 tempest-FloatingIPsAssociationNegativeTestJSON-572957506 tempest-FloatingIPsAssociationNegativeTestJSON-572957506-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 742.797491] env[61957]: INFO nova.compute.manager [None req-58fe0400-f1d0-4e1d-a459-7541ba7aeeed tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] [instance: c369d148-2093-4724-82cc-d80a67131fea] Took 1.02 seconds to deallocate network for instance. [ 742.944992] env[61957]: DEBUG nova.compute.manager [None req-660095f0-9d5a-4797-b621-e0585ffddbba tempest-ServerActionsTestOtherB-303558607 tempest-ServerActionsTestOtherB-303558607-project-member] [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] Starting instance... {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 743.073924] env[61957]: DEBUG oslo_concurrency.lockutils [None req-5ca5fdda-dbef-4c48-b38f-9d2aa78e9841 tempest-FloatingIPsAssociationNegativeTestJSON-572957506 tempest-FloatingIPsAssociationNegativeTestJSON-572957506-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.406s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 743.074481] env[61957]: DEBUG nova.compute.manager [None req-5ca5fdda-dbef-4c48-b38f-9d2aa78e9841 tempest-FloatingIPsAssociationNegativeTestJSON-572957506 tempest-FloatingIPsAssociationNegativeTestJSON-572957506-project-member] [instance: 348ea580-2268-4082-a5f1-33c954063e4d] Start building networks asynchronously for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 743.076993] env[61957]: DEBUG oslo_concurrency.lockutils [None req-707bfa11-3857-45e5-a4be-581ffade0e04 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.950s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 743.468330] env[61957]: DEBUG oslo_concurrency.lockutils [None req-660095f0-9d5a-4797-b621-e0585ffddbba tempest-ServerActionsTestOtherB-303558607 tempest-ServerActionsTestOtherB-303558607-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 743.581949] env[61957]: DEBUG nova.compute.utils [None req-5ca5fdda-dbef-4c48-b38f-9d2aa78e9841 tempest-FloatingIPsAssociationNegativeTestJSON-572957506 tempest-FloatingIPsAssociationNegativeTestJSON-572957506-project-member] Using /dev/sd instead of None {{(pid=61957) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 743.587014] env[61957]: DEBUG nova.compute.manager [None req-5ca5fdda-dbef-4c48-b38f-9d2aa78e9841 tempest-FloatingIPsAssociationNegativeTestJSON-572957506 tempest-FloatingIPsAssociationNegativeTestJSON-572957506-project-member] [instance: 348ea580-2268-4082-a5f1-33c954063e4d] Allocating IP information in the background. {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 743.587014] env[61957]: DEBUG nova.network.neutron [None req-5ca5fdda-dbef-4c48-b38f-9d2aa78e9841 tempest-FloatingIPsAssociationNegativeTestJSON-572957506 tempest-FloatingIPsAssociationNegativeTestJSON-572957506-project-member] [instance: 348ea580-2268-4082-a5f1-33c954063e4d] allocate_for_instance() {{(pid=61957) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 743.642804] env[61957]: DEBUG nova.policy [None req-5ca5fdda-dbef-4c48-b38f-9d2aa78e9841 tempest-FloatingIPsAssociationNegativeTestJSON-572957506 tempest-FloatingIPsAssociationNegativeTestJSON-572957506-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '361798e45172423ab5f325e7a19b23dd', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2381420424ec42ed96c00b3a284fa850', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61957) authorize /opt/stack/nova/nova/policy.py:203}} [ 743.825168] env[61957]: INFO nova.scheduler.client.report [None req-58fe0400-f1d0-4e1d-a459-7541ba7aeeed tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] Deleted allocations for instance c369d148-2093-4724-82cc-d80a67131fea [ 743.915892] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa06fb4c-4bab-4148-8e6d-32bff989dec6 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.923998] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49dad274-1850-40e4-8cc9-0b6bd52972b7 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.960656] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c14b1818-525d-4d23-8c74-1ea21b9ec784 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.969658] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d7f4e77-ac40-4944-af5c-5176861764fa {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.984607] env[61957]: DEBUG nova.compute.provider_tree [None req-707bfa11-3857-45e5-a4be-581ffade0e04 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 744.029650] env[61957]: DEBUG oslo_concurrency.lockutils [None req-227f29c0-ef6f-447c-b621-3f65bc420b4e tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] Acquiring lock "37a9e1c1-9f6c-4047-9ff5-e141d0ca383e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 744.029915] env[61957]: DEBUG oslo_concurrency.lockutils [None req-227f29c0-ef6f-447c-b621-3f65bc420b4e tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] Lock "37a9e1c1-9f6c-4047-9ff5-e141d0ca383e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 744.086828] env[61957]: DEBUG nova.compute.manager [None req-5ca5fdda-dbef-4c48-b38f-9d2aa78e9841 tempest-FloatingIPsAssociationNegativeTestJSON-572957506 tempest-FloatingIPsAssociationNegativeTestJSON-572957506-project-member] [instance: 348ea580-2268-4082-a5f1-33c954063e4d] Start building block device mappings for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 744.111609] env[61957]: DEBUG nova.network.neutron [None req-5ca5fdda-dbef-4c48-b38f-9d2aa78e9841 tempest-FloatingIPsAssociationNegativeTestJSON-572957506 tempest-FloatingIPsAssociationNegativeTestJSON-572957506-project-member] [instance: 348ea580-2268-4082-a5f1-33c954063e4d] Successfully created port: 89bc10a8-611c-42c9-ac0f-b4d98e2005d2 {{(pid=61957) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 744.335138] env[61957]: DEBUG oslo_concurrency.lockutils [None req-58fe0400-f1d0-4e1d-a459-7541ba7aeeed tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] Lock "c369d148-2093-4724-82cc-d80a67131fea" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 116.573s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 744.487663] env[61957]: DEBUG nova.scheduler.client.report [None req-707bfa11-3857-45e5-a4be-581ffade0e04 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 744.842994] env[61957]: DEBUG nova.compute.manager [None req-3d8587e3-7ad5-40a6-88af-e8ed44c49baa tempest-ServerActionsV293TestJSON-1269989829 tempest-ServerActionsV293TestJSON-1269989829-project-member] [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] Starting instance... {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 744.850422] env[61957]: DEBUG nova.compute.manager [req-fc90a278-48de-43b1-9689-d094a9b2b723 req-b810bf8a-537c-43df-b7ec-b66f8303692f service nova] [instance: 348ea580-2268-4082-a5f1-33c954063e4d] Received event network-changed-89bc10a8-611c-42c9-ac0f-b4d98e2005d2 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 744.850616] env[61957]: DEBUG nova.compute.manager [req-fc90a278-48de-43b1-9689-d094a9b2b723 req-b810bf8a-537c-43df-b7ec-b66f8303692f service nova] [instance: 348ea580-2268-4082-a5f1-33c954063e4d] Refreshing instance network info cache due to event network-changed-89bc10a8-611c-42c9-ac0f-b4d98e2005d2. {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 744.850853] env[61957]: DEBUG oslo_concurrency.lockutils [req-fc90a278-48de-43b1-9689-d094a9b2b723 req-b810bf8a-537c-43df-b7ec-b66f8303692f service nova] Acquiring lock "refresh_cache-348ea580-2268-4082-a5f1-33c954063e4d" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 744.850994] env[61957]: DEBUG oslo_concurrency.lockutils [req-fc90a278-48de-43b1-9689-d094a9b2b723 req-b810bf8a-537c-43df-b7ec-b66f8303692f service nova] Acquired lock "refresh_cache-348ea580-2268-4082-a5f1-33c954063e4d" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 744.851166] env[61957]: DEBUG nova.network.neutron [req-fc90a278-48de-43b1-9689-d094a9b2b723 req-b810bf8a-537c-43df-b7ec-b66f8303692f service nova] [instance: 348ea580-2268-4082-a5f1-33c954063e4d] Refreshing network info cache for port 89bc10a8-611c-42c9-ac0f-b4d98e2005d2 {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 744.994900] env[61957]: DEBUG oslo_concurrency.lockutils [None req-707bfa11-3857-45e5-a4be-581ffade0e04 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.916s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 744.994900] env[61957]: ERROR nova.compute.manager [None req-707bfa11-3857-45e5-a4be-581ffade0e04 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 9392bc30-2697-41c7-83dd-b8963c4dc6c3, please check neutron logs for more information. [ 744.994900] env[61957]: ERROR nova.compute.manager [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] Traceback (most recent call last): [ 744.994900] env[61957]: ERROR nova.compute.manager [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 744.994900] env[61957]: ERROR nova.compute.manager [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] self.driver.spawn(context, instance, image_meta, [ 744.994900] env[61957]: ERROR nova.compute.manager [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 744.994900] env[61957]: ERROR nova.compute.manager [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] self._vmops.spawn(context, instance, image_meta, injected_files, [ 744.994900] env[61957]: ERROR nova.compute.manager [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 744.994900] env[61957]: ERROR nova.compute.manager [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] vm_ref = self.build_virtual_machine(instance, [ 744.995420] env[61957]: ERROR nova.compute.manager [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 744.995420] env[61957]: ERROR nova.compute.manager [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] vif_infos = vmwarevif.get_vif_info(self._session, [ 744.995420] env[61957]: ERROR nova.compute.manager [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 744.995420] env[61957]: ERROR nova.compute.manager [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] for vif in network_info: [ 744.995420] env[61957]: ERROR nova.compute.manager [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 744.995420] env[61957]: ERROR nova.compute.manager [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] return self._sync_wrapper(fn, *args, **kwargs) [ 744.995420] env[61957]: ERROR nova.compute.manager [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 744.995420] env[61957]: ERROR nova.compute.manager [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] self.wait() [ 744.995420] env[61957]: ERROR nova.compute.manager [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 744.995420] env[61957]: ERROR nova.compute.manager [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] self[:] = self._gt.wait() [ 744.995420] env[61957]: ERROR nova.compute.manager [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 744.995420] env[61957]: ERROR nova.compute.manager [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] return self._exit_event.wait() [ 744.995420] env[61957]: ERROR nova.compute.manager [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 744.995758] env[61957]: ERROR nova.compute.manager [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] current.throw(*self._exc) [ 744.995758] env[61957]: ERROR nova.compute.manager [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 744.995758] env[61957]: ERROR nova.compute.manager [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] result = function(*args, **kwargs) [ 744.995758] env[61957]: ERROR nova.compute.manager [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 744.995758] env[61957]: ERROR nova.compute.manager [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] return func(*args, **kwargs) [ 744.995758] env[61957]: ERROR nova.compute.manager [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 744.995758] env[61957]: ERROR nova.compute.manager [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] raise e [ 744.995758] env[61957]: ERROR nova.compute.manager [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 744.995758] env[61957]: ERROR nova.compute.manager [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] nwinfo = self.network_api.allocate_for_instance( [ 744.995758] env[61957]: ERROR nova.compute.manager [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 744.995758] env[61957]: ERROR nova.compute.manager [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] created_port_ids = self._update_ports_for_instance( [ 744.995758] env[61957]: ERROR nova.compute.manager [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 744.995758] env[61957]: ERROR nova.compute.manager [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] with excutils.save_and_reraise_exception(): [ 744.996125] env[61957]: ERROR nova.compute.manager [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 744.996125] env[61957]: ERROR nova.compute.manager [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] self.force_reraise() [ 744.996125] env[61957]: ERROR nova.compute.manager [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 744.996125] env[61957]: ERROR nova.compute.manager [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] raise self.value [ 744.996125] env[61957]: ERROR nova.compute.manager [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 744.996125] env[61957]: ERROR nova.compute.manager [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] updated_port = self._update_port( [ 744.996125] env[61957]: ERROR nova.compute.manager [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 744.996125] env[61957]: ERROR nova.compute.manager [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] _ensure_no_port_binding_failure(port) [ 744.996125] env[61957]: ERROR nova.compute.manager [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 744.996125] env[61957]: ERROR nova.compute.manager [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] raise exception.PortBindingFailed(port_id=port['id']) [ 744.996125] env[61957]: ERROR nova.compute.manager [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] nova.exception.PortBindingFailed: Binding failed for port 9392bc30-2697-41c7-83dd-b8963c4dc6c3, please check neutron logs for more information. [ 744.996125] env[61957]: ERROR nova.compute.manager [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] [ 744.996447] env[61957]: DEBUG nova.compute.utils [None req-707bfa11-3857-45e5-a4be-581ffade0e04 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] Binding failed for port 9392bc30-2697-41c7-83dd-b8963c4dc6c3, please check neutron logs for more information. {{(pid=61957) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 744.996447] env[61957]: DEBUG oslo_concurrency.lockutils [None req-838e8ad7-4b6c-4dbc-8d4b-8380b652e559 tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.219s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 744.997451] env[61957]: INFO nova.compute.claims [None req-838e8ad7-4b6c-4dbc-8d4b-8380b652e559 tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] [instance: 042425a0-5398-47a6-9842-380d5b3299c0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 745.001461] env[61957]: DEBUG nova.compute.manager [None req-707bfa11-3857-45e5-a4be-581ffade0e04 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] Build of instance 2252ef6e-19ec-4356-8221-33168c01bef0 was re-scheduled: Binding failed for port 9392bc30-2697-41c7-83dd-b8963c4dc6c3, please check neutron logs for more information. {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 745.001461] env[61957]: DEBUG nova.compute.manager [None req-707bfa11-3857-45e5-a4be-581ffade0e04 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] Unplugging VIFs for instance {{(pid=61957) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 745.001461] env[61957]: DEBUG oslo_concurrency.lockutils [None req-707bfa11-3857-45e5-a4be-581ffade0e04 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Acquiring lock "refresh_cache-2252ef6e-19ec-4356-8221-33168c01bef0" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 745.001461] env[61957]: DEBUG oslo_concurrency.lockutils [None req-707bfa11-3857-45e5-a4be-581ffade0e04 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Acquired lock "refresh_cache-2252ef6e-19ec-4356-8221-33168c01bef0" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 745.001461] env[61957]: DEBUG nova.network.neutron [None req-707bfa11-3857-45e5-a4be-581ffade0e04 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 745.011350] env[61957]: ERROR nova.compute.manager [None req-5ca5fdda-dbef-4c48-b38f-9d2aa78e9841 tempest-FloatingIPsAssociationNegativeTestJSON-572957506 tempest-FloatingIPsAssociationNegativeTestJSON-572957506-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 89bc10a8-611c-42c9-ac0f-b4d98e2005d2, please check neutron logs for more information. [ 745.011350] env[61957]: ERROR nova.compute.manager Traceback (most recent call last): [ 745.011350] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 745.011350] env[61957]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 745.011350] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 745.011350] env[61957]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 745.011350] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 745.011350] env[61957]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 745.011350] env[61957]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 745.011350] env[61957]: ERROR nova.compute.manager self.force_reraise() [ 745.011350] env[61957]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 745.011350] env[61957]: ERROR nova.compute.manager raise self.value [ 745.011350] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 745.011350] env[61957]: ERROR nova.compute.manager updated_port = self._update_port( [ 745.011350] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 745.011350] env[61957]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 745.011885] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 745.011885] env[61957]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 745.011885] env[61957]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 89bc10a8-611c-42c9-ac0f-b4d98e2005d2, please check neutron logs for more information. [ 745.011885] env[61957]: ERROR nova.compute.manager [ 745.011885] env[61957]: Traceback (most recent call last): [ 745.011885] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 745.011885] env[61957]: listener.cb(fileno) [ 745.011885] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 745.011885] env[61957]: result = function(*args, **kwargs) [ 745.011885] env[61957]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 745.011885] env[61957]: return func(*args, **kwargs) [ 745.011885] env[61957]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 745.011885] env[61957]: raise e [ 745.011885] env[61957]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 745.011885] env[61957]: nwinfo = self.network_api.allocate_for_instance( [ 745.011885] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 745.011885] env[61957]: created_port_ids = self._update_ports_for_instance( [ 745.011885] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 745.011885] env[61957]: with excutils.save_and_reraise_exception(): [ 745.011885] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 745.011885] env[61957]: self.force_reraise() [ 745.011885] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 745.011885] env[61957]: raise self.value [ 745.011885] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 745.011885] env[61957]: updated_port = self._update_port( [ 745.011885] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 745.011885] env[61957]: _ensure_no_port_binding_failure(port) [ 745.011885] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 745.011885] env[61957]: raise exception.PortBindingFailed(port_id=port['id']) [ 745.012739] env[61957]: nova.exception.PortBindingFailed: Binding failed for port 89bc10a8-611c-42c9-ac0f-b4d98e2005d2, please check neutron logs for more information. [ 745.012739] env[61957]: Removing descriptor: 16 [ 745.097623] env[61957]: DEBUG nova.compute.manager [None req-5ca5fdda-dbef-4c48-b38f-9d2aa78e9841 tempest-FloatingIPsAssociationNegativeTestJSON-572957506 tempest-FloatingIPsAssociationNegativeTestJSON-572957506-project-member] [instance: 348ea580-2268-4082-a5f1-33c954063e4d] Start spawning the instance on the hypervisor. {{(pid=61957) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 745.122266] env[61957]: DEBUG nova.virt.hardware [None req-5ca5fdda-dbef-4c48-b38f-9d2aa78e9841 tempest-FloatingIPsAssociationNegativeTestJSON-572957506 tempest-FloatingIPsAssociationNegativeTestJSON-572957506-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T17:22:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T17:21:50Z,direct_url=,disk_format='vmdk',id=11c76a2c-f705-470a-ba9d-4657858bab38,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='81d3096d0f094373913fc3dc8f5c3c6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T17:21:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 745.122390] env[61957]: DEBUG nova.virt.hardware [None req-5ca5fdda-dbef-4c48-b38f-9d2aa78e9841 tempest-FloatingIPsAssociationNegativeTestJSON-572957506 tempest-FloatingIPsAssociationNegativeTestJSON-572957506-project-member] Flavor limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 745.122521] env[61957]: DEBUG nova.virt.hardware [None req-5ca5fdda-dbef-4c48-b38f-9d2aa78e9841 tempest-FloatingIPsAssociationNegativeTestJSON-572957506 tempest-FloatingIPsAssociationNegativeTestJSON-572957506-project-member] Image limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 745.122711] env[61957]: DEBUG nova.virt.hardware [None req-5ca5fdda-dbef-4c48-b38f-9d2aa78e9841 tempest-FloatingIPsAssociationNegativeTestJSON-572957506 tempest-FloatingIPsAssociationNegativeTestJSON-572957506-project-member] Flavor pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 745.122858] env[61957]: DEBUG nova.virt.hardware [None req-5ca5fdda-dbef-4c48-b38f-9d2aa78e9841 tempest-FloatingIPsAssociationNegativeTestJSON-572957506 tempest-FloatingIPsAssociationNegativeTestJSON-572957506-project-member] Image pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 745.123059] env[61957]: DEBUG nova.virt.hardware [None req-5ca5fdda-dbef-4c48-b38f-9d2aa78e9841 tempest-FloatingIPsAssociationNegativeTestJSON-572957506 tempest-FloatingIPsAssociationNegativeTestJSON-572957506-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 745.124179] env[61957]: DEBUG nova.virt.hardware [None req-5ca5fdda-dbef-4c48-b38f-9d2aa78e9841 tempest-FloatingIPsAssociationNegativeTestJSON-572957506 tempest-FloatingIPsAssociationNegativeTestJSON-572957506-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 745.124367] env[61957]: DEBUG nova.virt.hardware [None req-5ca5fdda-dbef-4c48-b38f-9d2aa78e9841 tempest-FloatingIPsAssociationNegativeTestJSON-572957506 tempest-FloatingIPsAssociationNegativeTestJSON-572957506-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 745.124638] env[61957]: DEBUG nova.virt.hardware [None req-5ca5fdda-dbef-4c48-b38f-9d2aa78e9841 tempest-FloatingIPsAssociationNegativeTestJSON-572957506 tempest-FloatingIPsAssociationNegativeTestJSON-572957506-project-member] Got 1 possible topologies {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 745.124732] env[61957]: DEBUG nova.virt.hardware [None req-5ca5fdda-dbef-4c48-b38f-9d2aa78e9841 tempest-FloatingIPsAssociationNegativeTestJSON-572957506 tempest-FloatingIPsAssociationNegativeTestJSON-572957506-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 745.124906] env[61957]: DEBUG nova.virt.hardware [None req-5ca5fdda-dbef-4c48-b38f-9d2aa78e9841 tempest-FloatingIPsAssociationNegativeTestJSON-572957506 tempest-FloatingIPsAssociationNegativeTestJSON-572957506-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 745.125824] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4b7b311-37a0-4fd8-8128-cfb34c73b526 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.133792] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f0666b2-1578-4757-aa5c-77b18a106825 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.147855] env[61957]: ERROR nova.compute.manager [None req-5ca5fdda-dbef-4c48-b38f-9d2aa78e9841 tempest-FloatingIPsAssociationNegativeTestJSON-572957506 tempest-FloatingIPsAssociationNegativeTestJSON-572957506-project-member] [instance: 348ea580-2268-4082-a5f1-33c954063e4d] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 89bc10a8-611c-42c9-ac0f-b4d98e2005d2, please check neutron logs for more information. [ 745.147855] env[61957]: ERROR nova.compute.manager [instance: 348ea580-2268-4082-a5f1-33c954063e4d] Traceback (most recent call last): [ 745.147855] env[61957]: ERROR nova.compute.manager [instance: 348ea580-2268-4082-a5f1-33c954063e4d] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 745.147855] env[61957]: ERROR nova.compute.manager [instance: 348ea580-2268-4082-a5f1-33c954063e4d] yield resources [ 745.147855] env[61957]: ERROR nova.compute.manager [instance: 348ea580-2268-4082-a5f1-33c954063e4d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 745.147855] env[61957]: ERROR nova.compute.manager [instance: 348ea580-2268-4082-a5f1-33c954063e4d] self.driver.spawn(context, instance, image_meta, [ 745.147855] env[61957]: ERROR nova.compute.manager [instance: 348ea580-2268-4082-a5f1-33c954063e4d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 745.147855] env[61957]: ERROR nova.compute.manager [instance: 348ea580-2268-4082-a5f1-33c954063e4d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 745.147855] env[61957]: ERROR nova.compute.manager [instance: 348ea580-2268-4082-a5f1-33c954063e4d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 745.147855] env[61957]: ERROR nova.compute.manager [instance: 348ea580-2268-4082-a5f1-33c954063e4d] vm_ref = self.build_virtual_machine(instance, [ 745.147855] env[61957]: ERROR nova.compute.manager [instance: 348ea580-2268-4082-a5f1-33c954063e4d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 745.148338] env[61957]: ERROR nova.compute.manager [instance: 348ea580-2268-4082-a5f1-33c954063e4d] vif_infos = vmwarevif.get_vif_info(self._session, [ 745.148338] env[61957]: ERROR nova.compute.manager [instance: 348ea580-2268-4082-a5f1-33c954063e4d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 745.148338] env[61957]: ERROR nova.compute.manager [instance: 348ea580-2268-4082-a5f1-33c954063e4d] for vif in network_info: [ 745.148338] env[61957]: ERROR nova.compute.manager [instance: 348ea580-2268-4082-a5f1-33c954063e4d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 745.148338] env[61957]: ERROR nova.compute.manager [instance: 348ea580-2268-4082-a5f1-33c954063e4d] return self._sync_wrapper(fn, *args, **kwargs) [ 745.148338] env[61957]: ERROR nova.compute.manager [instance: 348ea580-2268-4082-a5f1-33c954063e4d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 745.148338] env[61957]: ERROR nova.compute.manager [instance: 348ea580-2268-4082-a5f1-33c954063e4d] self.wait() [ 745.148338] env[61957]: ERROR nova.compute.manager [instance: 348ea580-2268-4082-a5f1-33c954063e4d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 745.148338] env[61957]: ERROR nova.compute.manager [instance: 348ea580-2268-4082-a5f1-33c954063e4d] self[:] = self._gt.wait() [ 745.148338] env[61957]: ERROR nova.compute.manager [instance: 348ea580-2268-4082-a5f1-33c954063e4d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 745.148338] env[61957]: ERROR nova.compute.manager [instance: 348ea580-2268-4082-a5f1-33c954063e4d] return self._exit_event.wait() [ 745.148338] env[61957]: ERROR nova.compute.manager [instance: 348ea580-2268-4082-a5f1-33c954063e4d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 745.148338] env[61957]: ERROR nova.compute.manager [instance: 348ea580-2268-4082-a5f1-33c954063e4d] current.throw(*self._exc) [ 745.148753] env[61957]: ERROR nova.compute.manager [instance: 348ea580-2268-4082-a5f1-33c954063e4d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 745.148753] env[61957]: ERROR nova.compute.manager [instance: 348ea580-2268-4082-a5f1-33c954063e4d] result = function(*args, **kwargs) [ 745.148753] env[61957]: ERROR nova.compute.manager [instance: 348ea580-2268-4082-a5f1-33c954063e4d] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 745.148753] env[61957]: ERROR nova.compute.manager [instance: 348ea580-2268-4082-a5f1-33c954063e4d] return func(*args, **kwargs) [ 745.148753] env[61957]: ERROR nova.compute.manager [instance: 348ea580-2268-4082-a5f1-33c954063e4d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 745.148753] env[61957]: ERROR nova.compute.manager [instance: 348ea580-2268-4082-a5f1-33c954063e4d] raise e [ 745.148753] env[61957]: ERROR nova.compute.manager [instance: 348ea580-2268-4082-a5f1-33c954063e4d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 745.148753] env[61957]: ERROR nova.compute.manager [instance: 348ea580-2268-4082-a5f1-33c954063e4d] nwinfo = self.network_api.allocate_for_instance( [ 745.148753] env[61957]: ERROR nova.compute.manager [instance: 348ea580-2268-4082-a5f1-33c954063e4d] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 745.148753] env[61957]: ERROR nova.compute.manager [instance: 348ea580-2268-4082-a5f1-33c954063e4d] created_port_ids = self._update_ports_for_instance( [ 745.148753] env[61957]: ERROR nova.compute.manager [instance: 348ea580-2268-4082-a5f1-33c954063e4d] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 745.148753] env[61957]: ERROR nova.compute.manager [instance: 348ea580-2268-4082-a5f1-33c954063e4d] with excutils.save_and_reraise_exception(): [ 745.148753] env[61957]: ERROR nova.compute.manager [instance: 348ea580-2268-4082-a5f1-33c954063e4d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 745.149191] env[61957]: ERROR nova.compute.manager [instance: 348ea580-2268-4082-a5f1-33c954063e4d] self.force_reraise() [ 745.149191] env[61957]: ERROR nova.compute.manager [instance: 348ea580-2268-4082-a5f1-33c954063e4d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 745.149191] env[61957]: ERROR nova.compute.manager [instance: 348ea580-2268-4082-a5f1-33c954063e4d] raise self.value [ 745.149191] env[61957]: ERROR nova.compute.manager [instance: 348ea580-2268-4082-a5f1-33c954063e4d] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 745.149191] env[61957]: ERROR nova.compute.manager [instance: 348ea580-2268-4082-a5f1-33c954063e4d] updated_port = self._update_port( [ 745.149191] env[61957]: ERROR nova.compute.manager [instance: 348ea580-2268-4082-a5f1-33c954063e4d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 745.149191] env[61957]: ERROR nova.compute.manager [instance: 348ea580-2268-4082-a5f1-33c954063e4d] _ensure_no_port_binding_failure(port) [ 745.149191] env[61957]: ERROR nova.compute.manager [instance: 348ea580-2268-4082-a5f1-33c954063e4d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 745.149191] env[61957]: ERROR nova.compute.manager [instance: 348ea580-2268-4082-a5f1-33c954063e4d] raise exception.PortBindingFailed(port_id=port['id']) [ 745.149191] env[61957]: ERROR nova.compute.manager [instance: 348ea580-2268-4082-a5f1-33c954063e4d] nova.exception.PortBindingFailed: Binding failed for port 89bc10a8-611c-42c9-ac0f-b4d98e2005d2, please check neutron logs for more information. [ 745.149191] env[61957]: ERROR nova.compute.manager [instance: 348ea580-2268-4082-a5f1-33c954063e4d] [ 745.149191] env[61957]: INFO nova.compute.manager [None req-5ca5fdda-dbef-4c48-b38f-9d2aa78e9841 tempest-FloatingIPsAssociationNegativeTestJSON-572957506 tempest-FloatingIPsAssociationNegativeTestJSON-572957506-project-member] [instance: 348ea580-2268-4082-a5f1-33c954063e4d] Terminating instance [ 745.150844] env[61957]: DEBUG oslo_concurrency.lockutils [None req-5ca5fdda-dbef-4c48-b38f-9d2aa78e9841 tempest-FloatingIPsAssociationNegativeTestJSON-572957506 tempest-FloatingIPsAssociationNegativeTestJSON-572957506-project-member] Acquiring lock "refresh_cache-348ea580-2268-4082-a5f1-33c954063e4d" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 745.372274] env[61957]: DEBUG nova.network.neutron [req-fc90a278-48de-43b1-9689-d094a9b2b723 req-b810bf8a-537c-43df-b7ec-b66f8303692f service nova] [instance: 348ea580-2268-4082-a5f1-33c954063e4d] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 745.375083] env[61957]: DEBUG oslo_concurrency.lockutils [None req-3d8587e3-7ad5-40a6-88af-e8ed44c49baa tempest-ServerActionsV293TestJSON-1269989829 tempest-ServerActionsV293TestJSON-1269989829-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 745.474188] env[61957]: DEBUG nova.network.neutron [req-fc90a278-48de-43b1-9689-d094a9b2b723 req-b810bf8a-537c-43df-b7ec-b66f8303692f service nova] [instance: 348ea580-2268-4082-a5f1-33c954063e4d] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 745.519238] env[61957]: DEBUG nova.network.neutron [None req-707bfa11-3857-45e5-a4be-581ffade0e04 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 745.594323] env[61957]: DEBUG nova.network.neutron [None req-707bfa11-3857-45e5-a4be-581ffade0e04 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 745.977191] env[61957]: DEBUG oslo_concurrency.lockutils [req-fc90a278-48de-43b1-9689-d094a9b2b723 req-b810bf8a-537c-43df-b7ec-b66f8303692f service nova] Releasing lock "refresh_cache-348ea580-2268-4082-a5f1-33c954063e4d" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 745.977680] env[61957]: DEBUG oslo_concurrency.lockutils [None req-5ca5fdda-dbef-4c48-b38f-9d2aa78e9841 tempest-FloatingIPsAssociationNegativeTestJSON-572957506 tempest-FloatingIPsAssociationNegativeTestJSON-572957506-project-member] Acquired lock "refresh_cache-348ea580-2268-4082-a5f1-33c954063e4d" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 745.977879] env[61957]: DEBUG nova.network.neutron [None req-5ca5fdda-dbef-4c48-b38f-9d2aa78e9841 tempest-FloatingIPsAssociationNegativeTestJSON-572957506 tempest-FloatingIPsAssociationNegativeTestJSON-572957506-project-member] [instance: 348ea580-2268-4082-a5f1-33c954063e4d] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 746.098938] env[61957]: DEBUG oslo_concurrency.lockutils [None req-707bfa11-3857-45e5-a4be-581ffade0e04 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Releasing lock "refresh_cache-2252ef6e-19ec-4356-8221-33168c01bef0" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 746.098938] env[61957]: DEBUG nova.compute.manager [None req-707bfa11-3857-45e5-a4be-581ffade0e04 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61957) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 746.098938] env[61957]: DEBUG nova.compute.manager [None req-707bfa11-3857-45e5-a4be-581ffade0e04 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 746.099204] env[61957]: DEBUG nova.network.neutron [None req-707bfa11-3857-45e5-a4be-581ffade0e04 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 746.115805] env[61957]: DEBUG nova.network.neutron [None req-707bfa11-3857-45e5-a4be-581ffade0e04 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 746.299338] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3617cba8-ec38-4954-9fbd-6f946efbf57e {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.306341] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abe6a932-5688-40fd-8870-3f0184b9d58c {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.337021] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5243a50-7437-4785-9142-db8ffc70243a {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.344061] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1cbf5f2d-4770-4049-a685-57f27b588ad6 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.356808] env[61957]: DEBUG nova.compute.provider_tree [None req-838e8ad7-4b6c-4dbc-8d4b-8380b652e559 tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 746.498943] env[61957]: DEBUG nova.network.neutron [None req-5ca5fdda-dbef-4c48-b38f-9d2aa78e9841 tempest-FloatingIPsAssociationNegativeTestJSON-572957506 tempest-FloatingIPsAssociationNegativeTestJSON-572957506-project-member] [instance: 348ea580-2268-4082-a5f1-33c954063e4d] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 746.593438] env[61957]: DEBUG nova.network.neutron [None req-5ca5fdda-dbef-4c48-b38f-9d2aa78e9841 tempest-FloatingIPsAssociationNegativeTestJSON-572957506 tempest-FloatingIPsAssociationNegativeTestJSON-572957506-project-member] [instance: 348ea580-2268-4082-a5f1-33c954063e4d] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 746.617524] env[61957]: DEBUG nova.network.neutron [None req-707bfa11-3857-45e5-a4be-581ffade0e04 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 746.859676] env[61957]: DEBUG nova.scheduler.client.report [None req-838e8ad7-4b6c-4dbc-8d4b-8380b652e559 tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 746.866883] env[61957]: DEBUG nova.compute.manager [req-8cd7ce8e-6324-4d6d-9cab-b60442c40c63 req-0e0509f4-4b53-4af6-a7d8-11f49a7ce042 service nova] [instance: 348ea580-2268-4082-a5f1-33c954063e4d] Received event network-vif-deleted-89bc10a8-611c-42c9-ac0f-b4d98e2005d2 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 747.096411] env[61957]: DEBUG oslo_concurrency.lockutils [None req-5ca5fdda-dbef-4c48-b38f-9d2aa78e9841 tempest-FloatingIPsAssociationNegativeTestJSON-572957506 tempest-FloatingIPsAssociationNegativeTestJSON-572957506-project-member] Releasing lock "refresh_cache-348ea580-2268-4082-a5f1-33c954063e4d" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 747.096911] env[61957]: DEBUG nova.compute.manager [None req-5ca5fdda-dbef-4c48-b38f-9d2aa78e9841 tempest-FloatingIPsAssociationNegativeTestJSON-572957506 tempest-FloatingIPsAssociationNegativeTestJSON-572957506-project-member] [instance: 348ea580-2268-4082-a5f1-33c954063e4d] Start destroying the instance on the hypervisor. {{(pid=61957) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 747.097829] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-5ca5fdda-dbef-4c48-b38f-9d2aa78e9841 tempest-FloatingIPsAssociationNegativeTestJSON-572957506 tempest-FloatingIPsAssociationNegativeTestJSON-572957506-project-member] [instance: 348ea580-2268-4082-a5f1-33c954063e4d] Destroying instance {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 747.097829] env[61957]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3d9a01a5-444e-420b-9e9c-ad817a71412e {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.107314] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bffd630a-5e3d-48d5-aee5-7dda4ef7fe49 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.122420] env[61957]: INFO nova.compute.manager [None req-707bfa11-3857-45e5-a4be-581ffade0e04 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: 2252ef6e-19ec-4356-8221-33168c01bef0] Took 1.02 seconds to deallocate network for instance. [ 747.136755] env[61957]: WARNING nova.virt.vmwareapi.vmops [None req-5ca5fdda-dbef-4c48-b38f-9d2aa78e9841 tempest-FloatingIPsAssociationNegativeTestJSON-572957506 tempest-FloatingIPsAssociationNegativeTestJSON-572957506-project-member] [instance: 348ea580-2268-4082-a5f1-33c954063e4d] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 348ea580-2268-4082-a5f1-33c954063e4d could not be found. [ 747.136755] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-5ca5fdda-dbef-4c48-b38f-9d2aa78e9841 tempest-FloatingIPsAssociationNegativeTestJSON-572957506 tempest-FloatingIPsAssociationNegativeTestJSON-572957506-project-member] [instance: 348ea580-2268-4082-a5f1-33c954063e4d] Instance destroyed {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 747.136755] env[61957]: INFO nova.compute.manager [None req-5ca5fdda-dbef-4c48-b38f-9d2aa78e9841 tempest-FloatingIPsAssociationNegativeTestJSON-572957506 tempest-FloatingIPsAssociationNegativeTestJSON-572957506-project-member] [instance: 348ea580-2268-4082-a5f1-33c954063e4d] Took 0.04 seconds to destroy the instance on the hypervisor. [ 747.136755] env[61957]: DEBUG oslo.service.loopingcall [None req-5ca5fdda-dbef-4c48-b38f-9d2aa78e9841 tempest-FloatingIPsAssociationNegativeTestJSON-572957506 tempest-FloatingIPsAssociationNegativeTestJSON-572957506-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 747.136755] env[61957]: DEBUG nova.compute.manager [-] [instance: 348ea580-2268-4082-a5f1-33c954063e4d] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 747.136956] env[61957]: DEBUG nova.network.neutron [-] [instance: 348ea580-2268-4082-a5f1-33c954063e4d] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 747.153186] env[61957]: DEBUG nova.network.neutron [-] [instance: 348ea580-2268-4082-a5f1-33c954063e4d] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 747.365095] env[61957]: DEBUG oslo_concurrency.lockutils [None req-838e8ad7-4b6c-4dbc-8d4b-8380b652e559 tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.369s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 747.365612] env[61957]: DEBUG nova.compute.manager [None req-838e8ad7-4b6c-4dbc-8d4b-8380b652e559 tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] [instance: 042425a0-5398-47a6-9842-380d5b3299c0] Start building networks asynchronously for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 747.368510] env[61957]: DEBUG oslo_concurrency.lockutils [None req-2b9492ff-1cb8-4ba9-ba5e-b513c4b108ad tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.056s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 747.655341] env[61957]: DEBUG nova.network.neutron [-] [instance: 348ea580-2268-4082-a5f1-33c954063e4d] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 747.872596] env[61957]: DEBUG nova.compute.utils [None req-838e8ad7-4b6c-4dbc-8d4b-8380b652e559 tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] Using /dev/sd instead of None {{(pid=61957) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 747.876968] env[61957]: DEBUG nova.compute.manager [None req-838e8ad7-4b6c-4dbc-8d4b-8380b652e559 tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] [instance: 042425a0-5398-47a6-9842-380d5b3299c0] Allocating IP information in the background. {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 747.877204] env[61957]: DEBUG nova.network.neutron [None req-838e8ad7-4b6c-4dbc-8d4b-8380b652e559 tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] [instance: 042425a0-5398-47a6-9842-380d5b3299c0] allocate_for_instance() {{(pid=61957) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 747.961156] env[61957]: DEBUG nova.policy [None req-838e8ad7-4b6c-4dbc-8d4b-8380b652e559 tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '167bf8ddcb234f65bc90bca1b3cd4bab', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'cd3e0ac57aa342ec98fc6a5fd06f2a30', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61957) authorize /opt/stack/nova/nova/policy.py:203}} [ 748.157642] env[61957]: INFO nova.compute.manager [-] [instance: 348ea580-2268-4082-a5f1-33c954063e4d] Took 1.02 seconds to deallocate network for instance. [ 748.160121] env[61957]: DEBUG nova.compute.claims [None req-5ca5fdda-dbef-4c48-b38f-9d2aa78e9841 tempest-FloatingIPsAssociationNegativeTestJSON-572957506 tempest-FloatingIPsAssociationNegativeTestJSON-572957506-project-member] [instance: 348ea580-2268-4082-a5f1-33c954063e4d] Aborting claim: {{(pid=61957) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 748.160121] env[61957]: DEBUG oslo_concurrency.lockutils [None req-5ca5fdda-dbef-4c48-b38f-9d2aa78e9841 tempest-FloatingIPsAssociationNegativeTestJSON-572957506 tempest-FloatingIPsAssociationNegativeTestJSON-572957506-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 748.161774] env[61957]: INFO nova.scheduler.client.report [None req-707bfa11-3857-45e5-a4be-581ffade0e04 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Deleted allocations for instance 2252ef6e-19ec-4356-8221-33168c01bef0 [ 748.186111] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ea8ce7d-0bc7-4209-8c58-a031cd5353e9 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.194506] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a282aaf-36b1-49f2-99d5-dbae3278073c {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.225889] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba40bf18-2b2f-49db-9b31-53dbcd1bab39 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.233396] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7668d5ea-0d22-46bb-99a3-60464ec187ae {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.247613] env[61957]: DEBUG nova.compute.provider_tree [None req-2b9492ff-1cb8-4ba9-ba5e-b513c4b108ad tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 748.378085] env[61957]: DEBUG nova.compute.manager [None req-838e8ad7-4b6c-4dbc-8d4b-8380b652e559 tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] [instance: 042425a0-5398-47a6-9842-380d5b3299c0] Start building block device mappings for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 748.438853] env[61957]: DEBUG nova.network.neutron [None req-838e8ad7-4b6c-4dbc-8d4b-8380b652e559 tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] [instance: 042425a0-5398-47a6-9842-380d5b3299c0] Successfully created port: 7c31a5be-6578-4eb8-833f-0722e63cb7e3 {{(pid=61957) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 748.672107] env[61957]: DEBUG oslo_concurrency.lockutils [None req-707bfa11-3857-45e5-a4be-581ffade0e04 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Lock "2252ef6e-19ec-4356-8221-33168c01bef0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 119.875s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 748.752428] env[61957]: DEBUG nova.scheduler.client.report [None req-2b9492ff-1cb8-4ba9-ba5e-b513c4b108ad tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 749.174759] env[61957]: DEBUG nova.compute.manager [None req-394366c7-3cf4-481b-bb9f-4da67892af40 tempest-ServersTestManualDisk-495750068 tempest-ServersTestManualDisk-495750068-project-member] [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] Starting instance... {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 749.257734] env[61957]: DEBUG oslo_concurrency.lockutils [None req-2b9492ff-1cb8-4ba9-ba5e-b513c4b108ad tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.889s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 749.258418] env[61957]: ERROR nova.compute.manager [None req-2b9492ff-1cb8-4ba9-ba5e-b513c4b108ad tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port d5344278-a5e5-4fe0-ab8d-80293d0e444b, please check neutron logs for more information. [ 749.258418] env[61957]: ERROR nova.compute.manager [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] Traceback (most recent call last): [ 749.258418] env[61957]: ERROR nova.compute.manager [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 749.258418] env[61957]: ERROR nova.compute.manager [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] self.driver.spawn(context, instance, image_meta, [ 749.258418] env[61957]: ERROR nova.compute.manager [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 749.258418] env[61957]: ERROR nova.compute.manager [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 749.258418] env[61957]: ERROR nova.compute.manager [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 749.258418] env[61957]: ERROR nova.compute.manager [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] vm_ref = self.build_virtual_machine(instance, [ 749.258418] env[61957]: ERROR nova.compute.manager [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 749.258418] env[61957]: ERROR nova.compute.manager [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] vif_infos = vmwarevif.get_vif_info(self._session, [ 749.258418] env[61957]: ERROR nova.compute.manager [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 749.258840] env[61957]: ERROR nova.compute.manager [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] for vif in network_info: [ 749.258840] env[61957]: ERROR nova.compute.manager [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 749.258840] env[61957]: ERROR nova.compute.manager [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] return self._sync_wrapper(fn, *args, **kwargs) [ 749.258840] env[61957]: ERROR nova.compute.manager [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 749.258840] env[61957]: ERROR nova.compute.manager [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] self.wait() [ 749.258840] env[61957]: ERROR nova.compute.manager [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 749.258840] env[61957]: ERROR nova.compute.manager [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] self[:] = self._gt.wait() [ 749.258840] env[61957]: ERROR nova.compute.manager [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 749.258840] env[61957]: ERROR nova.compute.manager [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] return self._exit_event.wait() [ 749.258840] env[61957]: ERROR nova.compute.manager [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 749.258840] env[61957]: ERROR nova.compute.manager [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] current.throw(*self._exc) [ 749.258840] env[61957]: ERROR nova.compute.manager [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 749.258840] env[61957]: ERROR nova.compute.manager [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] result = function(*args, **kwargs) [ 749.259254] env[61957]: ERROR nova.compute.manager [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 749.259254] env[61957]: ERROR nova.compute.manager [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] return func(*args, **kwargs) [ 749.259254] env[61957]: ERROR nova.compute.manager [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 749.259254] env[61957]: ERROR nova.compute.manager [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] raise e [ 749.259254] env[61957]: ERROR nova.compute.manager [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 749.259254] env[61957]: ERROR nova.compute.manager [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] nwinfo = self.network_api.allocate_for_instance( [ 749.259254] env[61957]: ERROR nova.compute.manager [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 749.259254] env[61957]: ERROR nova.compute.manager [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] created_port_ids = self._update_ports_for_instance( [ 749.259254] env[61957]: ERROR nova.compute.manager [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 749.259254] env[61957]: ERROR nova.compute.manager [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] with excutils.save_and_reraise_exception(): [ 749.259254] env[61957]: ERROR nova.compute.manager [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 749.259254] env[61957]: ERROR nova.compute.manager [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] self.force_reraise() [ 749.259254] env[61957]: ERROR nova.compute.manager [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 749.259676] env[61957]: ERROR nova.compute.manager [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] raise self.value [ 749.259676] env[61957]: ERROR nova.compute.manager [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 749.259676] env[61957]: ERROR nova.compute.manager [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] updated_port = self._update_port( [ 749.259676] env[61957]: ERROR nova.compute.manager [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 749.259676] env[61957]: ERROR nova.compute.manager [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] _ensure_no_port_binding_failure(port) [ 749.259676] env[61957]: ERROR nova.compute.manager [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 749.259676] env[61957]: ERROR nova.compute.manager [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] raise exception.PortBindingFailed(port_id=port['id']) [ 749.259676] env[61957]: ERROR nova.compute.manager [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] nova.exception.PortBindingFailed: Binding failed for port d5344278-a5e5-4fe0-ab8d-80293d0e444b, please check neutron logs for more information. [ 749.259676] env[61957]: ERROR nova.compute.manager [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] [ 749.259676] env[61957]: DEBUG nova.compute.utils [None req-2b9492ff-1cb8-4ba9-ba5e-b513c4b108ad tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] Binding failed for port d5344278-a5e5-4fe0-ab8d-80293d0e444b, please check neutron logs for more information. {{(pid=61957) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 749.260639] env[61957]: DEBUG oslo_concurrency.lockutils [None req-47718518-520f-4177-b41f-8f0cae78b6f6 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.110s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 749.262568] env[61957]: INFO nova.compute.claims [None req-47718518-520f-4177-b41f-8f0cae78b6f6 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] [instance: 7f3850c0-a455-4e5b-a329-fb3b02f9725f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 749.265818] env[61957]: DEBUG nova.compute.manager [None req-2b9492ff-1cb8-4ba9-ba5e-b513c4b108ad tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] Build of instance f8235178-7ae0-4733-bb05-3515d5489a5b was re-scheduled: Binding failed for port d5344278-a5e5-4fe0-ab8d-80293d0e444b, please check neutron logs for more information. {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 749.266447] env[61957]: DEBUG nova.compute.manager [None req-2b9492ff-1cb8-4ba9-ba5e-b513c4b108ad tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] Unplugging VIFs for instance {{(pid=61957) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 749.266629] env[61957]: DEBUG oslo_concurrency.lockutils [None req-2b9492ff-1cb8-4ba9-ba5e-b513c4b108ad tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] Acquiring lock "refresh_cache-f8235178-7ae0-4733-bb05-3515d5489a5b" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 749.266833] env[61957]: DEBUG oslo_concurrency.lockutils [None req-2b9492ff-1cb8-4ba9-ba5e-b513c4b108ad tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] Acquired lock "refresh_cache-f8235178-7ae0-4733-bb05-3515d5489a5b" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 749.267089] env[61957]: DEBUG nova.network.neutron [None req-2b9492ff-1cb8-4ba9-ba5e-b513c4b108ad tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 749.393781] env[61957]: DEBUG nova.compute.manager [None req-838e8ad7-4b6c-4dbc-8d4b-8380b652e559 tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] [instance: 042425a0-5398-47a6-9842-380d5b3299c0] Start spawning the instance on the hypervisor. {{(pid=61957) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 749.437861] env[61957]: DEBUG nova.virt.hardware [None req-838e8ad7-4b6c-4dbc-8d4b-8380b652e559 tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T17:24:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='47c6d361-c775-4a1e-8fdd-9e07aadebabb',id=37,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-706201999',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T17:21:50Z,direct_url=,disk_format='vmdk',id=11c76a2c-f705-470a-ba9d-4657858bab38,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='81d3096d0f094373913fc3dc8f5c3c6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T17:21:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 749.438160] env[61957]: DEBUG nova.virt.hardware [None req-838e8ad7-4b6c-4dbc-8d4b-8380b652e559 tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] Flavor limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 749.438321] env[61957]: DEBUG nova.virt.hardware [None req-838e8ad7-4b6c-4dbc-8d4b-8380b652e559 tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] Image limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 749.438512] env[61957]: DEBUG nova.virt.hardware [None req-838e8ad7-4b6c-4dbc-8d4b-8380b652e559 tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] Flavor pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 749.438660] env[61957]: DEBUG nova.virt.hardware [None req-838e8ad7-4b6c-4dbc-8d4b-8380b652e559 tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] Image pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 749.438833] env[61957]: DEBUG nova.virt.hardware [None req-838e8ad7-4b6c-4dbc-8d4b-8380b652e559 tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 749.439053] env[61957]: DEBUG nova.virt.hardware [None req-838e8ad7-4b6c-4dbc-8d4b-8380b652e559 tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 749.439215] env[61957]: DEBUG nova.virt.hardware [None req-838e8ad7-4b6c-4dbc-8d4b-8380b652e559 tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 749.439379] env[61957]: DEBUG nova.virt.hardware [None req-838e8ad7-4b6c-4dbc-8d4b-8380b652e559 tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] Got 1 possible topologies {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 749.439538] env[61957]: DEBUG nova.virt.hardware [None req-838e8ad7-4b6c-4dbc-8d4b-8380b652e559 tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 749.439728] env[61957]: DEBUG nova.virt.hardware [None req-838e8ad7-4b6c-4dbc-8d4b-8380b652e559 tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 749.440984] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f471a08b-241a-469f-b69a-d0091c59fe86 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.450362] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f077faa1-65f5-4324-acd7-137c1ea09a99 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.699677] env[61957]: DEBUG oslo_concurrency.lockutils [None req-394366c7-3cf4-481b-bb9f-4da67892af40 tempest-ServersTestManualDisk-495750068 tempest-ServersTestManualDisk-495750068-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 749.766105] env[61957]: DEBUG nova.compute.manager [req-46841913-8067-4efa-bbe0-db0e8c8fca36 req-3b2630af-481f-4e4e-8520-c740e3a930f3 service nova] [instance: 042425a0-5398-47a6-9842-380d5b3299c0] Received event network-changed-7c31a5be-6578-4eb8-833f-0722e63cb7e3 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 749.766317] env[61957]: DEBUG nova.compute.manager [req-46841913-8067-4efa-bbe0-db0e8c8fca36 req-3b2630af-481f-4e4e-8520-c740e3a930f3 service nova] [instance: 042425a0-5398-47a6-9842-380d5b3299c0] Refreshing instance network info cache due to event network-changed-7c31a5be-6578-4eb8-833f-0722e63cb7e3. {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 749.766526] env[61957]: DEBUG oslo_concurrency.lockutils [req-46841913-8067-4efa-bbe0-db0e8c8fca36 req-3b2630af-481f-4e4e-8520-c740e3a930f3 service nova] Acquiring lock "refresh_cache-042425a0-5398-47a6-9842-380d5b3299c0" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 749.766706] env[61957]: DEBUG oslo_concurrency.lockutils [req-46841913-8067-4efa-bbe0-db0e8c8fca36 req-3b2630af-481f-4e4e-8520-c740e3a930f3 service nova] Acquired lock "refresh_cache-042425a0-5398-47a6-9842-380d5b3299c0" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 749.766878] env[61957]: DEBUG nova.network.neutron [req-46841913-8067-4efa-bbe0-db0e8c8fca36 req-3b2630af-481f-4e4e-8520-c740e3a930f3 service nova] [instance: 042425a0-5398-47a6-9842-380d5b3299c0] Refreshing network info cache for port 7c31a5be-6578-4eb8-833f-0722e63cb7e3 {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 749.796863] env[61957]: DEBUG nova.network.neutron [None req-2b9492ff-1cb8-4ba9-ba5e-b513c4b108ad tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 749.931129] env[61957]: DEBUG nova.network.neutron [None req-2b9492ff-1cb8-4ba9-ba5e-b513c4b108ad tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 750.015983] env[61957]: DEBUG oslo_concurrency.lockutils [None req-e4109c08-00a8-4666-8e84-183191cbc2fb tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Acquiring lock "0adae8e1-8c2f-4110-805b-1f286debc833" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 750.016314] env[61957]: DEBUG oslo_concurrency.lockutils [None req-e4109c08-00a8-4666-8e84-183191cbc2fb tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Lock "0adae8e1-8c2f-4110-805b-1f286debc833" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 750.044836] env[61957]: ERROR nova.compute.manager [None req-838e8ad7-4b6c-4dbc-8d4b-8380b652e559 tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 7c31a5be-6578-4eb8-833f-0722e63cb7e3, please check neutron logs for more information. [ 750.044836] env[61957]: ERROR nova.compute.manager Traceback (most recent call last): [ 750.044836] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 750.044836] env[61957]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 750.044836] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 750.044836] env[61957]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 750.044836] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 750.044836] env[61957]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 750.044836] env[61957]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 750.044836] env[61957]: ERROR nova.compute.manager self.force_reraise() [ 750.044836] env[61957]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 750.044836] env[61957]: ERROR nova.compute.manager raise self.value [ 750.044836] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 750.044836] env[61957]: ERROR nova.compute.manager updated_port = self._update_port( [ 750.044836] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 750.044836] env[61957]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 750.045316] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 750.045316] env[61957]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 750.045316] env[61957]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 7c31a5be-6578-4eb8-833f-0722e63cb7e3, please check neutron logs for more information. [ 750.045316] env[61957]: ERROR nova.compute.manager [ 750.045316] env[61957]: Traceback (most recent call last): [ 750.045316] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 750.045316] env[61957]: listener.cb(fileno) [ 750.045316] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 750.045316] env[61957]: result = function(*args, **kwargs) [ 750.045316] env[61957]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 750.045316] env[61957]: return func(*args, **kwargs) [ 750.045316] env[61957]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 750.045316] env[61957]: raise e [ 750.045316] env[61957]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 750.045316] env[61957]: nwinfo = self.network_api.allocate_for_instance( [ 750.045316] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 750.045316] env[61957]: created_port_ids = self._update_ports_for_instance( [ 750.045316] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 750.045316] env[61957]: with excutils.save_and_reraise_exception(): [ 750.045316] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 750.045316] env[61957]: self.force_reraise() [ 750.045316] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 750.045316] env[61957]: raise self.value [ 750.045316] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 750.045316] env[61957]: updated_port = self._update_port( [ 750.045316] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 750.045316] env[61957]: _ensure_no_port_binding_failure(port) [ 750.045316] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 750.045316] env[61957]: raise exception.PortBindingFailed(port_id=port['id']) [ 750.046464] env[61957]: nova.exception.PortBindingFailed: Binding failed for port 7c31a5be-6578-4eb8-833f-0722e63cb7e3, please check neutron logs for more information. [ 750.046464] env[61957]: Removing descriptor: 16 [ 750.046464] env[61957]: ERROR nova.compute.manager [None req-838e8ad7-4b6c-4dbc-8d4b-8380b652e559 tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] [instance: 042425a0-5398-47a6-9842-380d5b3299c0] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 7c31a5be-6578-4eb8-833f-0722e63cb7e3, please check neutron logs for more information. [ 750.046464] env[61957]: ERROR nova.compute.manager [instance: 042425a0-5398-47a6-9842-380d5b3299c0] Traceback (most recent call last): [ 750.046464] env[61957]: ERROR nova.compute.manager [instance: 042425a0-5398-47a6-9842-380d5b3299c0] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 750.046464] env[61957]: ERROR nova.compute.manager [instance: 042425a0-5398-47a6-9842-380d5b3299c0] yield resources [ 750.046464] env[61957]: ERROR nova.compute.manager [instance: 042425a0-5398-47a6-9842-380d5b3299c0] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 750.046464] env[61957]: ERROR nova.compute.manager [instance: 042425a0-5398-47a6-9842-380d5b3299c0] self.driver.spawn(context, instance, image_meta, [ 750.046464] env[61957]: ERROR nova.compute.manager [instance: 042425a0-5398-47a6-9842-380d5b3299c0] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 750.046464] env[61957]: ERROR nova.compute.manager [instance: 042425a0-5398-47a6-9842-380d5b3299c0] self._vmops.spawn(context, instance, image_meta, injected_files, [ 750.046464] env[61957]: ERROR nova.compute.manager [instance: 042425a0-5398-47a6-9842-380d5b3299c0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 750.046464] env[61957]: ERROR nova.compute.manager [instance: 042425a0-5398-47a6-9842-380d5b3299c0] vm_ref = self.build_virtual_machine(instance, [ 750.046802] env[61957]: ERROR nova.compute.manager [instance: 042425a0-5398-47a6-9842-380d5b3299c0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 750.046802] env[61957]: ERROR nova.compute.manager [instance: 042425a0-5398-47a6-9842-380d5b3299c0] vif_infos = vmwarevif.get_vif_info(self._session, [ 750.046802] env[61957]: ERROR nova.compute.manager [instance: 042425a0-5398-47a6-9842-380d5b3299c0] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 750.046802] env[61957]: ERROR nova.compute.manager [instance: 042425a0-5398-47a6-9842-380d5b3299c0] for vif in network_info: [ 750.046802] env[61957]: ERROR nova.compute.manager [instance: 042425a0-5398-47a6-9842-380d5b3299c0] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 750.046802] env[61957]: ERROR nova.compute.manager [instance: 042425a0-5398-47a6-9842-380d5b3299c0] return self._sync_wrapper(fn, *args, **kwargs) [ 750.046802] env[61957]: ERROR nova.compute.manager [instance: 042425a0-5398-47a6-9842-380d5b3299c0] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 750.046802] env[61957]: ERROR nova.compute.manager [instance: 042425a0-5398-47a6-9842-380d5b3299c0] self.wait() [ 750.046802] env[61957]: ERROR nova.compute.manager [instance: 042425a0-5398-47a6-9842-380d5b3299c0] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 750.046802] env[61957]: ERROR nova.compute.manager [instance: 042425a0-5398-47a6-9842-380d5b3299c0] self[:] = self._gt.wait() [ 750.046802] env[61957]: ERROR nova.compute.manager [instance: 042425a0-5398-47a6-9842-380d5b3299c0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 750.046802] env[61957]: ERROR nova.compute.manager [instance: 042425a0-5398-47a6-9842-380d5b3299c0] return self._exit_event.wait() [ 750.046802] env[61957]: ERROR nova.compute.manager [instance: 042425a0-5398-47a6-9842-380d5b3299c0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 750.047161] env[61957]: ERROR nova.compute.manager [instance: 042425a0-5398-47a6-9842-380d5b3299c0] result = hub.switch() [ 750.047161] env[61957]: ERROR nova.compute.manager [instance: 042425a0-5398-47a6-9842-380d5b3299c0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 750.047161] env[61957]: ERROR nova.compute.manager [instance: 042425a0-5398-47a6-9842-380d5b3299c0] return self.greenlet.switch() [ 750.047161] env[61957]: ERROR nova.compute.manager [instance: 042425a0-5398-47a6-9842-380d5b3299c0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 750.047161] env[61957]: ERROR nova.compute.manager [instance: 042425a0-5398-47a6-9842-380d5b3299c0] result = function(*args, **kwargs) [ 750.047161] env[61957]: ERROR nova.compute.manager [instance: 042425a0-5398-47a6-9842-380d5b3299c0] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 750.047161] env[61957]: ERROR nova.compute.manager [instance: 042425a0-5398-47a6-9842-380d5b3299c0] return func(*args, **kwargs) [ 750.047161] env[61957]: ERROR nova.compute.manager [instance: 042425a0-5398-47a6-9842-380d5b3299c0] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 750.047161] env[61957]: ERROR nova.compute.manager [instance: 042425a0-5398-47a6-9842-380d5b3299c0] raise e [ 750.047161] env[61957]: ERROR nova.compute.manager [instance: 042425a0-5398-47a6-9842-380d5b3299c0] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 750.047161] env[61957]: ERROR nova.compute.manager [instance: 042425a0-5398-47a6-9842-380d5b3299c0] nwinfo = self.network_api.allocate_for_instance( [ 750.047161] env[61957]: ERROR nova.compute.manager [instance: 042425a0-5398-47a6-9842-380d5b3299c0] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 750.047161] env[61957]: ERROR nova.compute.manager [instance: 042425a0-5398-47a6-9842-380d5b3299c0] created_port_ids = self._update_ports_for_instance( [ 750.047515] env[61957]: ERROR nova.compute.manager [instance: 042425a0-5398-47a6-9842-380d5b3299c0] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 750.047515] env[61957]: ERROR nova.compute.manager [instance: 042425a0-5398-47a6-9842-380d5b3299c0] with excutils.save_and_reraise_exception(): [ 750.047515] env[61957]: ERROR nova.compute.manager [instance: 042425a0-5398-47a6-9842-380d5b3299c0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 750.047515] env[61957]: ERROR nova.compute.manager [instance: 042425a0-5398-47a6-9842-380d5b3299c0] self.force_reraise() [ 750.047515] env[61957]: ERROR nova.compute.manager [instance: 042425a0-5398-47a6-9842-380d5b3299c0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 750.047515] env[61957]: ERROR nova.compute.manager [instance: 042425a0-5398-47a6-9842-380d5b3299c0] raise self.value [ 750.047515] env[61957]: ERROR nova.compute.manager [instance: 042425a0-5398-47a6-9842-380d5b3299c0] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 750.047515] env[61957]: ERROR nova.compute.manager [instance: 042425a0-5398-47a6-9842-380d5b3299c0] updated_port = self._update_port( [ 750.047515] env[61957]: ERROR nova.compute.manager [instance: 042425a0-5398-47a6-9842-380d5b3299c0] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 750.047515] env[61957]: ERROR nova.compute.manager [instance: 042425a0-5398-47a6-9842-380d5b3299c0] _ensure_no_port_binding_failure(port) [ 750.047515] env[61957]: ERROR nova.compute.manager [instance: 042425a0-5398-47a6-9842-380d5b3299c0] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 750.047515] env[61957]: ERROR nova.compute.manager [instance: 042425a0-5398-47a6-9842-380d5b3299c0] raise exception.PortBindingFailed(port_id=port['id']) [ 750.047839] env[61957]: ERROR nova.compute.manager [instance: 042425a0-5398-47a6-9842-380d5b3299c0] nova.exception.PortBindingFailed: Binding failed for port 7c31a5be-6578-4eb8-833f-0722e63cb7e3, please check neutron logs for more information. [ 750.047839] env[61957]: ERROR nova.compute.manager [instance: 042425a0-5398-47a6-9842-380d5b3299c0] [ 750.047839] env[61957]: INFO nova.compute.manager [None req-838e8ad7-4b6c-4dbc-8d4b-8380b652e559 tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] [instance: 042425a0-5398-47a6-9842-380d5b3299c0] Terminating instance [ 750.048724] env[61957]: DEBUG oslo_concurrency.lockutils [None req-838e8ad7-4b6c-4dbc-8d4b-8380b652e559 tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] Acquiring lock "refresh_cache-042425a0-5398-47a6-9842-380d5b3299c0" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 750.298131] env[61957]: DEBUG nova.network.neutron [req-46841913-8067-4efa-bbe0-db0e8c8fca36 req-3b2630af-481f-4e4e-8520-c740e3a930f3 service nova] [instance: 042425a0-5398-47a6-9842-380d5b3299c0] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 750.381412] env[61957]: DEBUG nova.network.neutron [req-46841913-8067-4efa-bbe0-db0e8c8fca36 req-3b2630af-481f-4e4e-8520-c740e3a930f3 service nova] [instance: 042425a0-5398-47a6-9842-380d5b3299c0] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 750.433317] env[61957]: DEBUG oslo_concurrency.lockutils [None req-2b9492ff-1cb8-4ba9-ba5e-b513c4b108ad tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] Releasing lock "refresh_cache-f8235178-7ae0-4733-bb05-3515d5489a5b" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 750.433586] env[61957]: DEBUG nova.compute.manager [None req-2b9492ff-1cb8-4ba9-ba5e-b513c4b108ad tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61957) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 750.433771] env[61957]: DEBUG nova.compute.manager [None req-2b9492ff-1cb8-4ba9-ba5e-b513c4b108ad tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 750.433944] env[61957]: DEBUG nova.network.neutron [None req-2b9492ff-1cb8-4ba9-ba5e-b513c4b108ad tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 750.450023] env[61957]: DEBUG nova.network.neutron [None req-2b9492ff-1cb8-4ba9-ba5e-b513c4b108ad tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 750.621295] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a45e4395-6dc0-4f7e-a74c-3e6134d77a5a {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.629201] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-413ebd62-d308-40a0-ac10-610dfd23b423 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.661868] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ec6512d-41f9-4b00-af39-d894e5558a89 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.669705] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4707f97-7723-4b5d-b06a-018e8dbc515b {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.683424] env[61957]: DEBUG nova.compute.provider_tree [None req-47718518-520f-4177-b41f-8f0cae78b6f6 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 750.880686] env[61957]: DEBUG oslo_concurrency.lockutils [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Acquiring lock "6438fe40-046c-45d5-9986-8f182ecde49f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 750.880977] env[61957]: DEBUG oslo_concurrency.lockutils [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Lock "6438fe40-046c-45d5-9986-8f182ecde49f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 750.884870] env[61957]: DEBUG oslo_concurrency.lockutils [req-46841913-8067-4efa-bbe0-db0e8c8fca36 req-3b2630af-481f-4e4e-8520-c740e3a930f3 service nova] Releasing lock "refresh_cache-042425a0-5398-47a6-9842-380d5b3299c0" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 750.885215] env[61957]: DEBUG oslo_concurrency.lockutils [None req-838e8ad7-4b6c-4dbc-8d4b-8380b652e559 tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] Acquired lock "refresh_cache-042425a0-5398-47a6-9842-380d5b3299c0" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 750.885393] env[61957]: DEBUG nova.network.neutron [None req-838e8ad7-4b6c-4dbc-8d4b-8380b652e559 tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] [instance: 042425a0-5398-47a6-9842-380d5b3299c0] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 750.908836] env[61957]: DEBUG oslo_concurrency.lockutils [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Acquiring lock "f66db265-887e-4d61-b848-c609e5c884cb" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 750.909100] env[61957]: DEBUG oslo_concurrency.lockutils [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Lock "f66db265-887e-4d61-b848-c609e5c884cb" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 750.957825] env[61957]: DEBUG nova.network.neutron [None req-2b9492ff-1cb8-4ba9-ba5e-b513c4b108ad tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 751.187643] env[61957]: DEBUG nova.scheduler.client.report [None req-47718518-520f-4177-b41f-8f0cae78b6f6 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 751.405203] env[61957]: DEBUG nova.network.neutron [None req-838e8ad7-4b6c-4dbc-8d4b-8380b652e559 tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] [instance: 042425a0-5398-47a6-9842-380d5b3299c0] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 751.460008] env[61957]: INFO nova.compute.manager [None req-2b9492ff-1cb8-4ba9-ba5e-b513c4b108ad tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] [instance: f8235178-7ae0-4733-bb05-3515d5489a5b] Took 1.03 seconds to deallocate network for instance. [ 751.475748] env[61957]: DEBUG nova.network.neutron [None req-838e8ad7-4b6c-4dbc-8d4b-8380b652e559 tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] [instance: 042425a0-5398-47a6-9842-380d5b3299c0] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 751.693281] env[61957]: DEBUG oslo_concurrency.lockutils [None req-47718518-520f-4177-b41f-8f0cae78b6f6 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.432s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 751.693809] env[61957]: DEBUG nova.compute.manager [None req-47718518-520f-4177-b41f-8f0cae78b6f6 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] [instance: 7f3850c0-a455-4e5b-a329-fb3b02f9725f] Start building networks asynchronously for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 751.696344] env[61957]: DEBUG oslo_concurrency.lockutils [None req-f462e606-4c03-40e0-ac91-60efbd401559 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.426s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 751.795404] env[61957]: DEBUG nova.compute.manager [req-ee88cf0a-ec27-4d7e-a0cd-ba81cb09a88c req-1896da2e-bfab-4a09-ac4b-03626ddf42a9 service nova] [instance: 042425a0-5398-47a6-9842-380d5b3299c0] Received event network-vif-deleted-7c31a5be-6578-4eb8-833f-0722e63cb7e3 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 751.978318] env[61957]: DEBUG oslo_concurrency.lockutils [None req-838e8ad7-4b6c-4dbc-8d4b-8380b652e559 tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] Releasing lock "refresh_cache-042425a0-5398-47a6-9842-380d5b3299c0" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 751.978786] env[61957]: DEBUG nova.compute.manager [None req-838e8ad7-4b6c-4dbc-8d4b-8380b652e559 tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] [instance: 042425a0-5398-47a6-9842-380d5b3299c0] Start destroying the instance on the hypervisor. {{(pid=61957) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 751.978997] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-838e8ad7-4b6c-4dbc-8d4b-8380b652e559 tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] [instance: 042425a0-5398-47a6-9842-380d5b3299c0] Destroying instance {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 751.979301] env[61957]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c5580bac-db0a-4b24-9b7f-1031b2eda602 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.988384] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ae7d521-b890-4afb-85ec-d0bd1692d2d2 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.010254] env[61957]: WARNING nova.virt.vmwareapi.vmops [None req-838e8ad7-4b6c-4dbc-8d4b-8380b652e559 tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] [instance: 042425a0-5398-47a6-9842-380d5b3299c0] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 042425a0-5398-47a6-9842-380d5b3299c0 could not be found. [ 752.010456] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-838e8ad7-4b6c-4dbc-8d4b-8380b652e559 tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] [instance: 042425a0-5398-47a6-9842-380d5b3299c0] Instance destroyed {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 752.010624] env[61957]: INFO nova.compute.manager [None req-838e8ad7-4b6c-4dbc-8d4b-8380b652e559 tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] [instance: 042425a0-5398-47a6-9842-380d5b3299c0] Took 0.03 seconds to destroy the instance on the hypervisor. [ 752.010855] env[61957]: DEBUG oslo.service.loopingcall [None req-838e8ad7-4b6c-4dbc-8d4b-8380b652e559 tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 752.011069] env[61957]: DEBUG nova.compute.manager [-] [instance: 042425a0-5398-47a6-9842-380d5b3299c0] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 752.011167] env[61957]: DEBUG nova.network.neutron [-] [instance: 042425a0-5398-47a6-9842-380d5b3299c0] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 752.026668] env[61957]: DEBUG nova.network.neutron [-] [instance: 042425a0-5398-47a6-9842-380d5b3299c0] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 752.201272] env[61957]: DEBUG nova.compute.utils [None req-47718518-520f-4177-b41f-8f0cae78b6f6 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Using /dev/sd instead of None {{(pid=61957) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 752.205627] env[61957]: DEBUG nova.compute.manager [None req-47718518-520f-4177-b41f-8f0cae78b6f6 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] [instance: 7f3850c0-a455-4e5b-a329-fb3b02f9725f] Not allocating networking since 'none' was specified. {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 752.313679] env[61957]: DEBUG oslo_service.periodic_task [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61957) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 752.313899] env[61957]: DEBUG oslo_service.periodic_task [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61957) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 752.314054] env[61957]: DEBUG nova.compute.manager [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Starting heal instance info cache {{(pid=61957) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 752.314178] env[61957]: DEBUG nova.compute.manager [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Rebuilding the list of instances to heal {{(pid=61957) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9930}} [ 752.487504] env[61957]: INFO nova.scheduler.client.report [None req-2b9492ff-1cb8-4ba9-ba5e-b513c4b108ad tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] Deleted allocations for instance f8235178-7ae0-4733-bb05-3515d5489a5b [ 752.496222] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a12b3222-4eb1-491c-b324-42945c1316de {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.505163] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ea4b566-2e3e-4875-a293-0b0d7ef24534 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.536058] env[61957]: DEBUG nova.network.neutron [-] [instance: 042425a0-5398-47a6-9842-380d5b3299c0] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 752.537904] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2d25baa-e310-4838-8efd-a1fdb397a827 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.545748] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b89a5b8-29dd-44cd-852f-043f9b0ca215 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.559813] env[61957]: DEBUG nova.compute.provider_tree [None req-f462e606-4c03-40e0-ac91-60efbd401559 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 752.706602] env[61957]: DEBUG nova.compute.manager [None req-47718518-520f-4177-b41f-8f0cae78b6f6 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] [instance: 7f3850c0-a455-4e5b-a329-fb3b02f9725f] Start building block device mappings for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 752.819322] env[61957]: DEBUG nova.compute.manager [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] [instance: 348ea580-2268-4082-a5f1-33c954063e4d] Skipping network cache update for instance because it is Building. {{(pid=61957) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 752.819492] env[61957]: DEBUG nova.compute.manager [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] [instance: 042425a0-5398-47a6-9842-380d5b3299c0] Skipping network cache update for instance because it is Building. {{(pid=61957) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 752.819622] env[61957]: DEBUG nova.compute.manager [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] [instance: 7f3850c0-a455-4e5b-a329-fb3b02f9725f] Skipping network cache update for instance because it is Building. {{(pid=61957) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 752.819749] env[61957]: DEBUG nova.compute.manager [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Didn't find any instances for network info cache update. {{(pid=61957) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10012}} [ 752.819979] env[61957]: DEBUG oslo_service.periodic_task [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61957) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 752.820157] env[61957]: DEBUG oslo_service.periodic_task [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61957) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 752.820304] env[61957]: DEBUG oslo_service.periodic_task [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61957) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 752.820450] env[61957]: DEBUG oslo_service.periodic_task [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61957) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 752.820691] env[61957]: DEBUG oslo_service.periodic_task [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61957) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 752.820757] env[61957]: DEBUG oslo_service.periodic_task [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61957) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 752.820857] env[61957]: DEBUG nova.compute.manager [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61957) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 752.820984] env[61957]: DEBUG oslo_service.periodic_task [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61957) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 753.001041] env[61957]: DEBUG oslo_concurrency.lockutils [None req-2b9492ff-1cb8-4ba9-ba5e-b513c4b108ad tempest-ServerRescueNegativeTestJSON-1077609060 tempest-ServerRescueNegativeTestJSON-1077609060-project-member] Lock "f8235178-7ae0-4733-bb05-3515d5489a5b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 123.513s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 753.041336] env[61957]: INFO nova.compute.manager [-] [instance: 042425a0-5398-47a6-9842-380d5b3299c0] Took 1.03 seconds to deallocate network for instance. [ 753.043530] env[61957]: DEBUG nova.compute.claims [None req-838e8ad7-4b6c-4dbc-8d4b-8380b652e559 tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] [instance: 042425a0-5398-47a6-9842-380d5b3299c0] Aborting claim: {{(pid=61957) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 753.043702] env[61957]: DEBUG oslo_concurrency.lockutils [None req-838e8ad7-4b6c-4dbc-8d4b-8380b652e559 tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 753.063656] env[61957]: DEBUG nova.scheduler.client.report [None req-f462e606-4c03-40e0-ac91-60efbd401559 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 753.323750] env[61957]: DEBUG oslo_concurrency.lockutils [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 753.503205] env[61957]: DEBUG nova.compute.manager [None req-d6c9937c-2526-4768-bade-4ca2ced8b569 tempest-ServersTestFqdnHostnames-1308211839 tempest-ServersTestFqdnHostnames-1308211839-project-member] [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] Starting instance... {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 753.567672] env[61957]: DEBUG oslo_concurrency.lockutils [None req-f462e606-4c03-40e0-ac91-60efbd401559 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.871s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 753.571019] env[61957]: ERROR nova.compute.manager [None req-f462e606-4c03-40e0-ac91-60efbd401559 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port e0a84501-e615-4647-8dff-5c863b104dc5, please check neutron logs for more information. [ 753.571019] env[61957]: ERROR nova.compute.manager [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] Traceback (most recent call last): [ 753.571019] env[61957]: ERROR nova.compute.manager [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 753.571019] env[61957]: ERROR nova.compute.manager [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] self.driver.spawn(context, instance, image_meta, [ 753.571019] env[61957]: ERROR nova.compute.manager [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 753.571019] env[61957]: ERROR nova.compute.manager [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 753.571019] env[61957]: ERROR nova.compute.manager [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 753.571019] env[61957]: ERROR nova.compute.manager [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] vm_ref = self.build_virtual_machine(instance, [ 753.571019] env[61957]: ERROR nova.compute.manager [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 753.571019] env[61957]: ERROR nova.compute.manager [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] vif_infos = vmwarevif.get_vif_info(self._session, [ 753.571019] env[61957]: ERROR nova.compute.manager [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 753.571428] env[61957]: ERROR nova.compute.manager [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] for vif in network_info: [ 753.571428] env[61957]: ERROR nova.compute.manager [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 753.571428] env[61957]: ERROR nova.compute.manager [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] return self._sync_wrapper(fn, *args, **kwargs) [ 753.571428] env[61957]: ERROR nova.compute.manager [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 753.571428] env[61957]: ERROR nova.compute.manager [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] self.wait() [ 753.571428] env[61957]: ERROR nova.compute.manager [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 753.571428] env[61957]: ERROR nova.compute.manager [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] self[:] = self._gt.wait() [ 753.571428] env[61957]: ERROR nova.compute.manager [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 753.571428] env[61957]: ERROR nova.compute.manager [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] return self._exit_event.wait() [ 753.571428] env[61957]: ERROR nova.compute.manager [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 753.571428] env[61957]: ERROR nova.compute.manager [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] current.throw(*self._exc) [ 753.571428] env[61957]: ERROR nova.compute.manager [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 753.571428] env[61957]: ERROR nova.compute.manager [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] result = function(*args, **kwargs) [ 753.571762] env[61957]: ERROR nova.compute.manager [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 753.571762] env[61957]: ERROR nova.compute.manager [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] return func(*args, **kwargs) [ 753.571762] env[61957]: ERROR nova.compute.manager [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 753.571762] env[61957]: ERROR nova.compute.manager [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] raise e [ 753.571762] env[61957]: ERROR nova.compute.manager [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 753.571762] env[61957]: ERROR nova.compute.manager [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] nwinfo = self.network_api.allocate_for_instance( [ 753.571762] env[61957]: ERROR nova.compute.manager [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 753.571762] env[61957]: ERROR nova.compute.manager [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] created_port_ids = self._update_ports_for_instance( [ 753.571762] env[61957]: ERROR nova.compute.manager [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 753.571762] env[61957]: ERROR nova.compute.manager [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] with excutils.save_and_reraise_exception(): [ 753.571762] env[61957]: ERROR nova.compute.manager [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 753.571762] env[61957]: ERROR nova.compute.manager [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] self.force_reraise() [ 753.571762] env[61957]: ERROR nova.compute.manager [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 753.572106] env[61957]: ERROR nova.compute.manager [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] raise self.value [ 753.572106] env[61957]: ERROR nova.compute.manager [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 753.572106] env[61957]: ERROR nova.compute.manager [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] updated_port = self._update_port( [ 753.572106] env[61957]: ERROR nova.compute.manager [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 753.572106] env[61957]: ERROR nova.compute.manager [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] _ensure_no_port_binding_failure(port) [ 753.572106] env[61957]: ERROR nova.compute.manager [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 753.572106] env[61957]: ERROR nova.compute.manager [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] raise exception.PortBindingFailed(port_id=port['id']) [ 753.572106] env[61957]: ERROR nova.compute.manager [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] nova.exception.PortBindingFailed: Binding failed for port e0a84501-e615-4647-8dff-5c863b104dc5, please check neutron logs for more information. [ 753.572106] env[61957]: ERROR nova.compute.manager [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] [ 753.572106] env[61957]: DEBUG nova.compute.utils [None req-f462e606-4c03-40e0-ac91-60efbd401559 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] Binding failed for port e0a84501-e615-4647-8dff-5c863b104dc5, please check neutron logs for more information. {{(pid=61957) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 753.572371] env[61957]: DEBUG oslo_concurrency.lockutils [None req-09f1eb26-3ef8-431d-92bc-9c62b712a196 tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.895s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 753.572371] env[61957]: INFO nova.compute.claims [None req-09f1eb26-3ef8-431d-92bc-9c62b712a196 tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] [instance: 60ec4271-2614-470d-bf40-47b9955f544c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 753.574345] env[61957]: DEBUG nova.compute.manager [None req-f462e606-4c03-40e0-ac91-60efbd401559 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] Build of instance 115c1bb4-0e6b-4c6c-9d38-9584debea01f was re-scheduled: Binding failed for port e0a84501-e615-4647-8dff-5c863b104dc5, please check neutron logs for more information. {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 753.574867] env[61957]: DEBUG nova.compute.manager [None req-f462e606-4c03-40e0-ac91-60efbd401559 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] Unplugging VIFs for instance {{(pid=61957) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 753.574969] env[61957]: DEBUG oslo_concurrency.lockutils [None req-f462e606-4c03-40e0-ac91-60efbd401559 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Acquiring lock "refresh_cache-115c1bb4-0e6b-4c6c-9d38-9584debea01f" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 753.575117] env[61957]: DEBUG oslo_concurrency.lockutils [None req-f462e606-4c03-40e0-ac91-60efbd401559 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Acquired lock "refresh_cache-115c1bb4-0e6b-4c6c-9d38-9584debea01f" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 753.575270] env[61957]: DEBUG nova.network.neutron [None req-f462e606-4c03-40e0-ac91-60efbd401559 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 753.715768] env[61957]: DEBUG nova.compute.manager [None req-47718518-520f-4177-b41f-8f0cae78b6f6 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] [instance: 7f3850c0-a455-4e5b-a329-fb3b02f9725f] Start spawning the instance on the hypervisor. {{(pid=61957) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 753.745011] env[61957]: DEBUG nova.virt.hardware [None req-47718518-520f-4177-b41f-8f0cae78b6f6 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T17:22:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T17:21:50Z,direct_url=,disk_format='vmdk',id=11c76a2c-f705-470a-ba9d-4657858bab38,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='81d3096d0f094373913fc3dc8f5c3c6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T17:21:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 753.745288] env[61957]: DEBUG nova.virt.hardware [None req-47718518-520f-4177-b41f-8f0cae78b6f6 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Flavor limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 753.745440] env[61957]: DEBUG nova.virt.hardware [None req-47718518-520f-4177-b41f-8f0cae78b6f6 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Image limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 753.745617] env[61957]: DEBUG nova.virt.hardware [None req-47718518-520f-4177-b41f-8f0cae78b6f6 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Flavor pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 753.745761] env[61957]: DEBUG nova.virt.hardware [None req-47718518-520f-4177-b41f-8f0cae78b6f6 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Image pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 753.745908] env[61957]: DEBUG nova.virt.hardware [None req-47718518-520f-4177-b41f-8f0cae78b6f6 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 753.748258] env[61957]: DEBUG nova.virt.hardware [None req-47718518-520f-4177-b41f-8f0cae78b6f6 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 753.748258] env[61957]: DEBUG nova.virt.hardware [None req-47718518-520f-4177-b41f-8f0cae78b6f6 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 753.748258] env[61957]: DEBUG nova.virt.hardware [None req-47718518-520f-4177-b41f-8f0cae78b6f6 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Got 1 possible topologies {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 753.748258] env[61957]: DEBUG nova.virt.hardware [None req-47718518-520f-4177-b41f-8f0cae78b6f6 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 753.748258] env[61957]: DEBUG nova.virt.hardware [None req-47718518-520f-4177-b41f-8f0cae78b6f6 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 753.748986] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b634a642-b2be-47f9-865d-dc41409196a4 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.759767] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0eca396e-e8df-444a-8889-12f7ad722ad5 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.772744] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-47718518-520f-4177-b41f-8f0cae78b6f6 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] [instance: 7f3850c0-a455-4e5b-a329-fb3b02f9725f] Instance VIF info [] {{(pid=61957) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 753.778179] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-47718518-520f-4177-b41f-8f0cae78b6f6 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Creating folder: Project (1e04e2b61508424292bbecf1e9c66c8b). Parent ref: group-v274445. {{(pid=61957) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 753.778453] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-314656dd-a22a-4dd1-a033-84dea58253a2 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.789093] env[61957]: INFO nova.virt.vmwareapi.vm_util [None req-47718518-520f-4177-b41f-8f0cae78b6f6 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Created folder: Project (1e04e2b61508424292bbecf1e9c66c8b) in parent group-v274445. [ 753.789266] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-47718518-520f-4177-b41f-8f0cae78b6f6 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Creating folder: Instances. Parent ref: group-v274465. {{(pid=61957) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 753.789488] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9b012aaa-a486-403d-aa54-b68ae7bdb100 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.797888] env[61957]: INFO nova.virt.vmwareapi.vm_util [None req-47718518-520f-4177-b41f-8f0cae78b6f6 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Created folder: Instances in parent group-v274465. [ 753.798886] env[61957]: DEBUG oslo.service.loopingcall [None req-47718518-520f-4177-b41f-8f0cae78b6f6 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 753.798886] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7f3850c0-a455-4e5b-a329-fb3b02f9725f] Creating VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 753.798886] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e3ef58c4-0a83-48bf-a756-cd51fd6ce134 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.813869] env[61957]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 753.813869] env[61957]: value = "task-1277369" [ 753.813869] env[61957]: _type = "Task" [ 753.813869] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 753.821103] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277369, 'name': CreateVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 754.032159] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d6c9937c-2526-4768-bade-4ca2ced8b569 tempest-ServersTestFqdnHostnames-1308211839 tempest-ServersTestFqdnHostnames-1308211839-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 754.102698] env[61957]: DEBUG nova.network.neutron [None req-f462e606-4c03-40e0-ac91-60efbd401559 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 754.262154] env[61957]: DEBUG nova.network.neutron [None req-f462e606-4c03-40e0-ac91-60efbd401559 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 754.325752] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277369, 'name': CreateVM_Task, 'duration_secs': 0.259308} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 754.326419] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7f3850c0-a455-4e5b-a329-fb3b02f9725f] Created VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 754.326936] env[61957]: DEBUG oslo_concurrency.lockutils [None req-47718518-520f-4177-b41f-8f0cae78b6f6 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 754.327123] env[61957]: DEBUG oslo_concurrency.lockutils [None req-47718518-520f-4177-b41f-8f0cae78b6f6 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 754.327440] env[61957]: DEBUG oslo_concurrency.lockutils [None req-47718518-520f-4177-b41f-8f0cae78b6f6 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 754.327683] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2d2d9c6a-42c7-414b-834b-722353fe73a1 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.334021] env[61957]: DEBUG oslo_vmware.api [None req-47718518-520f-4177-b41f-8f0cae78b6f6 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Waiting for the task: (returnval){ [ 754.334021] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]520ff358-526a-6961-8fda-b0f9780b621a" [ 754.334021] env[61957]: _type = "Task" [ 754.334021] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 754.343041] env[61957]: DEBUG oslo_vmware.api [None req-47718518-520f-4177-b41f-8f0cae78b6f6 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]520ff358-526a-6961-8fda-b0f9780b621a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 754.767508] env[61957]: DEBUG oslo_concurrency.lockutils [None req-f462e606-4c03-40e0-ac91-60efbd401559 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Releasing lock "refresh_cache-115c1bb4-0e6b-4c6c-9d38-9584debea01f" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 754.767508] env[61957]: DEBUG nova.compute.manager [None req-f462e606-4c03-40e0-ac91-60efbd401559 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61957) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 754.767508] env[61957]: DEBUG nova.compute.manager [None req-f462e606-4c03-40e0-ac91-60efbd401559 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 754.767768] env[61957]: DEBUG nova.network.neutron [None req-f462e606-4c03-40e0-ac91-60efbd401559 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 754.784503] env[61957]: DEBUG nova.network.neutron [None req-f462e606-4c03-40e0-ac91-60efbd401559 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 754.845851] env[61957]: DEBUG oslo_vmware.api [None req-47718518-520f-4177-b41f-8f0cae78b6f6 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]520ff358-526a-6961-8fda-b0f9780b621a, 'name': SearchDatastore_Task, 'duration_secs': 0.009371} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 754.846283] env[61957]: DEBUG oslo_concurrency.lockutils [None req-47718518-520f-4177-b41f-8f0cae78b6f6 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 754.846625] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-47718518-520f-4177-b41f-8f0cae78b6f6 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] [instance: 7f3850c0-a455-4e5b-a329-fb3b02f9725f] Processing image 11c76a2c-f705-470a-ba9d-4657858bab38 {{(pid=61957) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 754.846914] env[61957]: DEBUG oslo_concurrency.lockutils [None req-47718518-520f-4177-b41f-8f0cae78b6f6 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 754.847103] env[61957]: DEBUG oslo_concurrency.lockutils [None req-47718518-520f-4177-b41f-8f0cae78b6f6 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 754.847332] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-47718518-520f-4177-b41f-8f0cae78b6f6 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 754.847620] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-34b7c987-994f-46b6-a64b-d6aac48debc4 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.859297] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-47718518-520f-4177-b41f-8f0cae78b6f6 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 754.859486] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-47718518-520f-4177-b41f-8f0cae78b6f6 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61957) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 754.860241] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b0c39b7f-458e-4285-9aa1-c10dbc09ae10 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.868526] env[61957]: DEBUG oslo_vmware.api [None req-47718518-520f-4177-b41f-8f0cae78b6f6 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Waiting for the task: (returnval){ [ 754.868526] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]528ec777-a672-9d6e-0702-0f772151c8cf" [ 754.868526] env[61957]: _type = "Task" [ 754.868526] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 754.876365] env[61957]: DEBUG oslo_vmware.api [None req-47718518-520f-4177-b41f-8f0cae78b6f6 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]528ec777-a672-9d6e-0702-0f772151c8cf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 754.954183] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c493d926-4820-4f19-b2f8-9295cc4e3498 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.962453] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-870fc15e-5138-4775-b018-1edfc92e422d {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.992202] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5603ab30-9d25-43a2-93ec-c5f819e1ce48 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.999761] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d568d0b-cf12-479a-adaf-4f8c2545e436 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.012339] env[61957]: DEBUG nova.compute.provider_tree [None req-09f1eb26-3ef8-431d-92bc-9c62b712a196 tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 755.287352] env[61957]: DEBUG nova.network.neutron [None req-f462e606-4c03-40e0-ac91-60efbd401559 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 755.378856] env[61957]: DEBUG oslo_vmware.api [None req-47718518-520f-4177-b41f-8f0cae78b6f6 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]528ec777-a672-9d6e-0702-0f772151c8cf, 'name': SearchDatastore_Task, 'duration_secs': 0.008618} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 755.379644] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-97d7868c-c888-4e27-9bb8-7e41a6dc016d {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.384553] env[61957]: DEBUG oslo_vmware.api [None req-47718518-520f-4177-b41f-8f0cae78b6f6 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Waiting for the task: (returnval){ [ 755.384553] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]5249d751-7f8d-e8bd-a51d-40520e09dcad" [ 755.384553] env[61957]: _type = "Task" [ 755.384553] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 755.391948] env[61957]: DEBUG oslo_vmware.api [None req-47718518-520f-4177-b41f-8f0cae78b6f6 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]5249d751-7f8d-e8bd-a51d-40520e09dcad, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 755.515539] env[61957]: DEBUG nova.scheduler.client.report [None req-09f1eb26-3ef8-431d-92bc-9c62b712a196 tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 755.789623] env[61957]: INFO nova.compute.manager [None req-f462e606-4c03-40e0-ac91-60efbd401559 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] [instance: 115c1bb4-0e6b-4c6c-9d38-9584debea01f] Took 1.02 seconds to deallocate network for instance. [ 755.895455] env[61957]: DEBUG oslo_vmware.api [None req-47718518-520f-4177-b41f-8f0cae78b6f6 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]5249d751-7f8d-e8bd-a51d-40520e09dcad, 'name': SearchDatastore_Task, 'duration_secs': 0.009622} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 755.895749] env[61957]: DEBUG oslo_concurrency.lockutils [None req-47718518-520f-4177-b41f-8f0cae78b6f6 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 755.896036] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-47718518-520f-4177-b41f-8f0cae78b6f6 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore2] 7f3850c0-a455-4e5b-a329-fb3b02f9725f/7f3850c0-a455-4e5b-a329-fb3b02f9725f.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 755.896550] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-fff909b0-b0b2-4e6c-8825-bd3d528720fa {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.903855] env[61957]: DEBUG oslo_vmware.api [None req-47718518-520f-4177-b41f-8f0cae78b6f6 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Waiting for the task: (returnval){ [ 755.903855] env[61957]: value = "task-1277370" [ 755.903855] env[61957]: _type = "Task" [ 755.903855] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 755.911958] env[61957]: DEBUG oslo_vmware.api [None req-47718518-520f-4177-b41f-8f0cae78b6f6 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Task: {'id': task-1277370, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 756.020621] env[61957]: DEBUG oslo_concurrency.lockutils [None req-09f1eb26-3ef8-431d-92bc-9c62b712a196 tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.450s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 756.021191] env[61957]: DEBUG nova.compute.manager [None req-09f1eb26-3ef8-431d-92bc-9c62b712a196 tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] [instance: 60ec4271-2614-470d-bf40-47b9955f544c] Start building networks asynchronously for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 756.023748] env[61957]: DEBUG oslo_concurrency.lockutils [None req-6b91bce8-d0f8-4c40-855e-ed34f3eb7e64 tempest-ServerActionsTestJSON-1567241756 tempest-ServerActionsTestJSON-1567241756-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.450s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 756.025211] env[61957]: INFO nova.compute.claims [None req-6b91bce8-d0f8-4c40-855e-ed34f3eb7e64 tempest-ServerActionsTestJSON-1567241756 tempest-ServerActionsTestJSON-1567241756-project-member] [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 756.413299] env[61957]: DEBUG oslo_vmware.api [None req-47718518-520f-4177-b41f-8f0cae78b6f6 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Task: {'id': task-1277370, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.439904} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 756.413599] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-47718518-520f-4177-b41f-8f0cae78b6f6 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore2] 7f3850c0-a455-4e5b-a329-fb3b02f9725f/7f3850c0-a455-4e5b-a329-fb3b02f9725f.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 756.413766] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-47718518-520f-4177-b41f-8f0cae78b6f6 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] [instance: 7f3850c0-a455-4e5b-a329-fb3b02f9725f] Extending root virtual disk to 1048576 {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 756.414015] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-cd8cc5e3-3ea5-4893-a84f-f514cb683d93 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.420583] env[61957]: DEBUG oslo_vmware.api [None req-47718518-520f-4177-b41f-8f0cae78b6f6 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Waiting for the task: (returnval){ [ 756.420583] env[61957]: value = "task-1277371" [ 756.420583] env[61957]: _type = "Task" [ 756.420583] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 756.428407] env[61957]: DEBUG oslo_vmware.api [None req-47718518-520f-4177-b41f-8f0cae78b6f6 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Task: {'id': task-1277371, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 756.533560] env[61957]: DEBUG nova.compute.utils [None req-09f1eb26-3ef8-431d-92bc-9c62b712a196 tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] Using /dev/sd instead of None {{(pid=61957) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 756.536891] env[61957]: DEBUG nova.compute.manager [None req-09f1eb26-3ef8-431d-92bc-9c62b712a196 tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] [instance: 60ec4271-2614-470d-bf40-47b9955f544c] Allocating IP information in the background. {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 756.537419] env[61957]: DEBUG nova.network.neutron [None req-09f1eb26-3ef8-431d-92bc-9c62b712a196 tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] [instance: 60ec4271-2614-470d-bf40-47b9955f544c] allocate_for_instance() {{(pid=61957) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 756.584699] env[61957]: DEBUG nova.policy [None req-09f1eb26-3ef8-431d-92bc-9c62b712a196 tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6871e6657be047708fb735ffb71dc19d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f1243661c0ff4c33b901de6aacfd5987', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61957) authorize /opt/stack/nova/nova/policy.py:203}} [ 756.827439] env[61957]: INFO nova.scheduler.client.report [None req-f462e606-4c03-40e0-ac91-60efbd401559 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Deleted allocations for instance 115c1bb4-0e6b-4c6c-9d38-9584debea01f [ 756.853781] env[61957]: DEBUG nova.network.neutron [None req-09f1eb26-3ef8-431d-92bc-9c62b712a196 tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] [instance: 60ec4271-2614-470d-bf40-47b9955f544c] Successfully created port: de4fd670-560d-4cad-ba00-671731d57410 {{(pid=61957) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 756.930532] env[61957]: DEBUG oslo_vmware.api [None req-47718518-520f-4177-b41f-8f0cae78b6f6 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Task: {'id': task-1277371, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.056975} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 756.930802] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-47718518-520f-4177-b41f-8f0cae78b6f6 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] [instance: 7f3850c0-a455-4e5b-a329-fb3b02f9725f] Extended root virtual disk {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 756.931607] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36c490cb-51ea-4de4-a0ac-974db608f56e {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.951474] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-47718518-520f-4177-b41f-8f0cae78b6f6 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] [instance: 7f3850c0-a455-4e5b-a329-fb3b02f9725f] Reconfiguring VM instance instance-00000028 to attach disk [datastore2] 7f3850c0-a455-4e5b-a329-fb3b02f9725f/7f3850c0-a455-4e5b-a329-fb3b02f9725f.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 756.951742] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-516de064-e7f9-456f-9a94-ba5ff09cfe20 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.972128] env[61957]: DEBUG oslo_vmware.api [None req-47718518-520f-4177-b41f-8f0cae78b6f6 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Waiting for the task: (returnval){ [ 756.972128] env[61957]: value = "task-1277372" [ 756.972128] env[61957]: _type = "Task" [ 756.972128] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 756.979708] env[61957]: DEBUG oslo_vmware.api [None req-47718518-520f-4177-b41f-8f0cae78b6f6 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Task: {'id': task-1277372, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 757.038360] env[61957]: DEBUG nova.compute.manager [None req-09f1eb26-3ef8-431d-92bc-9c62b712a196 tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] [instance: 60ec4271-2614-470d-bf40-47b9955f544c] Start building block device mappings for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 757.341603] env[61957]: DEBUG oslo_concurrency.lockutils [None req-f462e606-4c03-40e0-ac91-60efbd401559 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Lock "115c1bb4-0e6b-4c6c-9d38-9584debea01f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 126.296s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 757.379147] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3dc8b976-d62e-496a-b1bc-43407c233233 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.388399] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1329d86-2974-4fba-91fe-dab825928dd9 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.424018] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8254cca-5878-499a-aaca-b407be38783c {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.429316] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8d41451-1b7c-4f3e-8c96-632d6ba636dc {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.442701] env[61957]: DEBUG nova.compute.provider_tree [None req-6b91bce8-d0f8-4c40-855e-ed34f3eb7e64 tempest-ServerActionsTestJSON-1567241756 tempest-ServerActionsTestJSON-1567241756-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 757.481797] env[61957]: DEBUG oslo_vmware.api [None req-47718518-520f-4177-b41f-8f0cae78b6f6 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Task: {'id': task-1277372, 'name': ReconfigVM_Task, 'duration_secs': 0.274802} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 757.482087] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-47718518-520f-4177-b41f-8f0cae78b6f6 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] [instance: 7f3850c0-a455-4e5b-a329-fb3b02f9725f] Reconfigured VM instance instance-00000028 to attach disk [datastore2] 7f3850c0-a455-4e5b-a329-fb3b02f9725f/7f3850c0-a455-4e5b-a329-fb3b02f9725f.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 757.482739] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c40ef040-8cbd-4fd2-8458-2748ac977a4c {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.488634] env[61957]: DEBUG oslo_vmware.api [None req-47718518-520f-4177-b41f-8f0cae78b6f6 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Waiting for the task: (returnval){ [ 757.488634] env[61957]: value = "task-1277373" [ 757.488634] env[61957]: _type = "Task" [ 757.488634] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 757.495950] env[61957]: DEBUG oslo_vmware.api [None req-47718518-520f-4177-b41f-8f0cae78b6f6 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Task: {'id': task-1277373, 'name': Rename_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 757.601592] env[61957]: DEBUG nova.compute.manager [req-145815f0-f6eb-4bff-b208-90f675e8e9a7 req-4df81e78-e3f8-46c8-888f-38acbd161542 service nova] [instance: 60ec4271-2614-470d-bf40-47b9955f544c] Received event network-changed-de4fd670-560d-4cad-ba00-671731d57410 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 757.601810] env[61957]: DEBUG nova.compute.manager [req-145815f0-f6eb-4bff-b208-90f675e8e9a7 req-4df81e78-e3f8-46c8-888f-38acbd161542 service nova] [instance: 60ec4271-2614-470d-bf40-47b9955f544c] Refreshing instance network info cache due to event network-changed-de4fd670-560d-4cad-ba00-671731d57410. {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 757.602011] env[61957]: DEBUG oslo_concurrency.lockutils [req-145815f0-f6eb-4bff-b208-90f675e8e9a7 req-4df81e78-e3f8-46c8-888f-38acbd161542 service nova] Acquiring lock "refresh_cache-60ec4271-2614-470d-bf40-47b9955f544c" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 757.602169] env[61957]: DEBUG oslo_concurrency.lockutils [req-145815f0-f6eb-4bff-b208-90f675e8e9a7 req-4df81e78-e3f8-46c8-888f-38acbd161542 service nova] Acquired lock "refresh_cache-60ec4271-2614-470d-bf40-47b9955f544c" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 757.602330] env[61957]: DEBUG nova.network.neutron [req-145815f0-f6eb-4bff-b208-90f675e8e9a7 req-4df81e78-e3f8-46c8-888f-38acbd161542 service nova] [instance: 60ec4271-2614-470d-bf40-47b9955f544c] Refreshing network info cache for port de4fd670-560d-4cad-ba00-671731d57410 {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 757.740806] env[61957]: ERROR nova.compute.manager [None req-09f1eb26-3ef8-431d-92bc-9c62b712a196 tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port de4fd670-560d-4cad-ba00-671731d57410, please check neutron logs for more information. [ 757.740806] env[61957]: ERROR nova.compute.manager Traceback (most recent call last): [ 757.740806] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 757.740806] env[61957]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 757.740806] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 757.740806] env[61957]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 757.740806] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 757.740806] env[61957]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 757.740806] env[61957]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 757.740806] env[61957]: ERROR nova.compute.manager self.force_reraise() [ 757.740806] env[61957]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 757.740806] env[61957]: ERROR nova.compute.manager raise self.value [ 757.740806] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 757.740806] env[61957]: ERROR nova.compute.manager updated_port = self._update_port( [ 757.740806] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 757.740806] env[61957]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 757.741367] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 757.741367] env[61957]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 757.741367] env[61957]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port de4fd670-560d-4cad-ba00-671731d57410, please check neutron logs for more information. [ 757.741367] env[61957]: ERROR nova.compute.manager [ 757.741367] env[61957]: Traceback (most recent call last): [ 757.741367] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 757.741367] env[61957]: listener.cb(fileno) [ 757.741367] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 757.741367] env[61957]: result = function(*args, **kwargs) [ 757.741367] env[61957]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 757.741367] env[61957]: return func(*args, **kwargs) [ 757.741367] env[61957]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 757.741367] env[61957]: raise e [ 757.741367] env[61957]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 757.741367] env[61957]: nwinfo = self.network_api.allocate_for_instance( [ 757.741367] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 757.741367] env[61957]: created_port_ids = self._update_ports_for_instance( [ 757.741367] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 757.741367] env[61957]: with excutils.save_and_reraise_exception(): [ 757.741367] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 757.741367] env[61957]: self.force_reraise() [ 757.741367] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 757.741367] env[61957]: raise self.value [ 757.741367] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 757.741367] env[61957]: updated_port = self._update_port( [ 757.741367] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 757.741367] env[61957]: _ensure_no_port_binding_failure(port) [ 757.741367] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 757.741367] env[61957]: raise exception.PortBindingFailed(port_id=port['id']) [ 757.742166] env[61957]: nova.exception.PortBindingFailed: Binding failed for port de4fd670-560d-4cad-ba00-671731d57410, please check neutron logs for more information. [ 757.742166] env[61957]: Removing descriptor: 16 [ 757.843158] env[61957]: DEBUG nova.compute.manager [None req-bb8cb37b-74c5-4607-9ba8-e5615134a1da tempest-InstanceActionsTestJSON-565478614 tempest-InstanceActionsTestJSON-565478614-project-member] [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] Starting instance... {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 757.945914] env[61957]: DEBUG nova.scheduler.client.report [None req-6b91bce8-d0f8-4c40-855e-ed34f3eb7e64 tempest-ServerActionsTestJSON-1567241756 tempest-ServerActionsTestJSON-1567241756-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 757.999534] env[61957]: DEBUG oslo_vmware.api [None req-47718518-520f-4177-b41f-8f0cae78b6f6 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Task: {'id': task-1277373, 'name': Rename_Task, 'duration_secs': 0.1355} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 757.999861] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-47718518-520f-4177-b41f-8f0cae78b6f6 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] [instance: 7f3850c0-a455-4e5b-a329-fb3b02f9725f] Powering on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 758.000126] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f1a91bc4-06a8-4774-b24a-cf0d9fecb822 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.007096] env[61957]: DEBUG oslo_vmware.api [None req-47718518-520f-4177-b41f-8f0cae78b6f6 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Waiting for the task: (returnval){ [ 758.007096] env[61957]: value = "task-1277374" [ 758.007096] env[61957]: _type = "Task" [ 758.007096] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 758.015421] env[61957]: DEBUG oslo_vmware.api [None req-47718518-520f-4177-b41f-8f0cae78b6f6 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Task: {'id': task-1277374, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 758.050116] env[61957]: DEBUG nova.compute.manager [None req-09f1eb26-3ef8-431d-92bc-9c62b712a196 tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] [instance: 60ec4271-2614-470d-bf40-47b9955f544c] Start spawning the instance on the hypervisor. {{(pid=61957) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 758.080901] env[61957]: DEBUG nova.virt.hardware [None req-09f1eb26-3ef8-431d-92bc-9c62b712a196 tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T17:22:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T17:21:50Z,direct_url=,disk_format='vmdk',id=11c76a2c-f705-470a-ba9d-4657858bab38,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='81d3096d0f094373913fc3dc8f5c3c6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T17:21:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 758.081230] env[61957]: DEBUG nova.virt.hardware [None req-09f1eb26-3ef8-431d-92bc-9c62b712a196 tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] Flavor limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 758.081421] env[61957]: DEBUG nova.virt.hardware [None req-09f1eb26-3ef8-431d-92bc-9c62b712a196 tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] Image limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 758.081620] env[61957]: DEBUG nova.virt.hardware [None req-09f1eb26-3ef8-431d-92bc-9c62b712a196 tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] Flavor pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 758.081796] env[61957]: DEBUG nova.virt.hardware [None req-09f1eb26-3ef8-431d-92bc-9c62b712a196 tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] Image pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 758.081977] env[61957]: DEBUG nova.virt.hardware [None req-09f1eb26-3ef8-431d-92bc-9c62b712a196 tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 758.082473] env[61957]: DEBUG nova.virt.hardware [None req-09f1eb26-3ef8-431d-92bc-9c62b712a196 tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 758.082678] env[61957]: DEBUG nova.virt.hardware [None req-09f1eb26-3ef8-431d-92bc-9c62b712a196 tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 758.082881] env[61957]: DEBUG nova.virt.hardware [None req-09f1eb26-3ef8-431d-92bc-9c62b712a196 tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] Got 1 possible topologies {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 758.083102] env[61957]: DEBUG nova.virt.hardware [None req-09f1eb26-3ef8-431d-92bc-9c62b712a196 tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 758.083322] env[61957]: DEBUG nova.virt.hardware [None req-09f1eb26-3ef8-431d-92bc-9c62b712a196 tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 758.084235] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-617906e9-fbc5-4df6-a9fd-356f49b93ac6 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.096586] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bc63778-8978-42e8-9f9c-c3995e8c1b3e {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.115132] env[61957]: ERROR nova.compute.manager [None req-09f1eb26-3ef8-431d-92bc-9c62b712a196 tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] [instance: 60ec4271-2614-470d-bf40-47b9955f544c] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port de4fd670-560d-4cad-ba00-671731d57410, please check neutron logs for more information. [ 758.115132] env[61957]: ERROR nova.compute.manager [instance: 60ec4271-2614-470d-bf40-47b9955f544c] Traceback (most recent call last): [ 758.115132] env[61957]: ERROR nova.compute.manager [instance: 60ec4271-2614-470d-bf40-47b9955f544c] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 758.115132] env[61957]: ERROR nova.compute.manager [instance: 60ec4271-2614-470d-bf40-47b9955f544c] yield resources [ 758.115132] env[61957]: ERROR nova.compute.manager [instance: 60ec4271-2614-470d-bf40-47b9955f544c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 758.115132] env[61957]: ERROR nova.compute.manager [instance: 60ec4271-2614-470d-bf40-47b9955f544c] self.driver.spawn(context, instance, image_meta, [ 758.115132] env[61957]: ERROR nova.compute.manager [instance: 60ec4271-2614-470d-bf40-47b9955f544c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 758.115132] env[61957]: ERROR nova.compute.manager [instance: 60ec4271-2614-470d-bf40-47b9955f544c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 758.115132] env[61957]: ERROR nova.compute.manager [instance: 60ec4271-2614-470d-bf40-47b9955f544c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 758.115132] env[61957]: ERROR nova.compute.manager [instance: 60ec4271-2614-470d-bf40-47b9955f544c] vm_ref = self.build_virtual_machine(instance, [ 758.115132] env[61957]: ERROR nova.compute.manager [instance: 60ec4271-2614-470d-bf40-47b9955f544c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 758.115573] env[61957]: ERROR nova.compute.manager [instance: 60ec4271-2614-470d-bf40-47b9955f544c] vif_infos = vmwarevif.get_vif_info(self._session, [ 758.115573] env[61957]: ERROR nova.compute.manager [instance: 60ec4271-2614-470d-bf40-47b9955f544c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 758.115573] env[61957]: ERROR nova.compute.manager [instance: 60ec4271-2614-470d-bf40-47b9955f544c] for vif in network_info: [ 758.115573] env[61957]: ERROR nova.compute.manager [instance: 60ec4271-2614-470d-bf40-47b9955f544c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 758.115573] env[61957]: ERROR nova.compute.manager [instance: 60ec4271-2614-470d-bf40-47b9955f544c] return self._sync_wrapper(fn, *args, **kwargs) [ 758.115573] env[61957]: ERROR nova.compute.manager [instance: 60ec4271-2614-470d-bf40-47b9955f544c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 758.115573] env[61957]: ERROR nova.compute.manager [instance: 60ec4271-2614-470d-bf40-47b9955f544c] self.wait() [ 758.115573] env[61957]: ERROR nova.compute.manager [instance: 60ec4271-2614-470d-bf40-47b9955f544c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 758.115573] env[61957]: ERROR nova.compute.manager [instance: 60ec4271-2614-470d-bf40-47b9955f544c] self[:] = self._gt.wait() [ 758.115573] env[61957]: ERROR nova.compute.manager [instance: 60ec4271-2614-470d-bf40-47b9955f544c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 758.115573] env[61957]: ERROR nova.compute.manager [instance: 60ec4271-2614-470d-bf40-47b9955f544c] return self._exit_event.wait() [ 758.115573] env[61957]: ERROR nova.compute.manager [instance: 60ec4271-2614-470d-bf40-47b9955f544c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 758.115573] env[61957]: ERROR nova.compute.manager [instance: 60ec4271-2614-470d-bf40-47b9955f544c] current.throw(*self._exc) [ 758.116071] env[61957]: ERROR nova.compute.manager [instance: 60ec4271-2614-470d-bf40-47b9955f544c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 758.116071] env[61957]: ERROR nova.compute.manager [instance: 60ec4271-2614-470d-bf40-47b9955f544c] result = function(*args, **kwargs) [ 758.116071] env[61957]: ERROR nova.compute.manager [instance: 60ec4271-2614-470d-bf40-47b9955f544c] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 758.116071] env[61957]: ERROR nova.compute.manager [instance: 60ec4271-2614-470d-bf40-47b9955f544c] return func(*args, **kwargs) [ 758.116071] env[61957]: ERROR nova.compute.manager [instance: 60ec4271-2614-470d-bf40-47b9955f544c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 758.116071] env[61957]: ERROR nova.compute.manager [instance: 60ec4271-2614-470d-bf40-47b9955f544c] raise e [ 758.116071] env[61957]: ERROR nova.compute.manager [instance: 60ec4271-2614-470d-bf40-47b9955f544c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 758.116071] env[61957]: ERROR nova.compute.manager [instance: 60ec4271-2614-470d-bf40-47b9955f544c] nwinfo = self.network_api.allocate_for_instance( [ 758.116071] env[61957]: ERROR nova.compute.manager [instance: 60ec4271-2614-470d-bf40-47b9955f544c] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 758.116071] env[61957]: ERROR nova.compute.manager [instance: 60ec4271-2614-470d-bf40-47b9955f544c] created_port_ids = self._update_ports_for_instance( [ 758.116071] env[61957]: ERROR nova.compute.manager [instance: 60ec4271-2614-470d-bf40-47b9955f544c] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 758.116071] env[61957]: ERROR nova.compute.manager [instance: 60ec4271-2614-470d-bf40-47b9955f544c] with excutils.save_and_reraise_exception(): [ 758.116071] env[61957]: ERROR nova.compute.manager [instance: 60ec4271-2614-470d-bf40-47b9955f544c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 758.116453] env[61957]: ERROR nova.compute.manager [instance: 60ec4271-2614-470d-bf40-47b9955f544c] self.force_reraise() [ 758.116453] env[61957]: ERROR nova.compute.manager [instance: 60ec4271-2614-470d-bf40-47b9955f544c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 758.116453] env[61957]: ERROR nova.compute.manager [instance: 60ec4271-2614-470d-bf40-47b9955f544c] raise self.value [ 758.116453] env[61957]: ERROR nova.compute.manager [instance: 60ec4271-2614-470d-bf40-47b9955f544c] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 758.116453] env[61957]: ERROR nova.compute.manager [instance: 60ec4271-2614-470d-bf40-47b9955f544c] updated_port = self._update_port( [ 758.116453] env[61957]: ERROR nova.compute.manager [instance: 60ec4271-2614-470d-bf40-47b9955f544c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 758.116453] env[61957]: ERROR nova.compute.manager [instance: 60ec4271-2614-470d-bf40-47b9955f544c] _ensure_no_port_binding_failure(port) [ 758.116453] env[61957]: ERROR nova.compute.manager [instance: 60ec4271-2614-470d-bf40-47b9955f544c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 758.116453] env[61957]: ERROR nova.compute.manager [instance: 60ec4271-2614-470d-bf40-47b9955f544c] raise exception.PortBindingFailed(port_id=port['id']) [ 758.116453] env[61957]: ERROR nova.compute.manager [instance: 60ec4271-2614-470d-bf40-47b9955f544c] nova.exception.PortBindingFailed: Binding failed for port de4fd670-560d-4cad-ba00-671731d57410, please check neutron logs for more information. [ 758.116453] env[61957]: ERROR nova.compute.manager [instance: 60ec4271-2614-470d-bf40-47b9955f544c] [ 758.116453] env[61957]: INFO nova.compute.manager [None req-09f1eb26-3ef8-431d-92bc-9c62b712a196 tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] [instance: 60ec4271-2614-470d-bf40-47b9955f544c] Terminating instance [ 758.117572] env[61957]: DEBUG oslo_concurrency.lockutils [None req-09f1eb26-3ef8-431d-92bc-9c62b712a196 tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] Acquiring lock "refresh_cache-60ec4271-2614-470d-bf40-47b9955f544c" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 758.135360] env[61957]: DEBUG nova.network.neutron [req-145815f0-f6eb-4bff-b208-90f675e8e9a7 req-4df81e78-e3f8-46c8-888f-38acbd161542 service nova] [instance: 60ec4271-2614-470d-bf40-47b9955f544c] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 758.249432] env[61957]: DEBUG nova.network.neutron [req-145815f0-f6eb-4bff-b208-90f675e8e9a7 req-4df81e78-e3f8-46c8-888f-38acbd161542 service nova] [instance: 60ec4271-2614-470d-bf40-47b9955f544c] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 758.375021] env[61957]: DEBUG oslo_concurrency.lockutils [None req-bb8cb37b-74c5-4607-9ba8-e5615134a1da tempest-InstanceActionsTestJSON-565478614 tempest-InstanceActionsTestJSON-565478614-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 758.453136] env[61957]: DEBUG oslo_concurrency.lockutils [None req-6b91bce8-d0f8-4c40-855e-ed34f3eb7e64 tempest-ServerActionsTestJSON-1567241756 tempest-ServerActionsTestJSON-1567241756-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.427s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 758.453136] env[61957]: DEBUG nova.compute.manager [None req-6b91bce8-d0f8-4c40-855e-ed34f3eb7e64 tempest-ServerActionsTestJSON-1567241756 tempest-ServerActionsTestJSON-1567241756-project-member] [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] Start building networks asynchronously for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 758.456029] env[61957]: DEBUG oslo_concurrency.lockutils [None req-660095f0-9d5a-4797-b621-e0585ffddbba tempest-ServerActionsTestOtherB-303558607 tempest-ServerActionsTestOtherB-303558607-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.988s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 758.457062] env[61957]: INFO nova.compute.claims [None req-660095f0-9d5a-4797-b621-e0585ffddbba tempest-ServerActionsTestOtherB-303558607 tempest-ServerActionsTestOtherB-303558607-project-member] [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 758.519088] env[61957]: DEBUG oslo_vmware.api [None req-47718518-520f-4177-b41f-8f0cae78b6f6 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Task: {'id': task-1277374, 'name': PowerOnVM_Task, 'duration_secs': 0.445423} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 758.519559] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-47718518-520f-4177-b41f-8f0cae78b6f6 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] [instance: 7f3850c0-a455-4e5b-a329-fb3b02f9725f] Powered on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 758.523095] env[61957]: INFO nova.compute.manager [None req-47718518-520f-4177-b41f-8f0cae78b6f6 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] [instance: 7f3850c0-a455-4e5b-a329-fb3b02f9725f] Took 4.80 seconds to spawn the instance on the hypervisor. [ 758.523095] env[61957]: DEBUG nova.compute.manager [None req-47718518-520f-4177-b41f-8f0cae78b6f6 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] [instance: 7f3850c0-a455-4e5b-a329-fb3b02f9725f] Checking state {{(pid=61957) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 758.523095] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f46caf3-8c4f-499d-b227-647e1b3938f4 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.754018] env[61957]: DEBUG oslo_concurrency.lockutils [req-145815f0-f6eb-4bff-b208-90f675e8e9a7 req-4df81e78-e3f8-46c8-888f-38acbd161542 service nova] Releasing lock "refresh_cache-60ec4271-2614-470d-bf40-47b9955f544c" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 758.754018] env[61957]: DEBUG oslo_concurrency.lockutils [None req-09f1eb26-3ef8-431d-92bc-9c62b712a196 tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] Acquired lock "refresh_cache-60ec4271-2614-470d-bf40-47b9955f544c" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 758.754018] env[61957]: DEBUG nova.network.neutron [None req-09f1eb26-3ef8-431d-92bc-9c62b712a196 tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] [instance: 60ec4271-2614-470d-bf40-47b9955f544c] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 758.965181] env[61957]: DEBUG nova.compute.utils [None req-6b91bce8-d0f8-4c40-855e-ed34f3eb7e64 tempest-ServerActionsTestJSON-1567241756 tempest-ServerActionsTestJSON-1567241756-project-member] Using /dev/sd instead of None {{(pid=61957) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 758.966820] env[61957]: DEBUG nova.compute.manager [None req-6b91bce8-d0f8-4c40-855e-ed34f3eb7e64 tempest-ServerActionsTestJSON-1567241756 tempest-ServerActionsTestJSON-1567241756-project-member] [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] Allocating IP information in the background. {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 758.967200] env[61957]: DEBUG nova.network.neutron [None req-6b91bce8-d0f8-4c40-855e-ed34f3eb7e64 tempest-ServerActionsTestJSON-1567241756 tempest-ServerActionsTestJSON-1567241756-project-member] [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] allocate_for_instance() {{(pid=61957) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 759.010106] env[61957]: DEBUG nova.policy [None req-6b91bce8-d0f8-4c40-855e-ed34f3eb7e64 tempest-ServerActionsTestJSON-1567241756 tempest-ServerActionsTestJSON-1567241756-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9dcee3898f114a02a19211b35dc8149d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5b6658fd3386438eb5a44866066ebb65', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61957) authorize /opt/stack/nova/nova/policy.py:203}} [ 759.046253] env[61957]: INFO nova.compute.manager [None req-47718518-520f-4177-b41f-8f0cae78b6f6 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] [instance: 7f3850c0-a455-4e5b-a329-fb3b02f9725f] Took 23.91 seconds to build instance. [ 759.274489] env[61957]: DEBUG nova.network.neutron [None req-09f1eb26-3ef8-431d-92bc-9c62b712a196 tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] [instance: 60ec4271-2614-470d-bf40-47b9955f544c] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 759.340558] env[61957]: INFO nova.compute.manager [None req-3617f05d-8cb2-4824-b3d0-fbcde866b375 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] [instance: 7f3850c0-a455-4e5b-a329-fb3b02f9725f] Rebuilding instance [ 759.346765] env[61957]: DEBUG nova.network.neutron [None req-6b91bce8-d0f8-4c40-855e-ed34f3eb7e64 tempest-ServerActionsTestJSON-1567241756 tempest-ServerActionsTestJSON-1567241756-project-member] [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] Successfully created port: 3233fc75-1990-45ba-992a-b932f2d5eafc {{(pid=61957) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 759.389712] env[61957]: DEBUG nova.network.neutron [None req-09f1eb26-3ef8-431d-92bc-9c62b712a196 tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] [instance: 60ec4271-2614-470d-bf40-47b9955f544c] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 759.395824] env[61957]: DEBUG nova.compute.manager [None req-3617f05d-8cb2-4824-b3d0-fbcde866b375 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] [instance: 7f3850c0-a455-4e5b-a329-fb3b02f9725f] Checking state {{(pid=61957) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 759.399307] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9d4004e-2a46-478b-8377-161afcb34e85 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.468106] env[61957]: DEBUG nova.compute.manager [None req-6b91bce8-d0f8-4c40-855e-ed34f3eb7e64 tempest-ServerActionsTestJSON-1567241756 tempest-ServerActionsTestJSON-1567241756-project-member] [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] Start building block device mappings for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 759.546006] env[61957]: DEBUG oslo_concurrency.lockutils [None req-47718518-520f-4177-b41f-8f0cae78b6f6 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Lock "7f3850c0-a455-4e5b-a329-fb3b02f9725f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 125.353s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 759.764800] env[61957]: DEBUG nova.compute.manager [req-be77ef46-19b6-4b62-b770-4fc080ea227f req-64855694-df33-480b-956a-fb561429cc23 service nova] [instance: 60ec4271-2614-470d-bf40-47b9955f544c] Received event network-vif-deleted-de4fd670-560d-4cad-ba00-671731d57410 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 759.816386] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0553c520-f115-4ec2-bdbe-11b52b825736 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.825659] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0353fa1-ba20-4283-a54c-7b69e8d2cc32 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.855030] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc2a56eb-da7c-4c78-9f92-7e1723a21b76 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.862861] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d902f71-ab11-40f5-875f-15b69a7a9bc2 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.875985] env[61957]: DEBUG nova.compute.provider_tree [None req-660095f0-9d5a-4797-b621-e0585ffddbba tempest-ServerActionsTestOtherB-303558607 tempest-ServerActionsTestOtherB-303558607-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 759.901325] env[61957]: DEBUG oslo_concurrency.lockutils [None req-09f1eb26-3ef8-431d-92bc-9c62b712a196 tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] Releasing lock "refresh_cache-60ec4271-2614-470d-bf40-47b9955f544c" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 759.901740] env[61957]: DEBUG nova.compute.manager [None req-09f1eb26-3ef8-431d-92bc-9c62b712a196 tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] [instance: 60ec4271-2614-470d-bf40-47b9955f544c] Start destroying the instance on the hypervisor. {{(pid=61957) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 759.901930] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-09f1eb26-3ef8-431d-92bc-9c62b712a196 tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] [instance: 60ec4271-2614-470d-bf40-47b9955f544c] Destroying instance {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 759.902227] env[61957]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a9efe9d7-7c59-463f-8787-67ec1b536553 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.908669] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-3617f05d-8cb2-4824-b3d0-fbcde866b375 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] [instance: 7f3850c0-a455-4e5b-a329-fb3b02f9725f] Powering off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 759.908997] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1665b711-801d-4341-8fb5-54f9ceee9e17 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.913189] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75a0188d-615f-456e-a7b7-96084e142ae7 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.927414] env[61957]: DEBUG oslo_vmware.api [None req-3617f05d-8cb2-4824-b3d0-fbcde866b375 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Waiting for the task: (returnval){ [ 759.927414] env[61957]: value = "task-1277375" [ 759.927414] env[61957]: _type = "Task" [ 759.927414] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 759.936443] env[61957]: WARNING nova.virt.vmwareapi.vmops [None req-09f1eb26-3ef8-431d-92bc-9c62b712a196 tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] [instance: 60ec4271-2614-470d-bf40-47b9955f544c] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 60ec4271-2614-470d-bf40-47b9955f544c could not be found. [ 759.936714] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-09f1eb26-3ef8-431d-92bc-9c62b712a196 tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] [instance: 60ec4271-2614-470d-bf40-47b9955f544c] Instance destroyed {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 759.936893] env[61957]: INFO nova.compute.manager [None req-09f1eb26-3ef8-431d-92bc-9c62b712a196 tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] [instance: 60ec4271-2614-470d-bf40-47b9955f544c] Took 0.03 seconds to destroy the instance on the hypervisor. [ 759.937258] env[61957]: DEBUG oslo.service.loopingcall [None req-09f1eb26-3ef8-431d-92bc-9c62b712a196 tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 759.938041] env[61957]: DEBUG nova.compute.manager [-] [instance: 60ec4271-2614-470d-bf40-47b9955f544c] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 759.938041] env[61957]: DEBUG nova.network.neutron [-] [instance: 60ec4271-2614-470d-bf40-47b9955f544c] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 759.944673] env[61957]: DEBUG oslo_vmware.api [None req-3617f05d-8cb2-4824-b3d0-fbcde866b375 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Task: {'id': task-1277375, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 759.966225] env[61957]: DEBUG nova.network.neutron [-] [instance: 60ec4271-2614-470d-bf40-47b9955f544c] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 760.048170] env[61957]: DEBUG nova.compute.manager [None req-7860e034-eff8-489c-addf-33cf2776246c tempest-ServerActionsTestOtherA-1610268163 tempest-ServerActionsTestOtherA-1610268163-project-member] [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] Starting instance... {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 760.115961] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ac5c4b48-cfd7-420e-9aa1-225385350d11 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Acquiring lock "c23141ee-0cbb-4d1b-8390-c3073fe354f1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 760.116420] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ac5c4b48-cfd7-420e-9aa1-225385350d11 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Lock "c23141ee-0cbb-4d1b-8390-c3073fe354f1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 760.265824] env[61957]: ERROR nova.compute.manager [None req-6b91bce8-d0f8-4c40-855e-ed34f3eb7e64 tempest-ServerActionsTestJSON-1567241756 tempest-ServerActionsTestJSON-1567241756-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 3233fc75-1990-45ba-992a-b932f2d5eafc, please check neutron logs for more information. [ 760.265824] env[61957]: ERROR nova.compute.manager Traceback (most recent call last): [ 760.265824] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 760.265824] env[61957]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 760.265824] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 760.265824] env[61957]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 760.265824] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 760.265824] env[61957]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 760.265824] env[61957]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 760.265824] env[61957]: ERROR nova.compute.manager self.force_reraise() [ 760.265824] env[61957]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 760.265824] env[61957]: ERROR nova.compute.manager raise self.value [ 760.265824] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 760.265824] env[61957]: ERROR nova.compute.manager updated_port = self._update_port( [ 760.265824] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 760.265824] env[61957]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 760.266391] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 760.266391] env[61957]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 760.266391] env[61957]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 3233fc75-1990-45ba-992a-b932f2d5eafc, please check neutron logs for more information. [ 760.266391] env[61957]: ERROR nova.compute.manager [ 760.266391] env[61957]: Traceback (most recent call last): [ 760.266391] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 760.266391] env[61957]: listener.cb(fileno) [ 760.266391] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 760.266391] env[61957]: result = function(*args, **kwargs) [ 760.266391] env[61957]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 760.266391] env[61957]: return func(*args, **kwargs) [ 760.266391] env[61957]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 760.266391] env[61957]: raise e [ 760.266391] env[61957]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 760.266391] env[61957]: nwinfo = self.network_api.allocate_for_instance( [ 760.266391] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 760.266391] env[61957]: created_port_ids = self._update_ports_for_instance( [ 760.266391] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 760.266391] env[61957]: with excutils.save_and_reraise_exception(): [ 760.266391] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 760.266391] env[61957]: self.force_reraise() [ 760.266391] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 760.266391] env[61957]: raise self.value [ 760.266391] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 760.266391] env[61957]: updated_port = self._update_port( [ 760.266391] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 760.266391] env[61957]: _ensure_no_port_binding_failure(port) [ 760.266391] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 760.266391] env[61957]: raise exception.PortBindingFailed(port_id=port['id']) [ 760.267428] env[61957]: nova.exception.PortBindingFailed: Binding failed for port 3233fc75-1990-45ba-992a-b932f2d5eafc, please check neutron logs for more information. [ 760.267428] env[61957]: Removing descriptor: 16 [ 760.382022] env[61957]: DEBUG nova.scheduler.client.report [None req-660095f0-9d5a-4797-b621-e0585ffddbba tempest-ServerActionsTestOtherB-303558607 tempest-ServerActionsTestOtherB-303558607-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 760.438220] env[61957]: DEBUG oslo_vmware.api [None req-3617f05d-8cb2-4824-b3d0-fbcde866b375 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Task: {'id': task-1277375, 'name': PowerOffVM_Task, 'duration_secs': 0.111924} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 760.438577] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-3617f05d-8cb2-4824-b3d0-fbcde866b375 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] [instance: 7f3850c0-a455-4e5b-a329-fb3b02f9725f] Powered off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 760.438880] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-3617f05d-8cb2-4824-b3d0-fbcde866b375 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] [instance: 7f3850c0-a455-4e5b-a329-fb3b02f9725f] Destroying instance {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 760.439744] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-162ec531-5c4e-4996-8069-1d97ae1da1dd {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.447877] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-3617f05d-8cb2-4824-b3d0-fbcde866b375 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] [instance: 7f3850c0-a455-4e5b-a329-fb3b02f9725f] Unregistering the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 760.448252] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d582f45f-436b-442d-a462-15ba3e06d7f9 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.468387] env[61957]: DEBUG nova.network.neutron [-] [instance: 60ec4271-2614-470d-bf40-47b9955f544c] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 760.473386] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-3617f05d-8cb2-4824-b3d0-fbcde866b375 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] [instance: 7f3850c0-a455-4e5b-a329-fb3b02f9725f] Unregistered the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 760.473521] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-3617f05d-8cb2-4824-b3d0-fbcde866b375 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] [instance: 7f3850c0-a455-4e5b-a329-fb3b02f9725f] Deleting contents of the VM from datastore datastore2 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 760.473704] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-3617f05d-8cb2-4824-b3d0-fbcde866b375 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Deleting the datastore file [datastore2] 7f3850c0-a455-4e5b-a329-fb3b02f9725f {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 760.473952] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-508fb84b-115c-43c7-a1bb-9eb48d08bc4e {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.480339] env[61957]: DEBUG oslo_vmware.api [None req-3617f05d-8cb2-4824-b3d0-fbcde866b375 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Waiting for the task: (returnval){ [ 760.480339] env[61957]: value = "task-1277377" [ 760.480339] env[61957]: _type = "Task" [ 760.480339] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 760.481410] env[61957]: DEBUG nova.compute.manager [None req-6b91bce8-d0f8-4c40-855e-ed34f3eb7e64 tempest-ServerActionsTestJSON-1567241756 tempest-ServerActionsTestJSON-1567241756-project-member] [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] Start spawning the instance on the hypervisor. {{(pid=61957) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 760.491615] env[61957]: DEBUG oslo_vmware.api [None req-3617f05d-8cb2-4824-b3d0-fbcde866b375 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Task: {'id': task-1277377, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 760.506614] env[61957]: DEBUG nova.virt.hardware [None req-6b91bce8-d0f8-4c40-855e-ed34f3eb7e64 tempest-ServerActionsTestJSON-1567241756 tempest-ServerActionsTestJSON-1567241756-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T17:22:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T17:21:50Z,direct_url=,disk_format='vmdk',id=11c76a2c-f705-470a-ba9d-4657858bab38,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='81d3096d0f094373913fc3dc8f5c3c6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T17:21:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 760.506857] env[61957]: DEBUG nova.virt.hardware [None req-6b91bce8-d0f8-4c40-855e-ed34f3eb7e64 tempest-ServerActionsTestJSON-1567241756 tempest-ServerActionsTestJSON-1567241756-project-member] Flavor limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 760.507046] env[61957]: DEBUG nova.virt.hardware [None req-6b91bce8-d0f8-4c40-855e-ed34f3eb7e64 tempest-ServerActionsTestJSON-1567241756 tempest-ServerActionsTestJSON-1567241756-project-member] Image limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 760.507250] env[61957]: DEBUG nova.virt.hardware [None req-6b91bce8-d0f8-4c40-855e-ed34f3eb7e64 tempest-ServerActionsTestJSON-1567241756 tempest-ServerActionsTestJSON-1567241756-project-member] Flavor pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 760.507433] env[61957]: DEBUG nova.virt.hardware [None req-6b91bce8-d0f8-4c40-855e-ed34f3eb7e64 tempest-ServerActionsTestJSON-1567241756 tempest-ServerActionsTestJSON-1567241756-project-member] Image pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 760.507603] env[61957]: DEBUG nova.virt.hardware [None req-6b91bce8-d0f8-4c40-855e-ed34f3eb7e64 tempest-ServerActionsTestJSON-1567241756 tempest-ServerActionsTestJSON-1567241756-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 760.507826] env[61957]: DEBUG nova.virt.hardware [None req-6b91bce8-d0f8-4c40-855e-ed34f3eb7e64 tempest-ServerActionsTestJSON-1567241756 tempest-ServerActionsTestJSON-1567241756-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 760.507993] env[61957]: DEBUG nova.virt.hardware [None req-6b91bce8-d0f8-4c40-855e-ed34f3eb7e64 tempest-ServerActionsTestJSON-1567241756 tempest-ServerActionsTestJSON-1567241756-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 760.508185] env[61957]: DEBUG nova.virt.hardware [None req-6b91bce8-d0f8-4c40-855e-ed34f3eb7e64 tempest-ServerActionsTestJSON-1567241756 tempest-ServerActionsTestJSON-1567241756-project-member] Got 1 possible topologies {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 760.508349] env[61957]: DEBUG nova.virt.hardware [None req-6b91bce8-d0f8-4c40-855e-ed34f3eb7e64 tempest-ServerActionsTestJSON-1567241756 tempest-ServerActionsTestJSON-1567241756-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 760.508515] env[61957]: DEBUG nova.virt.hardware [None req-6b91bce8-d0f8-4c40-855e-ed34f3eb7e64 tempest-ServerActionsTestJSON-1567241756 tempest-ServerActionsTestJSON-1567241756-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 760.509366] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4b025c2-5da1-4e59-a8dc-f4df4dfcabba {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.516813] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71dd906d-456d-48b7-a7ad-169d3095c654 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.529955] env[61957]: ERROR nova.compute.manager [None req-6b91bce8-d0f8-4c40-855e-ed34f3eb7e64 tempest-ServerActionsTestJSON-1567241756 tempest-ServerActionsTestJSON-1567241756-project-member] [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 3233fc75-1990-45ba-992a-b932f2d5eafc, please check neutron logs for more information. [ 760.529955] env[61957]: ERROR nova.compute.manager [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] Traceback (most recent call last): [ 760.529955] env[61957]: ERROR nova.compute.manager [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 760.529955] env[61957]: ERROR nova.compute.manager [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] yield resources [ 760.529955] env[61957]: ERROR nova.compute.manager [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 760.529955] env[61957]: ERROR nova.compute.manager [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] self.driver.spawn(context, instance, image_meta, [ 760.529955] env[61957]: ERROR nova.compute.manager [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 760.529955] env[61957]: ERROR nova.compute.manager [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 760.529955] env[61957]: ERROR nova.compute.manager [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 760.529955] env[61957]: ERROR nova.compute.manager [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] vm_ref = self.build_virtual_machine(instance, [ 760.529955] env[61957]: ERROR nova.compute.manager [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 760.530464] env[61957]: ERROR nova.compute.manager [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] vif_infos = vmwarevif.get_vif_info(self._session, [ 760.530464] env[61957]: ERROR nova.compute.manager [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 760.530464] env[61957]: ERROR nova.compute.manager [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] for vif in network_info: [ 760.530464] env[61957]: ERROR nova.compute.manager [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 760.530464] env[61957]: ERROR nova.compute.manager [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] return self._sync_wrapper(fn, *args, **kwargs) [ 760.530464] env[61957]: ERROR nova.compute.manager [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 760.530464] env[61957]: ERROR nova.compute.manager [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] self.wait() [ 760.530464] env[61957]: ERROR nova.compute.manager [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 760.530464] env[61957]: ERROR nova.compute.manager [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] self[:] = self._gt.wait() [ 760.530464] env[61957]: ERROR nova.compute.manager [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 760.530464] env[61957]: ERROR nova.compute.manager [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] return self._exit_event.wait() [ 760.530464] env[61957]: ERROR nova.compute.manager [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 760.530464] env[61957]: ERROR nova.compute.manager [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] current.throw(*self._exc) [ 760.530966] env[61957]: ERROR nova.compute.manager [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 760.530966] env[61957]: ERROR nova.compute.manager [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] result = function(*args, **kwargs) [ 760.530966] env[61957]: ERROR nova.compute.manager [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 760.530966] env[61957]: ERROR nova.compute.manager [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] return func(*args, **kwargs) [ 760.530966] env[61957]: ERROR nova.compute.manager [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 760.530966] env[61957]: ERROR nova.compute.manager [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] raise e [ 760.530966] env[61957]: ERROR nova.compute.manager [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 760.530966] env[61957]: ERROR nova.compute.manager [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] nwinfo = self.network_api.allocate_for_instance( [ 760.530966] env[61957]: ERROR nova.compute.manager [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 760.530966] env[61957]: ERROR nova.compute.manager [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] created_port_ids = self._update_ports_for_instance( [ 760.530966] env[61957]: ERROR nova.compute.manager [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 760.530966] env[61957]: ERROR nova.compute.manager [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] with excutils.save_and_reraise_exception(): [ 760.530966] env[61957]: ERROR nova.compute.manager [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 760.531490] env[61957]: ERROR nova.compute.manager [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] self.force_reraise() [ 760.531490] env[61957]: ERROR nova.compute.manager [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 760.531490] env[61957]: ERROR nova.compute.manager [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] raise self.value [ 760.531490] env[61957]: ERROR nova.compute.manager [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 760.531490] env[61957]: ERROR nova.compute.manager [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] updated_port = self._update_port( [ 760.531490] env[61957]: ERROR nova.compute.manager [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 760.531490] env[61957]: ERROR nova.compute.manager [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] _ensure_no_port_binding_failure(port) [ 760.531490] env[61957]: ERROR nova.compute.manager [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 760.531490] env[61957]: ERROR nova.compute.manager [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] raise exception.PortBindingFailed(port_id=port['id']) [ 760.531490] env[61957]: ERROR nova.compute.manager [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] nova.exception.PortBindingFailed: Binding failed for port 3233fc75-1990-45ba-992a-b932f2d5eafc, please check neutron logs for more information. [ 760.531490] env[61957]: ERROR nova.compute.manager [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] [ 760.531490] env[61957]: INFO nova.compute.manager [None req-6b91bce8-d0f8-4c40-855e-ed34f3eb7e64 tempest-ServerActionsTestJSON-1567241756 tempest-ServerActionsTestJSON-1567241756-project-member] [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] Terminating instance [ 760.532301] env[61957]: DEBUG oslo_concurrency.lockutils [None req-6b91bce8-d0f8-4c40-855e-ed34f3eb7e64 tempest-ServerActionsTestJSON-1567241756 tempest-ServerActionsTestJSON-1567241756-project-member] Acquiring lock "refresh_cache-e69a7833-97b1-4063-a6e8-c4558980ee1e" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 760.532453] env[61957]: DEBUG oslo_concurrency.lockutils [None req-6b91bce8-d0f8-4c40-855e-ed34f3eb7e64 tempest-ServerActionsTestJSON-1567241756 tempest-ServerActionsTestJSON-1567241756-project-member] Acquired lock "refresh_cache-e69a7833-97b1-4063-a6e8-c4558980ee1e" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 760.532615] env[61957]: DEBUG nova.network.neutron [None req-6b91bce8-d0f8-4c40-855e-ed34f3eb7e64 tempest-ServerActionsTestJSON-1567241756 tempest-ServerActionsTestJSON-1567241756-project-member] [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 760.572255] env[61957]: DEBUG oslo_concurrency.lockutils [None req-7860e034-eff8-489c-addf-33cf2776246c tempest-ServerActionsTestOtherA-1610268163 tempest-ServerActionsTestOtherA-1610268163-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 760.885992] env[61957]: DEBUG oslo_concurrency.lockutils [None req-660095f0-9d5a-4797-b621-e0585ffddbba tempest-ServerActionsTestOtherB-303558607 tempest-ServerActionsTestOtherB-303558607-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.430s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 760.886348] env[61957]: DEBUG nova.compute.manager [None req-660095f0-9d5a-4797-b621-e0585ffddbba tempest-ServerActionsTestOtherB-303558607 tempest-ServerActionsTestOtherB-303558607-project-member] [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] Start building networks asynchronously for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 760.889107] env[61957]: DEBUG oslo_concurrency.lockutils [None req-3d8587e3-7ad5-40a6-88af-e8ed44c49baa tempest-ServerActionsV293TestJSON-1269989829 tempest-ServerActionsV293TestJSON-1269989829-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.514s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 760.890554] env[61957]: INFO nova.compute.claims [None req-3d8587e3-7ad5-40a6-88af-e8ed44c49baa tempest-ServerActionsV293TestJSON-1269989829 tempest-ServerActionsV293TestJSON-1269989829-project-member] [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 760.971081] env[61957]: INFO nova.compute.manager [-] [instance: 60ec4271-2614-470d-bf40-47b9955f544c] Took 1.03 seconds to deallocate network for instance. [ 760.973468] env[61957]: DEBUG nova.compute.claims [None req-09f1eb26-3ef8-431d-92bc-9c62b712a196 tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] [instance: 60ec4271-2614-470d-bf40-47b9955f544c] Aborting claim: {{(pid=61957) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 760.973468] env[61957]: DEBUG oslo_concurrency.lockutils [None req-09f1eb26-3ef8-431d-92bc-9c62b712a196 tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 760.992777] env[61957]: DEBUG oslo_vmware.api [None req-3617f05d-8cb2-4824-b3d0-fbcde866b375 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Task: {'id': task-1277377, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.116936} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 760.993285] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-3617f05d-8cb2-4824-b3d0-fbcde866b375 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Deleted the datastore file {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 760.993472] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-3617f05d-8cb2-4824-b3d0-fbcde866b375 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] [instance: 7f3850c0-a455-4e5b-a329-fb3b02f9725f] Deleted contents of the VM from datastore datastore2 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 760.993646] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-3617f05d-8cb2-4824-b3d0-fbcde866b375 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] [instance: 7f3850c0-a455-4e5b-a329-fb3b02f9725f] Instance destroyed {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 761.050569] env[61957]: DEBUG nova.network.neutron [None req-6b91bce8-d0f8-4c40-855e-ed34f3eb7e64 tempest-ServerActionsTestJSON-1567241756 tempest-ServerActionsTestJSON-1567241756-project-member] [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 761.133178] env[61957]: DEBUG nova.network.neutron [None req-6b91bce8-d0f8-4c40-855e-ed34f3eb7e64 tempest-ServerActionsTestJSON-1567241756 tempest-ServerActionsTestJSON-1567241756-project-member] [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 761.394660] env[61957]: DEBUG nova.compute.utils [None req-660095f0-9d5a-4797-b621-e0585ffddbba tempest-ServerActionsTestOtherB-303558607 tempest-ServerActionsTestOtherB-303558607-project-member] Using /dev/sd instead of None {{(pid=61957) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 761.399729] env[61957]: DEBUG nova.compute.manager [None req-660095f0-9d5a-4797-b621-e0585ffddbba tempest-ServerActionsTestOtherB-303558607 tempest-ServerActionsTestOtherB-303558607-project-member] [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] Allocating IP information in the background. {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 761.399729] env[61957]: DEBUG nova.network.neutron [None req-660095f0-9d5a-4797-b621-e0585ffddbba tempest-ServerActionsTestOtherB-303558607 tempest-ServerActionsTestOtherB-303558607-project-member] [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] allocate_for_instance() {{(pid=61957) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 761.441323] env[61957]: DEBUG nova.policy [None req-660095f0-9d5a-4797-b621-e0585ffddbba tempest-ServerActionsTestOtherB-303558607 tempest-ServerActionsTestOtherB-303558607-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4e8a657d2a204bf5b01f3ab880f76988', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '465de806e1714d739b9cf8eab184b82d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61957) authorize /opt/stack/nova/nova/policy.py:203}} [ 761.638061] env[61957]: DEBUG oslo_concurrency.lockutils [None req-6b91bce8-d0f8-4c40-855e-ed34f3eb7e64 tempest-ServerActionsTestJSON-1567241756 tempest-ServerActionsTestJSON-1567241756-project-member] Releasing lock "refresh_cache-e69a7833-97b1-4063-a6e8-c4558980ee1e" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 761.638061] env[61957]: DEBUG nova.compute.manager [None req-6b91bce8-d0f8-4c40-855e-ed34f3eb7e64 tempest-ServerActionsTestJSON-1567241756 tempest-ServerActionsTestJSON-1567241756-project-member] [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] Start destroying the instance on the hypervisor. {{(pid=61957) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 761.638061] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-6b91bce8-d0f8-4c40-855e-ed34f3eb7e64 tempest-ServerActionsTestJSON-1567241756 tempest-ServerActionsTestJSON-1567241756-project-member] [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] Destroying instance {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 761.638061] env[61957]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-11980d08-200b-413f-8576-3a32d47d09b6 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.648391] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0723cf22-0637-4b93-88c2-8095812abb96 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.669017] env[61957]: WARNING nova.virt.vmwareapi.vmops [None req-6b91bce8-d0f8-4c40-855e-ed34f3eb7e64 tempest-ServerActionsTestJSON-1567241756 tempest-ServerActionsTestJSON-1567241756-project-member] [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance e69a7833-97b1-4063-a6e8-c4558980ee1e could not be found. [ 761.669276] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-6b91bce8-d0f8-4c40-855e-ed34f3eb7e64 tempest-ServerActionsTestJSON-1567241756 tempest-ServerActionsTestJSON-1567241756-project-member] [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] Instance destroyed {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 761.669484] env[61957]: INFO nova.compute.manager [None req-6b91bce8-d0f8-4c40-855e-ed34f3eb7e64 tempest-ServerActionsTestJSON-1567241756 tempest-ServerActionsTestJSON-1567241756-project-member] [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] Took 0.03 seconds to destroy the instance on the hypervisor. [ 761.669720] env[61957]: DEBUG oslo.service.loopingcall [None req-6b91bce8-d0f8-4c40-855e-ed34f3eb7e64 tempest-ServerActionsTestJSON-1567241756 tempest-ServerActionsTestJSON-1567241756-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 761.669960] env[61957]: DEBUG nova.compute.manager [-] [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 761.670083] env[61957]: DEBUG nova.network.neutron [-] [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 761.687006] env[61957]: DEBUG nova.network.neutron [-] [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 761.745856] env[61957]: DEBUG nova.network.neutron [None req-660095f0-9d5a-4797-b621-e0585ffddbba tempest-ServerActionsTestOtherB-303558607 tempest-ServerActionsTestOtherB-303558607-project-member] [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] Successfully created port: 5d6fb25b-5ed5-4828-acbe-5d639e8419fa {{(pid=61957) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 761.820749] env[61957]: DEBUG nova.compute.manager [req-e848bb31-2ae8-4232-818b-598c747fa640 req-0e5af245-95f6-4330-aa6b-c39e7bb86a43 service nova] [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] Received event network-changed-3233fc75-1990-45ba-992a-b932f2d5eafc {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 761.820939] env[61957]: DEBUG nova.compute.manager [req-e848bb31-2ae8-4232-818b-598c747fa640 req-0e5af245-95f6-4330-aa6b-c39e7bb86a43 service nova] [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] Refreshing instance network info cache due to event network-changed-3233fc75-1990-45ba-992a-b932f2d5eafc. {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 761.821164] env[61957]: DEBUG oslo_concurrency.lockutils [req-e848bb31-2ae8-4232-818b-598c747fa640 req-0e5af245-95f6-4330-aa6b-c39e7bb86a43 service nova] Acquiring lock "refresh_cache-e69a7833-97b1-4063-a6e8-c4558980ee1e" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 761.821352] env[61957]: DEBUG oslo_concurrency.lockutils [req-e848bb31-2ae8-4232-818b-598c747fa640 req-0e5af245-95f6-4330-aa6b-c39e7bb86a43 service nova] Acquired lock "refresh_cache-e69a7833-97b1-4063-a6e8-c4558980ee1e" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 761.821453] env[61957]: DEBUG nova.network.neutron [req-e848bb31-2ae8-4232-818b-598c747fa640 req-0e5af245-95f6-4330-aa6b-c39e7bb86a43 service nova] [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] Refreshing network info cache for port 3233fc75-1990-45ba-992a-b932f2d5eafc {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 761.901452] env[61957]: DEBUG nova.compute.manager [None req-660095f0-9d5a-4797-b621-e0585ffddbba tempest-ServerActionsTestOtherB-303558607 tempest-ServerActionsTestOtherB-303558607-project-member] [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] Start building block device mappings for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 762.036042] env[61957]: DEBUG nova.virt.hardware [None req-3617f05d-8cb2-4824-b3d0-fbcde866b375 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T17:22:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T17:21:50Z,direct_url=,disk_format='vmdk',id=11c76a2c-f705-470a-ba9d-4657858bab38,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='81d3096d0f094373913fc3dc8f5c3c6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T17:21:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 762.036305] env[61957]: DEBUG nova.virt.hardware [None req-3617f05d-8cb2-4824-b3d0-fbcde866b375 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Flavor limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 762.036460] env[61957]: DEBUG nova.virt.hardware [None req-3617f05d-8cb2-4824-b3d0-fbcde866b375 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Image limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 762.036639] env[61957]: DEBUG nova.virt.hardware [None req-3617f05d-8cb2-4824-b3d0-fbcde866b375 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Flavor pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 762.036783] env[61957]: DEBUG nova.virt.hardware [None req-3617f05d-8cb2-4824-b3d0-fbcde866b375 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Image pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 762.036924] env[61957]: DEBUG nova.virt.hardware [None req-3617f05d-8cb2-4824-b3d0-fbcde866b375 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 762.037141] env[61957]: DEBUG nova.virt.hardware [None req-3617f05d-8cb2-4824-b3d0-fbcde866b375 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 762.037297] env[61957]: DEBUG nova.virt.hardware [None req-3617f05d-8cb2-4824-b3d0-fbcde866b375 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 762.037458] env[61957]: DEBUG nova.virt.hardware [None req-3617f05d-8cb2-4824-b3d0-fbcde866b375 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Got 1 possible topologies {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 762.037612] env[61957]: DEBUG nova.virt.hardware [None req-3617f05d-8cb2-4824-b3d0-fbcde866b375 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 762.037777] env[61957]: DEBUG nova.virt.hardware [None req-3617f05d-8cb2-4824-b3d0-fbcde866b375 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 762.038643] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b314a0d9-8540-44ec-af12-f5d133eb0fd1 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.047932] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a4452f4-c747-4a23-9762-1d100a65a6ef {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.078253] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-3617f05d-8cb2-4824-b3d0-fbcde866b375 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] [instance: 7f3850c0-a455-4e5b-a329-fb3b02f9725f] Instance VIF info [] {{(pid=61957) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 762.080626] env[61957]: DEBUG oslo.service.loopingcall [None req-3617f05d-8cb2-4824-b3d0-fbcde866b375 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 762.083669] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7f3850c0-a455-4e5b-a329-fb3b02f9725f] Creating VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 762.084443] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c97ab5bf-4b74-44ee-9852-67a7a34a3551 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.104222] env[61957]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 762.104222] env[61957]: value = "task-1277378" [ 762.104222] env[61957]: _type = "Task" [ 762.104222] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 762.111785] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277378, 'name': CreateVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.192373] env[61957]: DEBUG nova.network.neutron [-] [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 762.275894] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-549dc3ca-da53-4ad4-ad55-f95db6c648c4 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.283713] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-898fa780-9c68-48fa-a051-cb9adff86a64 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.313776] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfa37792-9dd9-4de1-86b9-c2a9981f5441 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.322717] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b1573f9-fa6b-460c-9304-ea582b0f28fb {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.342281] env[61957]: DEBUG nova.compute.provider_tree [None req-3d8587e3-7ad5-40a6-88af-e8ed44c49baa tempest-ServerActionsV293TestJSON-1269989829 tempest-ServerActionsV293TestJSON-1269989829-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 762.344714] env[61957]: DEBUG nova.network.neutron [req-e848bb31-2ae8-4232-818b-598c747fa640 req-0e5af245-95f6-4330-aa6b-c39e7bb86a43 service nova] [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 762.467188] env[61957]: DEBUG nova.network.neutron [req-e848bb31-2ae8-4232-818b-598c747fa640 req-0e5af245-95f6-4330-aa6b-c39e7bb86a43 service nova] [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 762.617203] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277378, 'name': CreateVM_Task, 'duration_secs': 0.31161} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 762.617378] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7f3850c0-a455-4e5b-a329-fb3b02f9725f] Created VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 762.617784] env[61957]: DEBUG oslo_concurrency.lockutils [None req-3617f05d-8cb2-4824-b3d0-fbcde866b375 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 762.617950] env[61957]: DEBUG oslo_concurrency.lockutils [None req-3617f05d-8cb2-4824-b3d0-fbcde866b375 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 762.618295] env[61957]: DEBUG oslo_concurrency.lockutils [None req-3617f05d-8cb2-4824-b3d0-fbcde866b375 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 762.618527] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e428009b-03b1-483e-a898-74ca9eefd2bc {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.623481] env[61957]: DEBUG oslo_vmware.api [None req-3617f05d-8cb2-4824-b3d0-fbcde866b375 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Waiting for the task: (returnval){ [ 762.623481] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]5240c58b-9153-77e0-4717-5213a047de6b" [ 762.623481] env[61957]: _type = "Task" [ 762.623481] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 762.631134] env[61957]: DEBUG oslo_vmware.api [None req-3617f05d-8cb2-4824-b3d0-fbcde866b375 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]5240c58b-9153-77e0-4717-5213a047de6b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.696930] env[61957]: INFO nova.compute.manager [-] [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] Took 1.03 seconds to deallocate network for instance. [ 762.700938] env[61957]: DEBUG nova.compute.claims [None req-6b91bce8-d0f8-4c40-855e-ed34f3eb7e64 tempest-ServerActionsTestJSON-1567241756 tempest-ServerActionsTestJSON-1567241756-project-member] [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] Aborting claim: {{(pid=61957) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 762.700938] env[61957]: DEBUG oslo_concurrency.lockutils [None req-6b91bce8-d0f8-4c40-855e-ed34f3eb7e64 tempest-ServerActionsTestJSON-1567241756 tempest-ServerActionsTestJSON-1567241756-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 762.815652] env[61957]: ERROR nova.compute.manager [None req-660095f0-9d5a-4797-b621-e0585ffddbba tempest-ServerActionsTestOtherB-303558607 tempest-ServerActionsTestOtherB-303558607-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 5d6fb25b-5ed5-4828-acbe-5d639e8419fa, please check neutron logs for more information. [ 762.815652] env[61957]: ERROR nova.compute.manager Traceback (most recent call last): [ 762.815652] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 762.815652] env[61957]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 762.815652] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 762.815652] env[61957]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 762.815652] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 762.815652] env[61957]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 762.815652] env[61957]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 762.815652] env[61957]: ERROR nova.compute.manager self.force_reraise() [ 762.815652] env[61957]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 762.815652] env[61957]: ERROR nova.compute.manager raise self.value [ 762.815652] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 762.815652] env[61957]: ERROR nova.compute.manager updated_port = self._update_port( [ 762.815652] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 762.815652] env[61957]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 762.816219] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 762.816219] env[61957]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 762.816219] env[61957]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 5d6fb25b-5ed5-4828-acbe-5d639e8419fa, please check neutron logs for more information. [ 762.816219] env[61957]: ERROR nova.compute.manager [ 762.816219] env[61957]: Traceback (most recent call last): [ 762.816219] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 762.816219] env[61957]: listener.cb(fileno) [ 762.816219] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 762.816219] env[61957]: result = function(*args, **kwargs) [ 762.816219] env[61957]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 762.816219] env[61957]: return func(*args, **kwargs) [ 762.816219] env[61957]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 762.816219] env[61957]: raise e [ 762.816219] env[61957]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 762.816219] env[61957]: nwinfo = self.network_api.allocate_for_instance( [ 762.816219] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 762.816219] env[61957]: created_port_ids = self._update_ports_for_instance( [ 762.816219] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 762.816219] env[61957]: with excutils.save_and_reraise_exception(): [ 762.816219] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 762.816219] env[61957]: self.force_reraise() [ 762.816219] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 762.816219] env[61957]: raise self.value [ 762.816219] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 762.816219] env[61957]: updated_port = self._update_port( [ 762.816219] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 762.816219] env[61957]: _ensure_no_port_binding_failure(port) [ 762.816219] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 762.816219] env[61957]: raise exception.PortBindingFailed(port_id=port['id']) [ 762.817190] env[61957]: nova.exception.PortBindingFailed: Binding failed for port 5d6fb25b-5ed5-4828-acbe-5d639e8419fa, please check neutron logs for more information. [ 762.817190] env[61957]: Removing descriptor: 16 [ 762.847465] env[61957]: DEBUG nova.scheduler.client.report [None req-3d8587e3-7ad5-40a6-88af-e8ed44c49baa tempest-ServerActionsV293TestJSON-1269989829 tempest-ServerActionsV293TestJSON-1269989829-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 762.917850] env[61957]: DEBUG nova.compute.manager [None req-660095f0-9d5a-4797-b621-e0585ffddbba tempest-ServerActionsTestOtherB-303558607 tempest-ServerActionsTestOtherB-303558607-project-member] [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] Start spawning the instance on the hypervisor. {{(pid=61957) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 762.945902] env[61957]: DEBUG nova.virt.hardware [None req-660095f0-9d5a-4797-b621-e0585ffddbba tempest-ServerActionsTestOtherB-303558607 tempest-ServerActionsTestOtherB-303558607-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T17:22:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T17:21:50Z,direct_url=,disk_format='vmdk',id=11c76a2c-f705-470a-ba9d-4657858bab38,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='81d3096d0f094373913fc3dc8f5c3c6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T17:21:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 762.946217] env[61957]: DEBUG nova.virt.hardware [None req-660095f0-9d5a-4797-b621-e0585ffddbba tempest-ServerActionsTestOtherB-303558607 tempest-ServerActionsTestOtherB-303558607-project-member] Flavor limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 762.946406] env[61957]: DEBUG nova.virt.hardware [None req-660095f0-9d5a-4797-b621-e0585ffddbba tempest-ServerActionsTestOtherB-303558607 tempest-ServerActionsTestOtherB-303558607-project-member] Image limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 762.946620] env[61957]: DEBUG nova.virt.hardware [None req-660095f0-9d5a-4797-b621-e0585ffddbba tempest-ServerActionsTestOtherB-303558607 tempest-ServerActionsTestOtherB-303558607-project-member] Flavor pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 762.946796] env[61957]: DEBUG nova.virt.hardware [None req-660095f0-9d5a-4797-b621-e0585ffddbba tempest-ServerActionsTestOtherB-303558607 tempest-ServerActionsTestOtherB-303558607-project-member] Image pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 762.946973] env[61957]: DEBUG nova.virt.hardware [None req-660095f0-9d5a-4797-b621-e0585ffddbba tempest-ServerActionsTestOtherB-303558607 tempest-ServerActionsTestOtherB-303558607-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 762.947250] env[61957]: DEBUG nova.virt.hardware [None req-660095f0-9d5a-4797-b621-e0585ffddbba tempest-ServerActionsTestOtherB-303558607 tempest-ServerActionsTestOtherB-303558607-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 762.947416] env[61957]: DEBUG nova.virt.hardware [None req-660095f0-9d5a-4797-b621-e0585ffddbba tempest-ServerActionsTestOtherB-303558607 tempest-ServerActionsTestOtherB-303558607-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 762.947588] env[61957]: DEBUG nova.virt.hardware [None req-660095f0-9d5a-4797-b621-e0585ffddbba tempest-ServerActionsTestOtherB-303558607 tempest-ServerActionsTestOtherB-303558607-project-member] Got 1 possible topologies {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 762.947744] env[61957]: DEBUG nova.virt.hardware [None req-660095f0-9d5a-4797-b621-e0585ffddbba tempest-ServerActionsTestOtherB-303558607 tempest-ServerActionsTestOtherB-303558607-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 762.947987] env[61957]: DEBUG nova.virt.hardware [None req-660095f0-9d5a-4797-b621-e0585ffddbba tempest-ServerActionsTestOtherB-303558607 tempest-ServerActionsTestOtherB-303558607-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 762.948778] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b98d4b2-d332-4131-aefe-90e68050c03e {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.956751] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ff01c74-4d3b-41ea-89af-ab3b05260ae6 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.969750] env[61957]: DEBUG oslo_concurrency.lockutils [req-e848bb31-2ae8-4232-818b-598c747fa640 req-0e5af245-95f6-4330-aa6b-c39e7bb86a43 service nova] Releasing lock "refresh_cache-e69a7833-97b1-4063-a6e8-c4558980ee1e" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 762.969962] env[61957]: DEBUG nova.compute.manager [req-e848bb31-2ae8-4232-818b-598c747fa640 req-0e5af245-95f6-4330-aa6b-c39e7bb86a43 service nova] [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] Received event network-vif-deleted-3233fc75-1990-45ba-992a-b932f2d5eafc {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 762.970847] env[61957]: ERROR nova.compute.manager [None req-660095f0-9d5a-4797-b621-e0585ffddbba tempest-ServerActionsTestOtherB-303558607 tempest-ServerActionsTestOtherB-303558607-project-member] [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 5d6fb25b-5ed5-4828-acbe-5d639e8419fa, please check neutron logs for more information. [ 762.970847] env[61957]: ERROR nova.compute.manager [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] Traceback (most recent call last): [ 762.970847] env[61957]: ERROR nova.compute.manager [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 762.970847] env[61957]: ERROR nova.compute.manager [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] yield resources [ 762.970847] env[61957]: ERROR nova.compute.manager [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 762.970847] env[61957]: ERROR nova.compute.manager [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] self.driver.spawn(context, instance, image_meta, [ 762.970847] env[61957]: ERROR nova.compute.manager [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 762.970847] env[61957]: ERROR nova.compute.manager [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] self._vmops.spawn(context, instance, image_meta, injected_files, [ 762.970847] env[61957]: ERROR nova.compute.manager [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 762.970847] env[61957]: ERROR nova.compute.manager [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] vm_ref = self.build_virtual_machine(instance, [ 762.970847] env[61957]: ERROR nova.compute.manager [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 762.971306] env[61957]: ERROR nova.compute.manager [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] vif_infos = vmwarevif.get_vif_info(self._session, [ 762.971306] env[61957]: ERROR nova.compute.manager [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 762.971306] env[61957]: ERROR nova.compute.manager [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] for vif in network_info: [ 762.971306] env[61957]: ERROR nova.compute.manager [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 762.971306] env[61957]: ERROR nova.compute.manager [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] return self._sync_wrapper(fn, *args, **kwargs) [ 762.971306] env[61957]: ERROR nova.compute.manager [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 762.971306] env[61957]: ERROR nova.compute.manager [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] self.wait() [ 762.971306] env[61957]: ERROR nova.compute.manager [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 762.971306] env[61957]: ERROR nova.compute.manager [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] self[:] = self._gt.wait() [ 762.971306] env[61957]: ERROR nova.compute.manager [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 762.971306] env[61957]: ERROR nova.compute.manager [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] return self._exit_event.wait() [ 762.971306] env[61957]: ERROR nova.compute.manager [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 762.971306] env[61957]: ERROR nova.compute.manager [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] current.throw(*self._exc) [ 762.971745] env[61957]: ERROR nova.compute.manager [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 762.971745] env[61957]: ERROR nova.compute.manager [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] result = function(*args, **kwargs) [ 762.971745] env[61957]: ERROR nova.compute.manager [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 762.971745] env[61957]: ERROR nova.compute.manager [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] return func(*args, **kwargs) [ 762.971745] env[61957]: ERROR nova.compute.manager [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 762.971745] env[61957]: ERROR nova.compute.manager [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] raise e [ 762.971745] env[61957]: ERROR nova.compute.manager [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 762.971745] env[61957]: ERROR nova.compute.manager [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] nwinfo = self.network_api.allocate_for_instance( [ 762.971745] env[61957]: ERROR nova.compute.manager [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 762.971745] env[61957]: ERROR nova.compute.manager [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] created_port_ids = self._update_ports_for_instance( [ 762.971745] env[61957]: ERROR nova.compute.manager [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 762.971745] env[61957]: ERROR nova.compute.manager [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] with excutils.save_and_reraise_exception(): [ 762.971745] env[61957]: ERROR nova.compute.manager [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 762.972132] env[61957]: ERROR nova.compute.manager [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] self.force_reraise() [ 762.972132] env[61957]: ERROR nova.compute.manager [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 762.972132] env[61957]: ERROR nova.compute.manager [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] raise self.value [ 762.972132] env[61957]: ERROR nova.compute.manager [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 762.972132] env[61957]: ERROR nova.compute.manager [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] updated_port = self._update_port( [ 762.972132] env[61957]: ERROR nova.compute.manager [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 762.972132] env[61957]: ERROR nova.compute.manager [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] _ensure_no_port_binding_failure(port) [ 762.972132] env[61957]: ERROR nova.compute.manager [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 762.972132] env[61957]: ERROR nova.compute.manager [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] raise exception.PortBindingFailed(port_id=port['id']) [ 762.972132] env[61957]: ERROR nova.compute.manager [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] nova.exception.PortBindingFailed: Binding failed for port 5d6fb25b-5ed5-4828-acbe-5d639e8419fa, please check neutron logs for more information. [ 762.972132] env[61957]: ERROR nova.compute.manager [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] [ 762.972132] env[61957]: INFO nova.compute.manager [None req-660095f0-9d5a-4797-b621-e0585ffddbba tempest-ServerActionsTestOtherB-303558607 tempest-ServerActionsTestOtherB-303558607-project-member] [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] Terminating instance [ 762.973425] env[61957]: DEBUG oslo_concurrency.lockutils [None req-660095f0-9d5a-4797-b621-e0585ffddbba tempest-ServerActionsTestOtherB-303558607 tempest-ServerActionsTestOtherB-303558607-project-member] Acquiring lock "refresh_cache-d77276ba-6d37-440b-a113-c4f62b2e7946" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 762.973506] env[61957]: DEBUG oslo_concurrency.lockutils [None req-660095f0-9d5a-4797-b621-e0585ffddbba tempest-ServerActionsTestOtherB-303558607 tempest-ServerActionsTestOtherB-303558607-project-member] Acquired lock "refresh_cache-d77276ba-6d37-440b-a113-c4f62b2e7946" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 762.973665] env[61957]: DEBUG nova.network.neutron [None req-660095f0-9d5a-4797-b621-e0585ffddbba tempest-ServerActionsTestOtherB-303558607 tempest-ServerActionsTestOtherB-303558607-project-member] [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 763.133201] env[61957]: DEBUG oslo_vmware.api [None req-3617f05d-8cb2-4824-b3d0-fbcde866b375 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]5240c58b-9153-77e0-4717-5213a047de6b, 'name': SearchDatastore_Task, 'duration_secs': 0.009084} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 763.133505] env[61957]: DEBUG oslo_concurrency.lockutils [None req-3617f05d-8cb2-4824-b3d0-fbcde866b375 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 763.133738] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-3617f05d-8cb2-4824-b3d0-fbcde866b375 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] [instance: 7f3850c0-a455-4e5b-a329-fb3b02f9725f] Processing image 11c76a2c-f705-470a-ba9d-4657858bab38 {{(pid=61957) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 763.133959] env[61957]: DEBUG oslo_concurrency.lockutils [None req-3617f05d-8cb2-4824-b3d0-fbcde866b375 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 763.134120] env[61957]: DEBUG oslo_concurrency.lockutils [None req-3617f05d-8cb2-4824-b3d0-fbcde866b375 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 763.134293] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-3617f05d-8cb2-4824-b3d0-fbcde866b375 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 763.134547] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c109bfc0-91fc-4e3c-8fd2-886b81934496 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.142754] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-3617f05d-8cb2-4824-b3d0-fbcde866b375 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 763.142938] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-3617f05d-8cb2-4824-b3d0-fbcde866b375 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61957) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 763.143660] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b835cdb4-90bc-4f51-a53e-3649a3f93b6a {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.148686] env[61957]: DEBUG oslo_vmware.api [None req-3617f05d-8cb2-4824-b3d0-fbcde866b375 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Waiting for the task: (returnval){ [ 763.148686] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]523c1157-2c59-00a5-233b-842be608dda1" [ 763.148686] env[61957]: _type = "Task" [ 763.148686] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 763.156526] env[61957]: DEBUG oslo_vmware.api [None req-3617f05d-8cb2-4824-b3d0-fbcde866b375 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]523c1157-2c59-00a5-233b-842be608dda1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 763.352057] env[61957]: DEBUG oslo_concurrency.lockutils [None req-3d8587e3-7ad5-40a6-88af-e8ed44c49baa tempest-ServerActionsV293TestJSON-1269989829 tempest-ServerActionsV293TestJSON-1269989829-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.463s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 763.352490] env[61957]: DEBUG nova.compute.manager [None req-3d8587e3-7ad5-40a6-88af-e8ed44c49baa tempest-ServerActionsV293TestJSON-1269989829 tempest-ServerActionsV293TestJSON-1269989829-project-member] [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] Start building networks asynchronously for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 763.355996] env[61957]: DEBUG oslo_concurrency.lockutils [None req-5ca5fdda-dbef-4c48-b38f-9d2aa78e9841 tempest-FloatingIPsAssociationNegativeTestJSON-572957506 tempest-FloatingIPsAssociationNegativeTestJSON-572957506-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.196s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 763.492057] env[61957]: DEBUG nova.network.neutron [None req-660095f0-9d5a-4797-b621-e0585ffddbba tempest-ServerActionsTestOtherB-303558607 tempest-ServerActionsTestOtherB-303558607-project-member] [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 763.573624] env[61957]: DEBUG nova.network.neutron [None req-660095f0-9d5a-4797-b621-e0585ffddbba tempest-ServerActionsTestOtherB-303558607 tempest-ServerActionsTestOtherB-303558607-project-member] [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 763.660248] env[61957]: DEBUG oslo_vmware.api [None req-3617f05d-8cb2-4824-b3d0-fbcde866b375 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]523c1157-2c59-00a5-233b-842be608dda1, 'name': SearchDatastore_Task, 'duration_secs': 0.008343} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 763.661119] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-56bd426f-aad2-4889-98fc-ce2a6fdffdd4 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.666373] env[61957]: DEBUG oslo_vmware.api [None req-3617f05d-8cb2-4824-b3d0-fbcde866b375 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Waiting for the task: (returnval){ [ 763.666373] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]5246f0ba-fb60-66ee-5e0b-6faf1cd2516c" [ 763.666373] env[61957]: _type = "Task" [ 763.666373] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 763.674214] env[61957]: DEBUG oslo_vmware.api [None req-3617f05d-8cb2-4824-b3d0-fbcde866b375 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]5246f0ba-fb60-66ee-5e0b-6faf1cd2516c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 763.863871] env[61957]: DEBUG nova.compute.utils [None req-3d8587e3-7ad5-40a6-88af-e8ed44c49baa tempest-ServerActionsV293TestJSON-1269989829 tempest-ServerActionsV293TestJSON-1269989829-project-member] Using /dev/sd instead of None {{(pid=61957) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 763.866592] env[61957]: DEBUG nova.compute.manager [None req-3d8587e3-7ad5-40a6-88af-e8ed44c49baa tempest-ServerActionsV293TestJSON-1269989829 tempest-ServerActionsV293TestJSON-1269989829-project-member] [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] Allocating IP information in the background. {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 763.866814] env[61957]: DEBUG nova.network.neutron [None req-3d8587e3-7ad5-40a6-88af-e8ed44c49baa tempest-ServerActionsV293TestJSON-1269989829 tempest-ServerActionsV293TestJSON-1269989829-project-member] [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] allocate_for_instance() {{(pid=61957) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 763.869326] env[61957]: DEBUG nova.compute.manager [req-64cba3fb-47a5-4568-ad49-3838fd3ae6ef req-1589246c-d11f-414d-b1cd-be85aaf97edc service nova] [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] Received event network-changed-5d6fb25b-5ed5-4828-acbe-5d639e8419fa {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 763.869505] env[61957]: DEBUG nova.compute.manager [req-64cba3fb-47a5-4568-ad49-3838fd3ae6ef req-1589246c-d11f-414d-b1cd-be85aaf97edc service nova] [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] Refreshing instance network info cache due to event network-changed-5d6fb25b-5ed5-4828-acbe-5d639e8419fa. {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 763.869687] env[61957]: DEBUG oslo_concurrency.lockutils [req-64cba3fb-47a5-4568-ad49-3838fd3ae6ef req-1589246c-d11f-414d-b1cd-be85aaf97edc service nova] Acquiring lock "refresh_cache-d77276ba-6d37-440b-a113-c4f62b2e7946" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 763.912141] env[61957]: DEBUG nova.policy [None req-3d8587e3-7ad5-40a6-88af-e8ed44c49baa tempest-ServerActionsV293TestJSON-1269989829 tempest-ServerActionsV293TestJSON-1269989829-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd5c8a2d08e7340ab861c9dd4eff85333', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '65868dd9d12643eda5a5c65171bcd205', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61957) authorize /opt/stack/nova/nova/policy.py:203}} [ 764.076341] env[61957]: DEBUG oslo_concurrency.lockutils [None req-660095f0-9d5a-4797-b621-e0585ffddbba tempest-ServerActionsTestOtherB-303558607 tempest-ServerActionsTestOtherB-303558607-project-member] Releasing lock "refresh_cache-d77276ba-6d37-440b-a113-c4f62b2e7946" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 764.076818] env[61957]: DEBUG nova.compute.manager [None req-660095f0-9d5a-4797-b621-e0585ffddbba tempest-ServerActionsTestOtherB-303558607 tempest-ServerActionsTestOtherB-303558607-project-member] [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] Start destroying the instance on the hypervisor. {{(pid=61957) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 764.077093] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-660095f0-9d5a-4797-b621-e0585ffddbba tempest-ServerActionsTestOtherB-303558607 tempest-ServerActionsTestOtherB-303558607-project-member] [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] Destroying instance {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 764.077413] env[61957]: DEBUG oslo_concurrency.lockutils [req-64cba3fb-47a5-4568-ad49-3838fd3ae6ef req-1589246c-d11f-414d-b1cd-be85aaf97edc service nova] Acquired lock "refresh_cache-d77276ba-6d37-440b-a113-c4f62b2e7946" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 764.077580] env[61957]: DEBUG nova.network.neutron [req-64cba3fb-47a5-4568-ad49-3838fd3ae6ef req-1589246c-d11f-414d-b1cd-be85aaf97edc service nova] [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] Refreshing network info cache for port 5d6fb25b-5ed5-4828-acbe-5d639e8419fa {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 764.078797] env[61957]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-bed0f3d7-01f5-4115-8e11-b6c7a59ea30e {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.090117] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8d7714b-0fdd-462c-be80-4d91897bc966 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.116613] env[61957]: WARNING nova.virt.vmwareapi.vmops [None req-660095f0-9d5a-4797-b621-e0585ffddbba tempest-ServerActionsTestOtherB-303558607 tempest-ServerActionsTestOtherB-303558607-project-member] [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance d77276ba-6d37-440b-a113-c4f62b2e7946 could not be found. [ 764.116841] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-660095f0-9d5a-4797-b621-e0585ffddbba tempest-ServerActionsTestOtherB-303558607 tempest-ServerActionsTestOtherB-303558607-project-member] [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] Instance destroyed {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 764.117048] env[61957]: INFO nova.compute.manager [None req-660095f0-9d5a-4797-b621-e0585ffddbba tempest-ServerActionsTestOtherB-303558607 tempest-ServerActionsTestOtherB-303558607-project-member] [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] Took 0.04 seconds to destroy the instance on the hypervisor. [ 764.117291] env[61957]: DEBUG oslo.service.loopingcall [None req-660095f0-9d5a-4797-b621-e0585ffddbba tempest-ServerActionsTestOtherB-303558607 tempest-ServerActionsTestOtherB-303558607-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 764.119857] env[61957]: DEBUG nova.compute.manager [-] [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 764.119955] env[61957]: DEBUG nova.network.neutron [-] [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 764.140983] env[61957]: DEBUG nova.network.neutron [-] [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 764.175593] env[61957]: DEBUG oslo_vmware.api [None req-3617f05d-8cb2-4824-b3d0-fbcde866b375 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]5246f0ba-fb60-66ee-5e0b-6faf1cd2516c, 'name': SearchDatastore_Task, 'duration_secs': 0.011991} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 764.178032] env[61957]: DEBUG oslo_concurrency.lockutils [None req-3617f05d-8cb2-4824-b3d0-fbcde866b375 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 764.178456] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-3617f05d-8cb2-4824-b3d0-fbcde866b375 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore2] 7f3850c0-a455-4e5b-a329-fb3b02f9725f/7f3850c0-a455-4e5b-a329-fb3b02f9725f.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 764.179075] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0c9fdfae-9bd1-4c20-91cc-d9324fa9a287 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.185295] env[61957]: DEBUG oslo_vmware.api [None req-3617f05d-8cb2-4824-b3d0-fbcde866b375 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Waiting for the task: (returnval){ [ 764.185295] env[61957]: value = "task-1277379" [ 764.185295] env[61957]: _type = "Task" [ 764.185295] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 764.195918] env[61957]: DEBUG oslo_vmware.api [None req-3617f05d-8cb2-4824-b3d0-fbcde866b375 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Task: {'id': task-1277379, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 764.238043] env[61957]: DEBUG nova.network.neutron [None req-3d8587e3-7ad5-40a6-88af-e8ed44c49baa tempest-ServerActionsV293TestJSON-1269989829 tempest-ServerActionsV293TestJSON-1269989829-project-member] [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] Successfully created port: 37cebbf5-42fc-4387-989c-dd71a77adabe {{(pid=61957) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 764.254682] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-150e0ca5-2a2f-4b87-8109-86004d7bc7ff {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.262817] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67a80edc-a2dc-47bd-80b2-b5c241b33a07 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.292379] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f98eb5e-c622-4fd6-b9ca-26eb923717b7 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.299820] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0a19280-3ec2-4191-8695-0cc571026151 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.312914] env[61957]: DEBUG nova.compute.provider_tree [None req-5ca5fdda-dbef-4c48-b38f-9d2aa78e9841 tempest-FloatingIPsAssociationNegativeTestJSON-572957506 tempest-FloatingIPsAssociationNegativeTestJSON-572957506-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 764.371828] env[61957]: DEBUG nova.compute.manager [None req-3d8587e3-7ad5-40a6-88af-e8ed44c49baa tempest-ServerActionsV293TestJSON-1269989829 tempest-ServerActionsV293TestJSON-1269989829-project-member] [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] Start building block device mappings for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 764.608776] env[61957]: DEBUG nova.network.neutron [req-64cba3fb-47a5-4568-ad49-3838fd3ae6ef req-1589246c-d11f-414d-b1cd-be85aaf97edc service nova] [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 764.644574] env[61957]: DEBUG nova.network.neutron [-] [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 764.701202] env[61957]: DEBUG oslo_vmware.api [None req-3617f05d-8cb2-4824-b3d0-fbcde866b375 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Task: {'id': task-1277379, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.466641} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 764.701495] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-3617f05d-8cb2-4824-b3d0-fbcde866b375 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore2] 7f3850c0-a455-4e5b-a329-fb3b02f9725f/7f3850c0-a455-4e5b-a329-fb3b02f9725f.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 764.701715] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-3617f05d-8cb2-4824-b3d0-fbcde866b375 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] [instance: 7f3850c0-a455-4e5b-a329-fb3b02f9725f] Extending root virtual disk to 1048576 {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 764.701935] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-00aa962f-f25c-428f-aca0-19a13a981b3c {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.708535] env[61957]: DEBUG oslo_vmware.api [None req-3617f05d-8cb2-4824-b3d0-fbcde866b375 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Waiting for the task: (returnval){ [ 764.708535] env[61957]: value = "task-1277380" [ 764.708535] env[61957]: _type = "Task" [ 764.708535] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 764.716704] env[61957]: DEBUG oslo_vmware.api [None req-3617f05d-8cb2-4824-b3d0-fbcde866b375 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Task: {'id': task-1277380, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 764.734632] env[61957]: DEBUG nova.network.neutron [req-64cba3fb-47a5-4568-ad49-3838fd3ae6ef req-1589246c-d11f-414d-b1cd-be85aaf97edc service nova] [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 764.816126] env[61957]: DEBUG nova.scheduler.client.report [None req-5ca5fdda-dbef-4c48-b38f-9d2aa78e9841 tempest-FloatingIPsAssociationNegativeTestJSON-572957506 tempest-FloatingIPsAssociationNegativeTestJSON-572957506-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 764.878041] env[61957]: INFO nova.virt.block_device [None req-3d8587e3-7ad5-40a6-88af-e8ed44c49baa tempest-ServerActionsV293TestJSON-1269989829 tempest-ServerActionsV293TestJSON-1269989829-project-member] [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] Booting with volume e259b2ae-e225-4b0f-8784-152c208d0379 at /dev/sda [ 764.924904] env[61957]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3626409b-216d-4b86-b62f-0561a93f6c28 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.934011] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-175f60e6-64f9-4ac9-aaf2-ba2b034577d4 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.963736] env[61957]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4af2a41d-237f-4e1a-8178-ccef3c0a4744 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.971595] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8de2c367-559c-43f7-8f44-8aa4ce47dec5 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.993225] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c790fb05-14b5-4d03-81a5-d7c7c8af6dd7 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.999981] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76edbe6b-b411-4bfe-bb80-8a9590b6ff82 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.013621] env[61957]: DEBUG nova.virt.block_device [None req-3d8587e3-7ad5-40a6-88af-e8ed44c49baa tempest-ServerActionsV293TestJSON-1269989829 tempest-ServerActionsV293TestJSON-1269989829-project-member] [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] Updating existing volume attachment record: 83c715f0-7eb0-4f98-98b3-e3b9d3d89d26 {{(pid=61957) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 765.148326] env[61957]: INFO nova.compute.manager [-] [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] Took 1.03 seconds to deallocate network for instance. [ 765.150752] env[61957]: DEBUG nova.compute.claims [None req-660095f0-9d5a-4797-b621-e0585ffddbba tempest-ServerActionsTestOtherB-303558607 tempest-ServerActionsTestOtherB-303558607-project-member] [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] Aborting claim: {{(pid=61957) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 765.150931] env[61957]: DEBUG oslo_concurrency.lockutils [None req-660095f0-9d5a-4797-b621-e0585ffddbba tempest-ServerActionsTestOtherB-303558607 tempest-ServerActionsTestOtherB-303558607-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 765.220188] env[61957]: DEBUG oslo_vmware.api [None req-3617f05d-8cb2-4824-b3d0-fbcde866b375 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Task: {'id': task-1277380, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064521} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 765.220188] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-3617f05d-8cb2-4824-b3d0-fbcde866b375 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] [instance: 7f3850c0-a455-4e5b-a329-fb3b02f9725f] Extended root virtual disk {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 765.220676] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7371fac-1339-45d7-b23d-1af2454c673c {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.224453] env[61957]: ERROR nova.compute.manager [None req-3d8587e3-7ad5-40a6-88af-e8ed44c49baa tempest-ServerActionsV293TestJSON-1269989829 tempest-ServerActionsV293TestJSON-1269989829-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 37cebbf5-42fc-4387-989c-dd71a77adabe, please check neutron logs for more information. [ 765.224453] env[61957]: ERROR nova.compute.manager Traceback (most recent call last): [ 765.224453] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 765.224453] env[61957]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 765.224453] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 765.224453] env[61957]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 765.224453] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 765.224453] env[61957]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 765.224453] env[61957]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 765.224453] env[61957]: ERROR nova.compute.manager self.force_reraise() [ 765.224453] env[61957]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 765.224453] env[61957]: ERROR nova.compute.manager raise self.value [ 765.224453] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 765.224453] env[61957]: ERROR nova.compute.manager updated_port = self._update_port( [ 765.224453] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 765.224453] env[61957]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 765.225054] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 765.225054] env[61957]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 765.225054] env[61957]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 37cebbf5-42fc-4387-989c-dd71a77adabe, please check neutron logs for more information. [ 765.225054] env[61957]: ERROR nova.compute.manager [ 765.225391] env[61957]: Traceback (most recent call last): [ 765.225430] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 765.225430] env[61957]: listener.cb(fileno) [ 765.225430] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 765.225430] env[61957]: result = function(*args, **kwargs) [ 765.225430] env[61957]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 765.225430] env[61957]: return func(*args, **kwargs) [ 765.225430] env[61957]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 765.225430] env[61957]: raise e [ 765.225430] env[61957]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 765.225430] env[61957]: nwinfo = self.network_api.allocate_for_instance( [ 765.225430] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 765.225430] env[61957]: created_port_ids = self._update_ports_for_instance( [ 765.225430] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 765.225430] env[61957]: with excutils.save_and_reraise_exception(): [ 765.225430] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 765.225430] env[61957]: self.force_reraise() [ 765.225430] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 765.225430] env[61957]: raise self.value [ 765.225430] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 765.225430] env[61957]: updated_port = self._update_port( [ 765.225430] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 765.225430] env[61957]: _ensure_no_port_binding_failure(port) [ 765.225430] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 765.225430] env[61957]: raise exception.PortBindingFailed(port_id=port['id']) [ 765.226275] env[61957]: nova.exception.PortBindingFailed: Binding failed for port 37cebbf5-42fc-4387-989c-dd71a77adabe, please check neutron logs for more information. [ 765.226275] env[61957]: Removing descriptor: 16 [ 765.237259] env[61957]: DEBUG oslo_concurrency.lockutils [req-64cba3fb-47a5-4568-ad49-3838fd3ae6ef req-1589246c-d11f-414d-b1cd-be85aaf97edc service nova] Releasing lock "refresh_cache-d77276ba-6d37-440b-a113-c4f62b2e7946" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 765.237520] env[61957]: DEBUG nova.compute.manager [req-64cba3fb-47a5-4568-ad49-3838fd3ae6ef req-1589246c-d11f-414d-b1cd-be85aaf97edc service nova] [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] Received event network-vif-deleted-5d6fb25b-5ed5-4828-acbe-5d639e8419fa {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 765.246147] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-3617f05d-8cb2-4824-b3d0-fbcde866b375 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] [instance: 7f3850c0-a455-4e5b-a329-fb3b02f9725f] Reconfiguring VM instance instance-00000028 to attach disk [datastore2] 7f3850c0-a455-4e5b-a329-fb3b02f9725f/7f3850c0-a455-4e5b-a329-fb3b02f9725f.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 765.246413] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-63c6ac57-be33-417f-905b-e06b32d10bf5 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.265203] env[61957]: DEBUG oslo_vmware.api [None req-3617f05d-8cb2-4824-b3d0-fbcde866b375 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Waiting for the task: (returnval){ [ 765.265203] env[61957]: value = "task-1277381" [ 765.265203] env[61957]: _type = "Task" [ 765.265203] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 765.272724] env[61957]: DEBUG oslo_vmware.api [None req-3617f05d-8cb2-4824-b3d0-fbcde866b375 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Task: {'id': task-1277381, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 765.321874] env[61957]: DEBUG oslo_concurrency.lockutils [None req-5ca5fdda-dbef-4c48-b38f-9d2aa78e9841 tempest-FloatingIPsAssociationNegativeTestJSON-572957506 tempest-FloatingIPsAssociationNegativeTestJSON-572957506-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.966s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 765.322650] env[61957]: ERROR nova.compute.manager [None req-5ca5fdda-dbef-4c48-b38f-9d2aa78e9841 tempest-FloatingIPsAssociationNegativeTestJSON-572957506 tempest-FloatingIPsAssociationNegativeTestJSON-572957506-project-member] [instance: 348ea580-2268-4082-a5f1-33c954063e4d] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 89bc10a8-611c-42c9-ac0f-b4d98e2005d2, please check neutron logs for more information. [ 765.322650] env[61957]: ERROR nova.compute.manager [instance: 348ea580-2268-4082-a5f1-33c954063e4d] Traceback (most recent call last): [ 765.322650] env[61957]: ERROR nova.compute.manager [instance: 348ea580-2268-4082-a5f1-33c954063e4d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 765.322650] env[61957]: ERROR nova.compute.manager [instance: 348ea580-2268-4082-a5f1-33c954063e4d] self.driver.spawn(context, instance, image_meta, [ 765.322650] env[61957]: ERROR nova.compute.manager [instance: 348ea580-2268-4082-a5f1-33c954063e4d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 765.322650] env[61957]: ERROR nova.compute.manager [instance: 348ea580-2268-4082-a5f1-33c954063e4d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 765.322650] env[61957]: ERROR nova.compute.manager [instance: 348ea580-2268-4082-a5f1-33c954063e4d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 765.322650] env[61957]: ERROR nova.compute.manager [instance: 348ea580-2268-4082-a5f1-33c954063e4d] vm_ref = self.build_virtual_machine(instance, [ 765.322650] env[61957]: ERROR nova.compute.manager [instance: 348ea580-2268-4082-a5f1-33c954063e4d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 765.322650] env[61957]: ERROR nova.compute.manager [instance: 348ea580-2268-4082-a5f1-33c954063e4d] vif_infos = vmwarevif.get_vif_info(self._session, [ 765.322650] env[61957]: ERROR nova.compute.manager [instance: 348ea580-2268-4082-a5f1-33c954063e4d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 765.323034] env[61957]: ERROR nova.compute.manager [instance: 348ea580-2268-4082-a5f1-33c954063e4d] for vif in network_info: [ 765.323034] env[61957]: ERROR nova.compute.manager [instance: 348ea580-2268-4082-a5f1-33c954063e4d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 765.323034] env[61957]: ERROR nova.compute.manager [instance: 348ea580-2268-4082-a5f1-33c954063e4d] return self._sync_wrapper(fn, *args, **kwargs) [ 765.323034] env[61957]: ERROR nova.compute.manager [instance: 348ea580-2268-4082-a5f1-33c954063e4d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 765.323034] env[61957]: ERROR nova.compute.manager [instance: 348ea580-2268-4082-a5f1-33c954063e4d] self.wait() [ 765.323034] env[61957]: ERROR nova.compute.manager [instance: 348ea580-2268-4082-a5f1-33c954063e4d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 765.323034] env[61957]: ERROR nova.compute.manager [instance: 348ea580-2268-4082-a5f1-33c954063e4d] self[:] = self._gt.wait() [ 765.323034] env[61957]: ERROR nova.compute.manager [instance: 348ea580-2268-4082-a5f1-33c954063e4d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 765.323034] env[61957]: ERROR nova.compute.manager [instance: 348ea580-2268-4082-a5f1-33c954063e4d] return self._exit_event.wait() [ 765.323034] env[61957]: ERROR nova.compute.manager [instance: 348ea580-2268-4082-a5f1-33c954063e4d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 765.323034] env[61957]: ERROR nova.compute.manager [instance: 348ea580-2268-4082-a5f1-33c954063e4d] current.throw(*self._exc) [ 765.323034] env[61957]: ERROR nova.compute.manager [instance: 348ea580-2268-4082-a5f1-33c954063e4d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 765.323034] env[61957]: ERROR nova.compute.manager [instance: 348ea580-2268-4082-a5f1-33c954063e4d] result = function(*args, **kwargs) [ 765.323705] env[61957]: ERROR nova.compute.manager [instance: 348ea580-2268-4082-a5f1-33c954063e4d] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 765.323705] env[61957]: ERROR nova.compute.manager [instance: 348ea580-2268-4082-a5f1-33c954063e4d] return func(*args, **kwargs) [ 765.323705] env[61957]: ERROR nova.compute.manager [instance: 348ea580-2268-4082-a5f1-33c954063e4d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 765.323705] env[61957]: ERROR nova.compute.manager [instance: 348ea580-2268-4082-a5f1-33c954063e4d] raise e [ 765.323705] env[61957]: ERROR nova.compute.manager [instance: 348ea580-2268-4082-a5f1-33c954063e4d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 765.323705] env[61957]: ERROR nova.compute.manager [instance: 348ea580-2268-4082-a5f1-33c954063e4d] nwinfo = self.network_api.allocate_for_instance( [ 765.323705] env[61957]: ERROR nova.compute.manager [instance: 348ea580-2268-4082-a5f1-33c954063e4d] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 765.323705] env[61957]: ERROR nova.compute.manager [instance: 348ea580-2268-4082-a5f1-33c954063e4d] created_port_ids = self._update_ports_for_instance( [ 765.323705] env[61957]: ERROR nova.compute.manager [instance: 348ea580-2268-4082-a5f1-33c954063e4d] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 765.323705] env[61957]: ERROR nova.compute.manager [instance: 348ea580-2268-4082-a5f1-33c954063e4d] with excutils.save_and_reraise_exception(): [ 765.323705] env[61957]: ERROR nova.compute.manager [instance: 348ea580-2268-4082-a5f1-33c954063e4d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 765.323705] env[61957]: ERROR nova.compute.manager [instance: 348ea580-2268-4082-a5f1-33c954063e4d] self.force_reraise() [ 765.323705] env[61957]: ERROR nova.compute.manager [instance: 348ea580-2268-4082-a5f1-33c954063e4d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 765.324380] env[61957]: ERROR nova.compute.manager [instance: 348ea580-2268-4082-a5f1-33c954063e4d] raise self.value [ 765.324380] env[61957]: ERROR nova.compute.manager [instance: 348ea580-2268-4082-a5f1-33c954063e4d] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 765.324380] env[61957]: ERROR nova.compute.manager [instance: 348ea580-2268-4082-a5f1-33c954063e4d] updated_port = self._update_port( [ 765.324380] env[61957]: ERROR nova.compute.manager [instance: 348ea580-2268-4082-a5f1-33c954063e4d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 765.324380] env[61957]: ERROR nova.compute.manager [instance: 348ea580-2268-4082-a5f1-33c954063e4d] _ensure_no_port_binding_failure(port) [ 765.324380] env[61957]: ERROR nova.compute.manager [instance: 348ea580-2268-4082-a5f1-33c954063e4d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 765.324380] env[61957]: ERROR nova.compute.manager [instance: 348ea580-2268-4082-a5f1-33c954063e4d] raise exception.PortBindingFailed(port_id=port['id']) [ 765.324380] env[61957]: ERROR nova.compute.manager [instance: 348ea580-2268-4082-a5f1-33c954063e4d] nova.exception.PortBindingFailed: Binding failed for port 89bc10a8-611c-42c9-ac0f-b4d98e2005d2, please check neutron logs for more information. [ 765.324380] env[61957]: ERROR nova.compute.manager [instance: 348ea580-2268-4082-a5f1-33c954063e4d] [ 765.324380] env[61957]: DEBUG nova.compute.utils [None req-5ca5fdda-dbef-4c48-b38f-9d2aa78e9841 tempest-FloatingIPsAssociationNegativeTestJSON-572957506 tempest-FloatingIPsAssociationNegativeTestJSON-572957506-project-member] [instance: 348ea580-2268-4082-a5f1-33c954063e4d] Binding failed for port 89bc10a8-611c-42c9-ac0f-b4d98e2005d2, please check neutron logs for more information. {{(pid=61957) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 765.324743] env[61957]: DEBUG oslo_concurrency.lockutils [None req-394366c7-3cf4-481b-bb9f-4da67892af40 tempest-ServersTestManualDisk-495750068 tempest-ServersTestManualDisk-495750068-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.625s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 765.326197] env[61957]: INFO nova.compute.claims [None req-394366c7-3cf4-481b-bb9f-4da67892af40 tempest-ServersTestManualDisk-495750068 tempest-ServersTestManualDisk-495750068-project-member] [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 765.329591] env[61957]: DEBUG nova.compute.manager [None req-5ca5fdda-dbef-4c48-b38f-9d2aa78e9841 tempest-FloatingIPsAssociationNegativeTestJSON-572957506 tempest-FloatingIPsAssociationNegativeTestJSON-572957506-project-member] [instance: 348ea580-2268-4082-a5f1-33c954063e4d] Build of instance 348ea580-2268-4082-a5f1-33c954063e4d was re-scheduled: Binding failed for port 89bc10a8-611c-42c9-ac0f-b4d98e2005d2, please check neutron logs for more information. {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 765.329591] env[61957]: DEBUG nova.compute.manager [None req-5ca5fdda-dbef-4c48-b38f-9d2aa78e9841 tempest-FloatingIPsAssociationNegativeTestJSON-572957506 tempest-FloatingIPsAssociationNegativeTestJSON-572957506-project-member] [instance: 348ea580-2268-4082-a5f1-33c954063e4d] Unplugging VIFs for instance {{(pid=61957) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 765.329830] env[61957]: DEBUG oslo_concurrency.lockutils [None req-5ca5fdda-dbef-4c48-b38f-9d2aa78e9841 tempest-FloatingIPsAssociationNegativeTestJSON-572957506 tempest-FloatingIPsAssociationNegativeTestJSON-572957506-project-member] Acquiring lock "refresh_cache-348ea580-2268-4082-a5f1-33c954063e4d" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 765.330392] env[61957]: DEBUG oslo_concurrency.lockutils [None req-5ca5fdda-dbef-4c48-b38f-9d2aa78e9841 tempest-FloatingIPsAssociationNegativeTestJSON-572957506 tempest-FloatingIPsAssociationNegativeTestJSON-572957506-project-member] Acquired lock "refresh_cache-348ea580-2268-4082-a5f1-33c954063e4d" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 765.330392] env[61957]: DEBUG nova.network.neutron [None req-5ca5fdda-dbef-4c48-b38f-9d2aa78e9841 tempest-FloatingIPsAssociationNegativeTestJSON-572957506 tempest-FloatingIPsAssociationNegativeTestJSON-572957506-project-member] [instance: 348ea580-2268-4082-a5f1-33c954063e4d] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 765.774627] env[61957]: DEBUG oslo_vmware.api [None req-3617f05d-8cb2-4824-b3d0-fbcde866b375 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Task: {'id': task-1277381, 'name': ReconfigVM_Task, 'duration_secs': 0.293816} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 765.775862] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-3617f05d-8cb2-4824-b3d0-fbcde866b375 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] [instance: 7f3850c0-a455-4e5b-a329-fb3b02f9725f] Reconfigured VM instance instance-00000028 to attach disk [datastore2] 7f3850c0-a455-4e5b-a329-fb3b02f9725f/7f3850c0-a455-4e5b-a329-fb3b02f9725f.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 765.775862] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b773b99c-e419-484e-901c-c1c470ad3030 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.782030] env[61957]: DEBUG oslo_vmware.api [None req-3617f05d-8cb2-4824-b3d0-fbcde866b375 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Waiting for the task: (returnval){ [ 765.782030] env[61957]: value = "task-1277382" [ 765.782030] env[61957]: _type = "Task" [ 765.782030] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 765.789686] env[61957]: DEBUG oslo_vmware.api [None req-3617f05d-8cb2-4824-b3d0-fbcde866b375 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Task: {'id': task-1277382, 'name': Rename_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 765.851926] env[61957]: DEBUG nova.network.neutron [None req-5ca5fdda-dbef-4c48-b38f-9d2aa78e9841 tempest-FloatingIPsAssociationNegativeTestJSON-572957506 tempest-FloatingIPsAssociationNegativeTestJSON-572957506-project-member] [instance: 348ea580-2268-4082-a5f1-33c954063e4d] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 765.891694] env[61957]: DEBUG nova.compute.manager [req-3e770225-491c-479b-bad1-da19611aaef8 req-23532920-1c24-4980-bec0-235b66f3d52e service nova] [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] Received event network-changed-37cebbf5-42fc-4387-989c-dd71a77adabe {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 765.891979] env[61957]: DEBUG nova.compute.manager [req-3e770225-491c-479b-bad1-da19611aaef8 req-23532920-1c24-4980-bec0-235b66f3d52e service nova] [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] Refreshing instance network info cache due to event network-changed-37cebbf5-42fc-4387-989c-dd71a77adabe. {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 765.892551] env[61957]: DEBUG oslo_concurrency.lockutils [req-3e770225-491c-479b-bad1-da19611aaef8 req-23532920-1c24-4980-bec0-235b66f3d52e service nova] Acquiring lock "refresh_cache-5dc993e9-5970-43ae-90be-f3a53d1f090a" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 765.892801] env[61957]: DEBUG oslo_concurrency.lockutils [req-3e770225-491c-479b-bad1-da19611aaef8 req-23532920-1c24-4980-bec0-235b66f3d52e service nova] Acquired lock "refresh_cache-5dc993e9-5970-43ae-90be-f3a53d1f090a" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 765.892851] env[61957]: DEBUG nova.network.neutron [req-3e770225-491c-479b-bad1-da19611aaef8 req-23532920-1c24-4980-bec0-235b66f3d52e service nova] [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] Refreshing network info cache for port 37cebbf5-42fc-4387-989c-dd71a77adabe {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 765.944824] env[61957]: DEBUG nova.network.neutron [None req-5ca5fdda-dbef-4c48-b38f-9d2aa78e9841 tempest-FloatingIPsAssociationNegativeTestJSON-572957506 tempest-FloatingIPsAssociationNegativeTestJSON-572957506-project-member] [instance: 348ea580-2268-4082-a5f1-33c954063e4d] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 766.291681] env[61957]: DEBUG oslo_vmware.api [None req-3617f05d-8cb2-4824-b3d0-fbcde866b375 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Task: {'id': task-1277382, 'name': Rename_Task, 'duration_secs': 0.135595} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 766.291979] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-3617f05d-8cb2-4824-b3d0-fbcde866b375 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] [instance: 7f3850c0-a455-4e5b-a329-fb3b02f9725f] Powering on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 766.292196] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0ced23b8-4b23-4e8b-95e8-f32dec86470c {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.298006] env[61957]: DEBUG oslo_vmware.api [None req-3617f05d-8cb2-4824-b3d0-fbcde866b375 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Waiting for the task: (returnval){ [ 766.298006] env[61957]: value = "task-1277383" [ 766.298006] env[61957]: _type = "Task" [ 766.298006] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 766.304900] env[61957]: DEBUG oslo_vmware.api [None req-3617f05d-8cb2-4824-b3d0-fbcde866b375 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Task: {'id': task-1277383, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 766.411067] env[61957]: DEBUG nova.network.neutron [req-3e770225-491c-479b-bad1-da19611aaef8 req-23532920-1c24-4980-bec0-235b66f3d52e service nova] [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 766.448248] env[61957]: DEBUG oslo_concurrency.lockutils [None req-5ca5fdda-dbef-4c48-b38f-9d2aa78e9841 tempest-FloatingIPsAssociationNegativeTestJSON-572957506 tempest-FloatingIPsAssociationNegativeTestJSON-572957506-project-member] Releasing lock "refresh_cache-348ea580-2268-4082-a5f1-33c954063e4d" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 766.448520] env[61957]: DEBUG nova.compute.manager [None req-5ca5fdda-dbef-4c48-b38f-9d2aa78e9841 tempest-FloatingIPsAssociationNegativeTestJSON-572957506 tempest-FloatingIPsAssociationNegativeTestJSON-572957506-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61957) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 766.448716] env[61957]: DEBUG nova.compute.manager [None req-5ca5fdda-dbef-4c48-b38f-9d2aa78e9841 tempest-FloatingIPsAssociationNegativeTestJSON-572957506 tempest-FloatingIPsAssociationNegativeTestJSON-572957506-project-member] [instance: 348ea580-2268-4082-a5f1-33c954063e4d] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 766.448923] env[61957]: DEBUG nova.network.neutron [None req-5ca5fdda-dbef-4c48-b38f-9d2aa78e9841 tempest-FloatingIPsAssociationNegativeTestJSON-572957506 tempest-FloatingIPsAssociationNegativeTestJSON-572957506-project-member] [instance: 348ea580-2268-4082-a5f1-33c954063e4d] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 766.477024] env[61957]: DEBUG nova.network.neutron [None req-5ca5fdda-dbef-4c48-b38f-9d2aa78e9841 tempest-FloatingIPsAssociationNegativeTestJSON-572957506 tempest-FloatingIPsAssociationNegativeTestJSON-572957506-project-member] [instance: 348ea580-2268-4082-a5f1-33c954063e4d] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 766.499413] env[61957]: DEBUG nova.network.neutron [req-3e770225-491c-479b-bad1-da19611aaef8 req-23532920-1c24-4980-bec0-235b66f3d52e service nova] [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 766.650486] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc4ef4bc-e584-41ad-9382-2b6d9d2269ea {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.659882] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d816fa81-16cb-4b13-af61-dc21b9dbb485 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.688574] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32df3c72-5364-45e6-b205-362776ca00bd {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.695771] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61ef5f7d-83b1-4667-bcf3-22a3bc8d79d3 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.708838] env[61957]: DEBUG nova.compute.provider_tree [None req-394366c7-3cf4-481b-bb9f-4da67892af40 tempest-ServersTestManualDisk-495750068 tempest-ServersTestManualDisk-495750068-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 766.808260] env[61957]: DEBUG oslo_vmware.api [None req-3617f05d-8cb2-4824-b3d0-fbcde866b375 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Task: {'id': task-1277383, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 766.979573] env[61957]: DEBUG nova.network.neutron [None req-5ca5fdda-dbef-4c48-b38f-9d2aa78e9841 tempest-FloatingIPsAssociationNegativeTestJSON-572957506 tempest-FloatingIPsAssociationNegativeTestJSON-572957506-project-member] [instance: 348ea580-2268-4082-a5f1-33c954063e4d] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 767.001776] env[61957]: DEBUG oslo_concurrency.lockutils [req-3e770225-491c-479b-bad1-da19611aaef8 req-23532920-1c24-4980-bec0-235b66f3d52e service nova] Releasing lock "refresh_cache-5dc993e9-5970-43ae-90be-f3a53d1f090a" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 767.002039] env[61957]: DEBUG nova.compute.manager [req-3e770225-491c-479b-bad1-da19611aaef8 req-23532920-1c24-4980-bec0-235b66f3d52e service nova] [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] Received event network-vif-deleted-37cebbf5-42fc-4387-989c-dd71a77adabe {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 767.099760] env[61957]: DEBUG nova.compute.manager [None req-3d8587e3-7ad5-40a6-88af-e8ed44c49baa tempest-ServerActionsV293TestJSON-1269989829 tempest-ServerActionsV293TestJSON-1269989829-project-member] [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] Start spawning the instance on the hypervisor. {{(pid=61957) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 767.100407] env[61957]: DEBUG nova.virt.hardware [None req-3d8587e3-7ad5-40a6-88af-e8ed44c49baa tempest-ServerActionsV293TestJSON-1269989829 tempest-ServerActionsV293TestJSON-1269989829-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T17:22:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 767.100685] env[61957]: DEBUG nova.virt.hardware [None req-3d8587e3-7ad5-40a6-88af-e8ed44c49baa tempest-ServerActionsV293TestJSON-1269989829 tempest-ServerActionsV293TestJSON-1269989829-project-member] Flavor limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 767.100879] env[61957]: DEBUG nova.virt.hardware [None req-3d8587e3-7ad5-40a6-88af-e8ed44c49baa tempest-ServerActionsV293TestJSON-1269989829 tempest-ServerActionsV293TestJSON-1269989829-project-member] Image limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 767.101088] env[61957]: DEBUG nova.virt.hardware [None req-3d8587e3-7ad5-40a6-88af-e8ed44c49baa tempest-ServerActionsV293TestJSON-1269989829 tempest-ServerActionsV293TestJSON-1269989829-project-member] Flavor pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 767.101235] env[61957]: DEBUG nova.virt.hardware [None req-3d8587e3-7ad5-40a6-88af-e8ed44c49baa tempest-ServerActionsV293TestJSON-1269989829 tempest-ServerActionsV293TestJSON-1269989829-project-member] Image pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 767.101379] env[61957]: DEBUG nova.virt.hardware [None req-3d8587e3-7ad5-40a6-88af-e8ed44c49baa tempest-ServerActionsV293TestJSON-1269989829 tempest-ServerActionsV293TestJSON-1269989829-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 767.101580] env[61957]: DEBUG nova.virt.hardware [None req-3d8587e3-7ad5-40a6-88af-e8ed44c49baa tempest-ServerActionsV293TestJSON-1269989829 tempest-ServerActionsV293TestJSON-1269989829-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 767.101735] env[61957]: DEBUG nova.virt.hardware [None req-3d8587e3-7ad5-40a6-88af-e8ed44c49baa tempest-ServerActionsV293TestJSON-1269989829 tempest-ServerActionsV293TestJSON-1269989829-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 767.101934] env[61957]: DEBUG nova.virt.hardware [None req-3d8587e3-7ad5-40a6-88af-e8ed44c49baa tempest-ServerActionsV293TestJSON-1269989829 tempest-ServerActionsV293TestJSON-1269989829-project-member] Got 1 possible topologies {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 767.102162] env[61957]: DEBUG nova.virt.hardware [None req-3d8587e3-7ad5-40a6-88af-e8ed44c49baa tempest-ServerActionsV293TestJSON-1269989829 tempest-ServerActionsV293TestJSON-1269989829-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 767.102370] env[61957]: DEBUG nova.virt.hardware [None req-3d8587e3-7ad5-40a6-88af-e8ed44c49baa tempest-ServerActionsV293TestJSON-1269989829 tempest-ServerActionsV293TestJSON-1269989829-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 767.103228] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e66ab53b-2d86-44fa-85d1-18dcd3d329af {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.112324] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33b36291-e64e-4868-97b2-19adf8ed9f8a {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.126056] env[61957]: ERROR nova.compute.manager [None req-3d8587e3-7ad5-40a6-88af-e8ed44c49baa tempest-ServerActionsV293TestJSON-1269989829 tempest-ServerActionsV293TestJSON-1269989829-project-member] [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 37cebbf5-42fc-4387-989c-dd71a77adabe, please check neutron logs for more information. [ 767.126056] env[61957]: ERROR nova.compute.manager [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] Traceback (most recent call last): [ 767.126056] env[61957]: ERROR nova.compute.manager [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 767.126056] env[61957]: ERROR nova.compute.manager [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] yield resources [ 767.126056] env[61957]: ERROR nova.compute.manager [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 767.126056] env[61957]: ERROR nova.compute.manager [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] self.driver.spawn(context, instance, image_meta, [ 767.126056] env[61957]: ERROR nova.compute.manager [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 767.126056] env[61957]: ERROR nova.compute.manager [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 767.126056] env[61957]: ERROR nova.compute.manager [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 767.126056] env[61957]: ERROR nova.compute.manager [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] vm_ref = self.build_virtual_machine(instance, [ 767.126056] env[61957]: ERROR nova.compute.manager [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 767.126449] env[61957]: ERROR nova.compute.manager [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] vif_infos = vmwarevif.get_vif_info(self._session, [ 767.126449] env[61957]: ERROR nova.compute.manager [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 767.126449] env[61957]: ERROR nova.compute.manager [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] for vif in network_info: [ 767.126449] env[61957]: ERROR nova.compute.manager [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 767.126449] env[61957]: ERROR nova.compute.manager [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] return self._sync_wrapper(fn, *args, **kwargs) [ 767.126449] env[61957]: ERROR nova.compute.manager [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 767.126449] env[61957]: ERROR nova.compute.manager [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] self.wait() [ 767.126449] env[61957]: ERROR nova.compute.manager [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 767.126449] env[61957]: ERROR nova.compute.manager [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] self[:] = self._gt.wait() [ 767.126449] env[61957]: ERROR nova.compute.manager [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 767.126449] env[61957]: ERROR nova.compute.manager [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] return self._exit_event.wait() [ 767.126449] env[61957]: ERROR nova.compute.manager [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 767.126449] env[61957]: ERROR nova.compute.manager [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] current.throw(*self._exc) [ 767.127026] env[61957]: ERROR nova.compute.manager [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 767.127026] env[61957]: ERROR nova.compute.manager [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] result = function(*args, **kwargs) [ 767.127026] env[61957]: ERROR nova.compute.manager [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 767.127026] env[61957]: ERROR nova.compute.manager [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] return func(*args, **kwargs) [ 767.127026] env[61957]: ERROR nova.compute.manager [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 767.127026] env[61957]: ERROR nova.compute.manager [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] raise e [ 767.127026] env[61957]: ERROR nova.compute.manager [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 767.127026] env[61957]: ERROR nova.compute.manager [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] nwinfo = self.network_api.allocate_for_instance( [ 767.127026] env[61957]: ERROR nova.compute.manager [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 767.127026] env[61957]: ERROR nova.compute.manager [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] created_port_ids = self._update_ports_for_instance( [ 767.127026] env[61957]: ERROR nova.compute.manager [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 767.127026] env[61957]: ERROR nova.compute.manager [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] with excutils.save_and_reraise_exception(): [ 767.127026] env[61957]: ERROR nova.compute.manager [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 767.127456] env[61957]: ERROR nova.compute.manager [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] self.force_reraise() [ 767.127456] env[61957]: ERROR nova.compute.manager [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 767.127456] env[61957]: ERROR nova.compute.manager [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] raise self.value [ 767.127456] env[61957]: ERROR nova.compute.manager [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 767.127456] env[61957]: ERROR nova.compute.manager [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] updated_port = self._update_port( [ 767.127456] env[61957]: ERROR nova.compute.manager [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 767.127456] env[61957]: ERROR nova.compute.manager [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] _ensure_no_port_binding_failure(port) [ 767.127456] env[61957]: ERROR nova.compute.manager [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 767.127456] env[61957]: ERROR nova.compute.manager [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] raise exception.PortBindingFailed(port_id=port['id']) [ 767.127456] env[61957]: ERROR nova.compute.manager [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] nova.exception.PortBindingFailed: Binding failed for port 37cebbf5-42fc-4387-989c-dd71a77adabe, please check neutron logs for more information. [ 767.127456] env[61957]: ERROR nova.compute.manager [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] [ 767.127456] env[61957]: INFO nova.compute.manager [None req-3d8587e3-7ad5-40a6-88af-e8ed44c49baa tempest-ServerActionsV293TestJSON-1269989829 tempest-ServerActionsV293TestJSON-1269989829-project-member] [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] Terminating instance [ 767.128310] env[61957]: DEBUG oslo_concurrency.lockutils [None req-3d8587e3-7ad5-40a6-88af-e8ed44c49baa tempest-ServerActionsV293TestJSON-1269989829 tempest-ServerActionsV293TestJSON-1269989829-project-member] Acquiring lock "refresh_cache-5dc993e9-5970-43ae-90be-f3a53d1f090a" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 767.128464] env[61957]: DEBUG oslo_concurrency.lockutils [None req-3d8587e3-7ad5-40a6-88af-e8ed44c49baa tempest-ServerActionsV293TestJSON-1269989829 tempest-ServerActionsV293TestJSON-1269989829-project-member] Acquired lock "refresh_cache-5dc993e9-5970-43ae-90be-f3a53d1f090a" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 767.128629] env[61957]: DEBUG nova.network.neutron [None req-3d8587e3-7ad5-40a6-88af-e8ed44c49baa tempest-ServerActionsV293TestJSON-1269989829 tempest-ServerActionsV293TestJSON-1269989829-project-member] [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 767.212797] env[61957]: DEBUG nova.scheduler.client.report [None req-394366c7-3cf4-481b-bb9f-4da67892af40 tempest-ServersTestManualDisk-495750068 tempest-ServersTestManualDisk-495750068-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 767.309350] env[61957]: DEBUG oslo_vmware.api [None req-3617f05d-8cb2-4824-b3d0-fbcde866b375 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Task: {'id': task-1277383, 'name': PowerOnVM_Task, 'duration_secs': 0.554069} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 767.309505] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-3617f05d-8cb2-4824-b3d0-fbcde866b375 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] [instance: 7f3850c0-a455-4e5b-a329-fb3b02f9725f] Powered on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 767.309612] env[61957]: DEBUG nova.compute.manager [None req-3617f05d-8cb2-4824-b3d0-fbcde866b375 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] [instance: 7f3850c0-a455-4e5b-a329-fb3b02f9725f] Checking state {{(pid=61957) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 767.310381] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1043be85-0156-4fb1-aa88-8aa7f8499fd1 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.482429] env[61957]: INFO nova.compute.manager [None req-5ca5fdda-dbef-4c48-b38f-9d2aa78e9841 tempest-FloatingIPsAssociationNegativeTestJSON-572957506 tempest-FloatingIPsAssociationNegativeTestJSON-572957506-project-member] [instance: 348ea580-2268-4082-a5f1-33c954063e4d] Took 1.03 seconds to deallocate network for instance. [ 767.647839] env[61957]: DEBUG nova.network.neutron [None req-3d8587e3-7ad5-40a6-88af-e8ed44c49baa tempest-ServerActionsV293TestJSON-1269989829 tempest-ServerActionsV293TestJSON-1269989829-project-member] [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 767.719887] env[61957]: DEBUG nova.network.neutron [None req-3d8587e3-7ad5-40a6-88af-e8ed44c49baa tempest-ServerActionsV293TestJSON-1269989829 tempest-ServerActionsV293TestJSON-1269989829-project-member] [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 767.722742] env[61957]: DEBUG oslo_concurrency.lockutils [None req-394366c7-3cf4-481b-bb9f-4da67892af40 tempest-ServersTestManualDisk-495750068 tempest-ServersTestManualDisk-495750068-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.398s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 767.723246] env[61957]: DEBUG nova.compute.manager [None req-394366c7-3cf4-481b-bb9f-4da67892af40 tempest-ServersTestManualDisk-495750068 tempest-ServersTestManualDisk-495750068-project-member] [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] Start building networks asynchronously for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 767.725867] env[61957]: DEBUG oslo_concurrency.lockutils [None req-838e8ad7-4b6c-4dbc-8d4b-8380b652e559 tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.682s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 767.827321] env[61957]: DEBUG oslo_concurrency.lockutils [None req-3617f05d-8cb2-4824-b3d0-fbcde866b375 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 768.225598] env[61957]: DEBUG oslo_concurrency.lockutils [None req-3d8587e3-7ad5-40a6-88af-e8ed44c49baa tempest-ServerActionsV293TestJSON-1269989829 tempest-ServerActionsV293TestJSON-1269989829-project-member] Releasing lock "refresh_cache-5dc993e9-5970-43ae-90be-f3a53d1f090a" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 768.225598] env[61957]: DEBUG nova.compute.manager [None req-3d8587e3-7ad5-40a6-88af-e8ed44c49baa tempest-ServerActionsV293TestJSON-1269989829 tempest-ServerActionsV293TestJSON-1269989829-project-member] [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] Start destroying the instance on the hypervisor. {{(pid=61957) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 768.225598] env[61957]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-20c470cb-75a8-446f-b229-b16fa0c39658 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.229748] env[61957]: DEBUG nova.compute.utils [None req-394366c7-3cf4-481b-bb9f-4da67892af40 tempest-ServersTestManualDisk-495750068 tempest-ServersTestManualDisk-495750068-project-member] Using /dev/sd instead of None {{(pid=61957) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 768.232693] env[61957]: DEBUG nova.compute.manager [None req-394366c7-3cf4-481b-bb9f-4da67892af40 tempest-ServersTestManualDisk-495750068 tempest-ServersTestManualDisk-495750068-project-member] [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] Allocating IP information in the background. {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 768.232693] env[61957]: DEBUG nova.network.neutron [None req-394366c7-3cf4-481b-bb9f-4da67892af40 tempest-ServersTestManualDisk-495750068 tempest-ServersTestManualDisk-495750068-project-member] [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] allocate_for_instance() {{(pid=61957) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 768.240640] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b22e51a-96a7-4ee3-8244-5f18f03fe30c {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.267852] env[61957]: WARNING nova.virt.vmwareapi.driver [None req-3d8587e3-7ad5-40a6-88af-e8ed44c49baa tempest-ServerActionsV293TestJSON-1269989829 tempest-ServerActionsV293TestJSON-1269989829-project-member] [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] Instance does not exists. Proceeding to delete instance properties on datastore: nova.exception.InstanceNotFound: Instance 5dc993e9-5970-43ae-90be-f3a53d1f090a could not be found. [ 768.268299] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-3d8587e3-7ad5-40a6-88af-e8ed44c49baa tempest-ServerActionsV293TestJSON-1269989829 tempest-ServerActionsV293TestJSON-1269989829-project-member] [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] Destroying instance {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 768.268667] env[61957]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a57017e1-ebc8-4050-a3ad-80053c3c6b72 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.279287] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e64973e8-8ed1-450e-8af1-0ad58cd97d48 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.306317] env[61957]: WARNING nova.virt.vmwareapi.vmops [None req-3d8587e3-7ad5-40a6-88af-e8ed44c49baa tempest-ServerActionsV293TestJSON-1269989829 tempest-ServerActionsV293TestJSON-1269989829-project-member] [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 5dc993e9-5970-43ae-90be-f3a53d1f090a could not be found. [ 768.306638] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-3d8587e3-7ad5-40a6-88af-e8ed44c49baa tempest-ServerActionsV293TestJSON-1269989829 tempest-ServerActionsV293TestJSON-1269989829-project-member] [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] Instance destroyed {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 768.306903] env[61957]: INFO nova.compute.manager [None req-3d8587e3-7ad5-40a6-88af-e8ed44c49baa tempest-ServerActionsV293TestJSON-1269989829 tempest-ServerActionsV293TestJSON-1269989829-project-member] [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] Took 0.08 seconds to destroy the instance on the hypervisor. [ 768.307215] env[61957]: DEBUG oslo.service.loopingcall [None req-3d8587e3-7ad5-40a6-88af-e8ed44c49baa tempest-ServerActionsV293TestJSON-1269989829 tempest-ServerActionsV293TestJSON-1269989829-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 768.310073] env[61957]: DEBUG nova.compute.manager [-] [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 768.310230] env[61957]: DEBUG nova.network.neutron [-] [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 768.319702] env[61957]: DEBUG nova.policy [None req-394366c7-3cf4-481b-bb9f-4da67892af40 tempest-ServersTestManualDisk-495750068 tempest-ServersTestManualDisk-495750068-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0e4a7bad5d764f3ab628d80a89e438d3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5d55f6883e2b4020a651c0987fb79f8b', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61957) authorize /opt/stack/nova/nova/policy.py:203}} [ 768.335105] env[61957]: DEBUG nova.network.neutron [-] [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 768.355825] env[61957]: DEBUG oslo_concurrency.lockutils [None req-709951f5-5291-4ca3-b836-f8744d6c583b tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Acquiring lock "7f3850c0-a455-4e5b-a329-fb3b02f9725f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 768.356091] env[61957]: DEBUG oslo_concurrency.lockutils [None req-709951f5-5291-4ca3-b836-f8744d6c583b tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Lock "7f3850c0-a455-4e5b-a329-fb3b02f9725f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 768.356301] env[61957]: DEBUG oslo_concurrency.lockutils [None req-709951f5-5291-4ca3-b836-f8744d6c583b tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Acquiring lock "7f3850c0-a455-4e5b-a329-fb3b02f9725f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 768.356494] env[61957]: DEBUG oslo_concurrency.lockutils [None req-709951f5-5291-4ca3-b836-f8744d6c583b tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Lock "7f3850c0-a455-4e5b-a329-fb3b02f9725f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 768.356654] env[61957]: DEBUG oslo_concurrency.lockutils [None req-709951f5-5291-4ca3-b836-f8744d6c583b tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Lock "7f3850c0-a455-4e5b-a329-fb3b02f9725f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 768.360898] env[61957]: INFO nova.compute.manager [None req-709951f5-5291-4ca3-b836-f8744d6c583b tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] [instance: 7f3850c0-a455-4e5b-a329-fb3b02f9725f] Terminating instance [ 768.363016] env[61957]: DEBUG oslo_concurrency.lockutils [None req-709951f5-5291-4ca3-b836-f8744d6c583b tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Acquiring lock "refresh_cache-7f3850c0-a455-4e5b-a329-fb3b02f9725f" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 768.363174] env[61957]: DEBUG oslo_concurrency.lockutils [None req-709951f5-5291-4ca3-b836-f8744d6c583b tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Acquired lock "refresh_cache-7f3850c0-a455-4e5b-a329-fb3b02f9725f" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 768.363338] env[61957]: DEBUG nova.network.neutron [None req-709951f5-5291-4ca3-b836-f8744d6c583b tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] [instance: 7f3850c0-a455-4e5b-a329-fb3b02f9725f] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 768.505522] env[61957]: INFO nova.scheduler.client.report [None req-5ca5fdda-dbef-4c48-b38f-9d2aa78e9841 tempest-FloatingIPsAssociationNegativeTestJSON-572957506 tempest-FloatingIPsAssociationNegativeTestJSON-572957506-project-member] Deleted allocations for instance 348ea580-2268-4082-a5f1-33c954063e4d [ 768.597186] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f75c0f03-1b9d-4a94-96a7-b1613411b81b {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.605869] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49fc3dd6-0ef3-4735-b7cc-fa9f71f2999e {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.638443] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48e0e3d3-f51e-41ca-948f-0f21442dc8e1 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.650043] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2febf4e-7683-4c21-9aff-1b6e9846dbed {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.666918] env[61957]: DEBUG nova.compute.provider_tree [None req-838e8ad7-4b6c-4dbc-8d4b-8380b652e559 tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 768.677694] env[61957]: DEBUG nova.network.neutron [None req-394366c7-3cf4-481b-bb9f-4da67892af40 tempest-ServersTestManualDisk-495750068 tempest-ServersTestManualDisk-495750068-project-member] [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] Successfully created port: da6abf85-4313-4087-bd58-112359201b6a {{(pid=61957) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 768.739026] env[61957]: DEBUG nova.compute.manager [None req-394366c7-3cf4-481b-bb9f-4da67892af40 tempest-ServersTestManualDisk-495750068 tempest-ServersTestManualDisk-495750068-project-member] [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] Start building block device mappings for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 768.846094] env[61957]: DEBUG nova.network.neutron [-] [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 768.883628] env[61957]: DEBUG nova.network.neutron [None req-709951f5-5291-4ca3-b836-f8744d6c583b tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] [instance: 7f3850c0-a455-4e5b-a329-fb3b02f9725f] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 768.960037] env[61957]: DEBUG nova.network.neutron [None req-709951f5-5291-4ca3-b836-f8744d6c583b tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] [instance: 7f3850c0-a455-4e5b-a329-fb3b02f9725f] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 769.014631] env[61957]: DEBUG oslo_concurrency.lockutils [None req-5ca5fdda-dbef-4c48-b38f-9d2aa78e9841 tempest-FloatingIPsAssociationNegativeTestJSON-572957506 tempest-FloatingIPsAssociationNegativeTestJSON-572957506-project-member] Lock "348ea580-2268-4082-a5f1-33c954063e4d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 136.338s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 769.170617] env[61957]: DEBUG nova.scheduler.client.report [None req-838e8ad7-4b6c-4dbc-8d4b-8380b652e559 tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 769.351999] env[61957]: INFO nova.compute.manager [-] [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] Took 1.04 seconds to deallocate network for instance. [ 769.463682] env[61957]: DEBUG oslo_concurrency.lockutils [None req-709951f5-5291-4ca3-b836-f8744d6c583b tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Releasing lock "refresh_cache-7f3850c0-a455-4e5b-a329-fb3b02f9725f" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 769.464153] env[61957]: DEBUG nova.compute.manager [None req-709951f5-5291-4ca3-b836-f8744d6c583b tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] [instance: 7f3850c0-a455-4e5b-a329-fb3b02f9725f] Start destroying the instance on the hypervisor. {{(pid=61957) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 769.464327] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-709951f5-5291-4ca3-b836-f8744d6c583b tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] [instance: 7f3850c0-a455-4e5b-a329-fb3b02f9725f] Destroying instance {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 769.469934] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0551bc1-8156-4b91-ab03-a960e29ea26b {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.476822] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-709951f5-5291-4ca3-b836-f8744d6c583b tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] [instance: 7f3850c0-a455-4e5b-a329-fb3b02f9725f] Powering off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 769.477097] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c4cfa390-20b2-484a-8413-236069ff0c75 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.484430] env[61957]: DEBUG oslo_vmware.api [None req-709951f5-5291-4ca3-b836-f8744d6c583b tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Waiting for the task: (returnval){ [ 769.484430] env[61957]: value = "task-1277384" [ 769.484430] env[61957]: _type = "Task" [ 769.484430] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 769.493973] env[61957]: DEBUG oslo_vmware.api [None req-709951f5-5291-4ca3-b836-f8744d6c583b tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Task: {'id': task-1277384, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 769.512659] env[61957]: DEBUG nova.compute.manager [req-851144c6-fcd5-49bb-aa8e-5114c55b2304 req-21d0ad6d-ddb2-4c03-b6c6-31da5769b23b service nova] [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] Received event network-changed-da6abf85-4313-4087-bd58-112359201b6a {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 769.512942] env[61957]: DEBUG nova.compute.manager [req-851144c6-fcd5-49bb-aa8e-5114c55b2304 req-21d0ad6d-ddb2-4c03-b6c6-31da5769b23b service nova] [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] Refreshing instance network info cache due to event network-changed-da6abf85-4313-4087-bd58-112359201b6a. {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 769.513306] env[61957]: DEBUG oslo_concurrency.lockutils [req-851144c6-fcd5-49bb-aa8e-5114c55b2304 req-21d0ad6d-ddb2-4c03-b6c6-31da5769b23b service nova] Acquiring lock "refresh_cache-517eeb71-e9e3-4284-a144-dc4b63319d2d" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 769.513527] env[61957]: DEBUG oslo_concurrency.lockutils [req-851144c6-fcd5-49bb-aa8e-5114c55b2304 req-21d0ad6d-ddb2-4c03-b6c6-31da5769b23b service nova] Acquired lock "refresh_cache-517eeb71-e9e3-4284-a144-dc4b63319d2d" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 769.513788] env[61957]: DEBUG nova.network.neutron [req-851144c6-fcd5-49bb-aa8e-5114c55b2304 req-21d0ad6d-ddb2-4c03-b6c6-31da5769b23b service nova] [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] Refreshing network info cache for port da6abf85-4313-4087-bd58-112359201b6a {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 769.516981] env[61957]: DEBUG nova.compute.manager [None req-79f56a2c-6af4-4a0f-b13f-0dc131681bf7 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] Starting instance... {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 769.676982] env[61957]: DEBUG oslo_concurrency.lockutils [None req-838e8ad7-4b6c-4dbc-8d4b-8380b652e559 tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.951s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 769.677862] env[61957]: ERROR nova.compute.manager [None req-838e8ad7-4b6c-4dbc-8d4b-8380b652e559 tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] [instance: 042425a0-5398-47a6-9842-380d5b3299c0] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 7c31a5be-6578-4eb8-833f-0722e63cb7e3, please check neutron logs for more information. [ 769.677862] env[61957]: ERROR nova.compute.manager [instance: 042425a0-5398-47a6-9842-380d5b3299c0] Traceback (most recent call last): [ 769.677862] env[61957]: ERROR nova.compute.manager [instance: 042425a0-5398-47a6-9842-380d5b3299c0] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 769.677862] env[61957]: ERROR nova.compute.manager [instance: 042425a0-5398-47a6-9842-380d5b3299c0] self.driver.spawn(context, instance, image_meta, [ 769.677862] env[61957]: ERROR nova.compute.manager [instance: 042425a0-5398-47a6-9842-380d5b3299c0] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 769.677862] env[61957]: ERROR nova.compute.manager [instance: 042425a0-5398-47a6-9842-380d5b3299c0] self._vmops.spawn(context, instance, image_meta, injected_files, [ 769.677862] env[61957]: ERROR nova.compute.manager [instance: 042425a0-5398-47a6-9842-380d5b3299c0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 769.677862] env[61957]: ERROR nova.compute.manager [instance: 042425a0-5398-47a6-9842-380d5b3299c0] vm_ref = self.build_virtual_machine(instance, [ 769.677862] env[61957]: ERROR nova.compute.manager [instance: 042425a0-5398-47a6-9842-380d5b3299c0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 769.677862] env[61957]: ERROR nova.compute.manager [instance: 042425a0-5398-47a6-9842-380d5b3299c0] vif_infos = vmwarevif.get_vif_info(self._session, [ 769.677862] env[61957]: ERROR nova.compute.manager [instance: 042425a0-5398-47a6-9842-380d5b3299c0] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 769.678241] env[61957]: ERROR nova.compute.manager [instance: 042425a0-5398-47a6-9842-380d5b3299c0] for vif in network_info: [ 769.678241] env[61957]: ERROR nova.compute.manager [instance: 042425a0-5398-47a6-9842-380d5b3299c0] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 769.678241] env[61957]: ERROR nova.compute.manager [instance: 042425a0-5398-47a6-9842-380d5b3299c0] return self._sync_wrapper(fn, *args, **kwargs) [ 769.678241] env[61957]: ERROR nova.compute.manager [instance: 042425a0-5398-47a6-9842-380d5b3299c0] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 769.678241] env[61957]: ERROR nova.compute.manager [instance: 042425a0-5398-47a6-9842-380d5b3299c0] self.wait() [ 769.678241] env[61957]: ERROR nova.compute.manager [instance: 042425a0-5398-47a6-9842-380d5b3299c0] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 769.678241] env[61957]: ERROR nova.compute.manager [instance: 042425a0-5398-47a6-9842-380d5b3299c0] self[:] = self._gt.wait() [ 769.678241] env[61957]: ERROR nova.compute.manager [instance: 042425a0-5398-47a6-9842-380d5b3299c0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 769.678241] env[61957]: ERROR nova.compute.manager [instance: 042425a0-5398-47a6-9842-380d5b3299c0] return self._exit_event.wait() [ 769.678241] env[61957]: ERROR nova.compute.manager [instance: 042425a0-5398-47a6-9842-380d5b3299c0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 769.678241] env[61957]: ERROR nova.compute.manager [instance: 042425a0-5398-47a6-9842-380d5b3299c0] result = hub.switch() [ 769.678241] env[61957]: ERROR nova.compute.manager [instance: 042425a0-5398-47a6-9842-380d5b3299c0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 769.678241] env[61957]: ERROR nova.compute.manager [instance: 042425a0-5398-47a6-9842-380d5b3299c0] return self.greenlet.switch() [ 769.678680] env[61957]: ERROR nova.compute.manager [instance: 042425a0-5398-47a6-9842-380d5b3299c0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 769.678680] env[61957]: ERROR nova.compute.manager [instance: 042425a0-5398-47a6-9842-380d5b3299c0] result = function(*args, **kwargs) [ 769.678680] env[61957]: ERROR nova.compute.manager [instance: 042425a0-5398-47a6-9842-380d5b3299c0] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 769.678680] env[61957]: ERROR nova.compute.manager [instance: 042425a0-5398-47a6-9842-380d5b3299c0] return func(*args, **kwargs) [ 769.678680] env[61957]: ERROR nova.compute.manager [instance: 042425a0-5398-47a6-9842-380d5b3299c0] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 769.678680] env[61957]: ERROR nova.compute.manager [instance: 042425a0-5398-47a6-9842-380d5b3299c0] raise e [ 769.678680] env[61957]: ERROR nova.compute.manager [instance: 042425a0-5398-47a6-9842-380d5b3299c0] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 769.678680] env[61957]: ERROR nova.compute.manager [instance: 042425a0-5398-47a6-9842-380d5b3299c0] nwinfo = self.network_api.allocate_for_instance( [ 769.678680] env[61957]: ERROR nova.compute.manager [instance: 042425a0-5398-47a6-9842-380d5b3299c0] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 769.678680] env[61957]: ERROR nova.compute.manager [instance: 042425a0-5398-47a6-9842-380d5b3299c0] created_port_ids = self._update_ports_for_instance( [ 769.678680] env[61957]: ERROR nova.compute.manager [instance: 042425a0-5398-47a6-9842-380d5b3299c0] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 769.678680] env[61957]: ERROR nova.compute.manager [instance: 042425a0-5398-47a6-9842-380d5b3299c0] with excutils.save_and_reraise_exception(): [ 769.678680] env[61957]: ERROR nova.compute.manager [instance: 042425a0-5398-47a6-9842-380d5b3299c0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 769.679184] env[61957]: ERROR nova.compute.manager [instance: 042425a0-5398-47a6-9842-380d5b3299c0] self.force_reraise() [ 769.679184] env[61957]: ERROR nova.compute.manager [instance: 042425a0-5398-47a6-9842-380d5b3299c0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 769.679184] env[61957]: ERROR nova.compute.manager [instance: 042425a0-5398-47a6-9842-380d5b3299c0] raise self.value [ 769.679184] env[61957]: ERROR nova.compute.manager [instance: 042425a0-5398-47a6-9842-380d5b3299c0] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 769.679184] env[61957]: ERROR nova.compute.manager [instance: 042425a0-5398-47a6-9842-380d5b3299c0] updated_port = self._update_port( [ 769.679184] env[61957]: ERROR nova.compute.manager [instance: 042425a0-5398-47a6-9842-380d5b3299c0] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 769.679184] env[61957]: ERROR nova.compute.manager [instance: 042425a0-5398-47a6-9842-380d5b3299c0] _ensure_no_port_binding_failure(port) [ 769.679184] env[61957]: ERROR nova.compute.manager [instance: 042425a0-5398-47a6-9842-380d5b3299c0] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 769.679184] env[61957]: ERROR nova.compute.manager [instance: 042425a0-5398-47a6-9842-380d5b3299c0] raise exception.PortBindingFailed(port_id=port['id']) [ 769.679184] env[61957]: ERROR nova.compute.manager [instance: 042425a0-5398-47a6-9842-380d5b3299c0] nova.exception.PortBindingFailed: Binding failed for port 7c31a5be-6578-4eb8-833f-0722e63cb7e3, please check neutron logs for more information. [ 769.679184] env[61957]: ERROR nova.compute.manager [instance: 042425a0-5398-47a6-9842-380d5b3299c0] [ 769.679524] env[61957]: DEBUG nova.compute.utils [None req-838e8ad7-4b6c-4dbc-8d4b-8380b652e559 tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] [instance: 042425a0-5398-47a6-9842-380d5b3299c0] Binding failed for port 7c31a5be-6578-4eb8-833f-0722e63cb7e3, please check neutron logs for more information. {{(pid=61957) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 769.680184] env[61957]: DEBUG oslo_concurrency.lockutils [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 16.357s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 769.680380] env[61957]: DEBUG oslo_concurrency.lockutils [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 769.680575] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61957) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 769.680974] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d6c9937c-2526-4768-bade-4ca2ced8b569 tempest-ServersTestFqdnHostnames-1308211839 tempest-ServersTestFqdnHostnames-1308211839-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.649s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 769.682807] env[61957]: INFO nova.compute.claims [None req-d6c9937c-2526-4768-bade-4ca2ced8b569 tempest-ServersTestFqdnHostnames-1308211839 tempest-ServersTestFqdnHostnames-1308211839-project-member] [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 769.686241] env[61957]: DEBUG nova.compute.manager [None req-838e8ad7-4b6c-4dbc-8d4b-8380b652e559 tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] [instance: 042425a0-5398-47a6-9842-380d5b3299c0] Build of instance 042425a0-5398-47a6-9842-380d5b3299c0 was re-scheduled: Binding failed for port 7c31a5be-6578-4eb8-833f-0722e63cb7e3, please check neutron logs for more information. {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 769.686713] env[61957]: DEBUG nova.compute.manager [None req-838e8ad7-4b6c-4dbc-8d4b-8380b652e559 tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] [instance: 042425a0-5398-47a6-9842-380d5b3299c0] Unplugging VIFs for instance {{(pid=61957) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 769.687020] env[61957]: DEBUG oslo_concurrency.lockutils [None req-838e8ad7-4b6c-4dbc-8d4b-8380b652e559 tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] Acquiring lock "refresh_cache-042425a0-5398-47a6-9842-380d5b3299c0" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 769.687184] env[61957]: DEBUG oslo_concurrency.lockutils [None req-838e8ad7-4b6c-4dbc-8d4b-8380b652e559 tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] Acquired lock "refresh_cache-042425a0-5398-47a6-9842-380d5b3299c0" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 769.687383] env[61957]: DEBUG nova.network.neutron [None req-838e8ad7-4b6c-4dbc-8d4b-8380b652e559 tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] [instance: 042425a0-5398-47a6-9842-380d5b3299c0] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 769.689392] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6d71d74-567d-4cc8-a3d8-62131abd96be {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.698962] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3250d314-c669-42e6-ac0e-9772d9925909 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.714804] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e0400e4-2093-48b0-b23a-94b2e0c9e249 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.721787] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efb08843-5783-45ba-bb81-d1f21749b35a {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.753975] env[61957]: DEBUG nova.compute.manager [None req-394366c7-3cf4-481b-bb9f-4da67892af40 tempest-ServersTestManualDisk-495750068 tempest-ServersTestManualDisk-495750068-project-member] [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] Start spawning the instance on the hypervisor. {{(pid=61957) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 769.755902] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181413MB free_disk=142GB free_vcpus=48 pci_devices=None {{(pid=61957) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 769.756022] env[61957]: DEBUG oslo_concurrency.lockutils [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 769.780558] env[61957]: DEBUG nova.virt.hardware [None req-394366c7-3cf4-481b-bb9f-4da67892af40 tempest-ServersTestManualDisk-495750068 tempest-ServersTestManualDisk-495750068-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T17:22:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T17:21:50Z,direct_url=,disk_format='vmdk',id=11c76a2c-f705-470a-ba9d-4657858bab38,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='81d3096d0f094373913fc3dc8f5c3c6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T17:21:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 769.780797] env[61957]: DEBUG nova.virt.hardware [None req-394366c7-3cf4-481b-bb9f-4da67892af40 tempest-ServersTestManualDisk-495750068 tempest-ServersTestManualDisk-495750068-project-member] Flavor limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 769.780949] env[61957]: DEBUG nova.virt.hardware [None req-394366c7-3cf4-481b-bb9f-4da67892af40 tempest-ServersTestManualDisk-495750068 tempest-ServersTestManualDisk-495750068-project-member] Image limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 769.781445] env[61957]: DEBUG nova.virt.hardware [None req-394366c7-3cf4-481b-bb9f-4da67892af40 tempest-ServersTestManualDisk-495750068 tempest-ServersTestManualDisk-495750068-project-member] Flavor pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 769.781519] env[61957]: DEBUG nova.virt.hardware [None req-394366c7-3cf4-481b-bb9f-4da67892af40 tempest-ServersTestManualDisk-495750068 tempest-ServersTestManualDisk-495750068-project-member] Image pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 769.781662] env[61957]: DEBUG nova.virt.hardware [None req-394366c7-3cf4-481b-bb9f-4da67892af40 tempest-ServersTestManualDisk-495750068 tempest-ServersTestManualDisk-495750068-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 769.781873] env[61957]: DEBUG nova.virt.hardware [None req-394366c7-3cf4-481b-bb9f-4da67892af40 tempest-ServersTestManualDisk-495750068 tempest-ServersTestManualDisk-495750068-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 769.782039] env[61957]: DEBUG nova.virt.hardware [None req-394366c7-3cf4-481b-bb9f-4da67892af40 tempest-ServersTestManualDisk-495750068 tempest-ServersTestManualDisk-495750068-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 769.782206] env[61957]: DEBUG nova.virt.hardware [None req-394366c7-3cf4-481b-bb9f-4da67892af40 tempest-ServersTestManualDisk-495750068 tempest-ServersTestManualDisk-495750068-project-member] Got 1 possible topologies {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 769.782365] env[61957]: DEBUG nova.virt.hardware [None req-394366c7-3cf4-481b-bb9f-4da67892af40 tempest-ServersTestManualDisk-495750068 tempest-ServersTestManualDisk-495750068-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 769.782619] env[61957]: DEBUG nova.virt.hardware [None req-394366c7-3cf4-481b-bb9f-4da67892af40 tempest-ServersTestManualDisk-495750068 tempest-ServersTestManualDisk-495750068-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 769.783611] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-503d1428-03d4-4e1d-98a9-a788d9dbc9ed {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.792478] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c81119a-f089-43da-9e12-32e834b3e5f8 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.816828] env[61957]: ERROR nova.compute.manager [None req-394366c7-3cf4-481b-bb9f-4da67892af40 tempest-ServersTestManualDisk-495750068 tempest-ServersTestManualDisk-495750068-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port da6abf85-4313-4087-bd58-112359201b6a, please check neutron logs for more information. [ 769.816828] env[61957]: ERROR nova.compute.manager Traceback (most recent call last): [ 769.816828] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 769.816828] env[61957]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 769.816828] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 769.816828] env[61957]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 769.816828] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 769.816828] env[61957]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 769.816828] env[61957]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 769.816828] env[61957]: ERROR nova.compute.manager self.force_reraise() [ 769.816828] env[61957]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 769.816828] env[61957]: ERROR nova.compute.manager raise self.value [ 769.816828] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 769.816828] env[61957]: ERROR nova.compute.manager updated_port = self._update_port( [ 769.816828] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 769.816828] env[61957]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 769.817666] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 769.817666] env[61957]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 769.817666] env[61957]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port da6abf85-4313-4087-bd58-112359201b6a, please check neutron logs for more information. [ 769.817666] env[61957]: ERROR nova.compute.manager [ 769.817666] env[61957]: Traceback (most recent call last): [ 769.817666] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 769.817666] env[61957]: listener.cb(fileno) [ 769.817666] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 769.817666] env[61957]: result = function(*args, **kwargs) [ 769.817666] env[61957]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 769.817666] env[61957]: return func(*args, **kwargs) [ 769.817666] env[61957]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 769.817666] env[61957]: raise e [ 769.817666] env[61957]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 769.817666] env[61957]: nwinfo = self.network_api.allocate_for_instance( [ 769.817666] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 769.817666] env[61957]: created_port_ids = self._update_ports_for_instance( [ 769.817666] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 769.817666] env[61957]: with excutils.save_and_reraise_exception(): [ 769.817666] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 769.817666] env[61957]: self.force_reraise() [ 769.817666] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 769.817666] env[61957]: raise self.value [ 769.817666] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 769.817666] env[61957]: updated_port = self._update_port( [ 769.817666] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 769.817666] env[61957]: _ensure_no_port_binding_failure(port) [ 769.817666] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 769.817666] env[61957]: raise exception.PortBindingFailed(port_id=port['id']) [ 769.818819] env[61957]: nova.exception.PortBindingFailed: Binding failed for port da6abf85-4313-4087-bd58-112359201b6a, please check neutron logs for more information. [ 769.818819] env[61957]: Removing descriptor: 16 [ 769.818819] env[61957]: ERROR nova.compute.manager [None req-394366c7-3cf4-481b-bb9f-4da67892af40 tempest-ServersTestManualDisk-495750068 tempest-ServersTestManualDisk-495750068-project-member] [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port da6abf85-4313-4087-bd58-112359201b6a, please check neutron logs for more information. [ 769.818819] env[61957]: ERROR nova.compute.manager [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] Traceback (most recent call last): [ 769.818819] env[61957]: ERROR nova.compute.manager [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 769.818819] env[61957]: ERROR nova.compute.manager [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] yield resources [ 769.818819] env[61957]: ERROR nova.compute.manager [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 769.818819] env[61957]: ERROR nova.compute.manager [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] self.driver.spawn(context, instance, image_meta, [ 769.818819] env[61957]: ERROR nova.compute.manager [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 769.818819] env[61957]: ERROR nova.compute.manager [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 769.818819] env[61957]: ERROR nova.compute.manager [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 769.818819] env[61957]: ERROR nova.compute.manager [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] vm_ref = self.build_virtual_machine(instance, [ 769.819342] env[61957]: ERROR nova.compute.manager [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 769.819342] env[61957]: ERROR nova.compute.manager [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] vif_infos = vmwarevif.get_vif_info(self._session, [ 769.819342] env[61957]: ERROR nova.compute.manager [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 769.819342] env[61957]: ERROR nova.compute.manager [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] for vif in network_info: [ 769.819342] env[61957]: ERROR nova.compute.manager [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 769.819342] env[61957]: ERROR nova.compute.manager [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] return self._sync_wrapper(fn, *args, **kwargs) [ 769.819342] env[61957]: ERROR nova.compute.manager [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 769.819342] env[61957]: ERROR nova.compute.manager [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] self.wait() [ 769.819342] env[61957]: ERROR nova.compute.manager [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 769.819342] env[61957]: ERROR nova.compute.manager [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] self[:] = self._gt.wait() [ 769.819342] env[61957]: ERROR nova.compute.manager [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 769.819342] env[61957]: ERROR nova.compute.manager [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] return self._exit_event.wait() [ 769.819342] env[61957]: ERROR nova.compute.manager [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 769.819713] env[61957]: ERROR nova.compute.manager [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] result = hub.switch() [ 769.819713] env[61957]: ERROR nova.compute.manager [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 769.819713] env[61957]: ERROR nova.compute.manager [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] return self.greenlet.switch() [ 769.819713] env[61957]: ERROR nova.compute.manager [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 769.819713] env[61957]: ERROR nova.compute.manager [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] result = function(*args, **kwargs) [ 769.819713] env[61957]: ERROR nova.compute.manager [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 769.819713] env[61957]: ERROR nova.compute.manager [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] return func(*args, **kwargs) [ 769.819713] env[61957]: ERROR nova.compute.manager [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 769.819713] env[61957]: ERROR nova.compute.manager [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] raise e [ 769.819713] env[61957]: ERROR nova.compute.manager [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 769.819713] env[61957]: ERROR nova.compute.manager [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] nwinfo = self.network_api.allocate_for_instance( [ 769.819713] env[61957]: ERROR nova.compute.manager [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 769.819713] env[61957]: ERROR nova.compute.manager [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] created_port_ids = self._update_ports_for_instance( [ 769.820171] env[61957]: ERROR nova.compute.manager [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 769.820171] env[61957]: ERROR nova.compute.manager [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] with excutils.save_and_reraise_exception(): [ 769.820171] env[61957]: ERROR nova.compute.manager [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 769.820171] env[61957]: ERROR nova.compute.manager [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] self.force_reraise() [ 769.820171] env[61957]: ERROR nova.compute.manager [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 769.820171] env[61957]: ERROR nova.compute.manager [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] raise self.value [ 769.820171] env[61957]: ERROR nova.compute.manager [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 769.820171] env[61957]: ERROR nova.compute.manager [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] updated_port = self._update_port( [ 769.820171] env[61957]: ERROR nova.compute.manager [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 769.820171] env[61957]: ERROR nova.compute.manager [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] _ensure_no_port_binding_failure(port) [ 769.820171] env[61957]: ERROR nova.compute.manager [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 769.820171] env[61957]: ERROR nova.compute.manager [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] raise exception.PortBindingFailed(port_id=port['id']) [ 769.820524] env[61957]: ERROR nova.compute.manager [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] nova.exception.PortBindingFailed: Binding failed for port da6abf85-4313-4087-bd58-112359201b6a, please check neutron logs for more information. [ 769.820524] env[61957]: ERROR nova.compute.manager [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] [ 769.820524] env[61957]: INFO nova.compute.manager [None req-394366c7-3cf4-481b-bb9f-4da67892af40 tempest-ServersTestManualDisk-495750068 tempest-ServersTestManualDisk-495750068-project-member] [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] Terminating instance [ 769.820524] env[61957]: DEBUG oslo_concurrency.lockutils [None req-394366c7-3cf4-481b-bb9f-4da67892af40 tempest-ServersTestManualDisk-495750068 tempest-ServersTestManualDisk-495750068-project-member] Acquiring lock "refresh_cache-517eeb71-e9e3-4284-a144-dc4b63319d2d" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 769.902399] env[61957]: INFO nova.compute.manager [None req-3d8587e3-7ad5-40a6-88af-e8ed44c49baa tempest-ServerActionsV293TestJSON-1269989829 tempest-ServerActionsV293TestJSON-1269989829-project-member] [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] Took 0.55 seconds to detach 1 volumes for instance. [ 769.904485] env[61957]: DEBUG nova.compute.claims [None req-3d8587e3-7ad5-40a6-88af-e8ed44c49baa tempest-ServerActionsV293TestJSON-1269989829 tempest-ServerActionsV293TestJSON-1269989829-project-member] [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] Aborting claim: {{(pid=61957) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 769.904705] env[61957]: DEBUG oslo_concurrency.lockutils [None req-3d8587e3-7ad5-40a6-88af-e8ed44c49baa tempest-ServerActionsV293TestJSON-1269989829 tempest-ServerActionsV293TestJSON-1269989829-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 769.997230] env[61957]: DEBUG oslo_vmware.api [None req-709951f5-5291-4ca3-b836-f8744d6c583b tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Task: {'id': task-1277384, 'name': PowerOffVM_Task, 'duration_secs': 0.141256} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 769.997562] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-709951f5-5291-4ca3-b836-f8744d6c583b tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] [instance: 7f3850c0-a455-4e5b-a329-fb3b02f9725f] Powered off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 769.998842] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-709951f5-5291-4ca3-b836-f8744d6c583b tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] [instance: 7f3850c0-a455-4e5b-a329-fb3b02f9725f] Unregistering the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 769.999208] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b9cc85dc-d37f-461b-8609-22a8b9f0f216 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.029165] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-709951f5-5291-4ca3-b836-f8744d6c583b tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] [instance: 7f3850c0-a455-4e5b-a329-fb3b02f9725f] Unregistered the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 770.029272] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-709951f5-5291-4ca3-b836-f8744d6c583b tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] [instance: 7f3850c0-a455-4e5b-a329-fb3b02f9725f] Deleting contents of the VM from datastore datastore2 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 770.029512] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-709951f5-5291-4ca3-b836-f8744d6c583b tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Deleting the datastore file [datastore2] 7f3850c0-a455-4e5b-a329-fb3b02f9725f {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 770.029789] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-45230838-8027-41f7-a6c1-0925143b2081 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.037807] env[61957]: DEBUG oslo_vmware.api [None req-709951f5-5291-4ca3-b836-f8744d6c583b tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Waiting for the task: (returnval){ [ 770.037807] env[61957]: value = "task-1277386" [ 770.037807] env[61957]: _type = "Task" [ 770.037807] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 770.047162] env[61957]: DEBUG oslo_vmware.api [None req-709951f5-5291-4ca3-b836-f8744d6c583b tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Task: {'id': task-1277386, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 770.048288] env[61957]: DEBUG oslo_concurrency.lockutils [None req-79f56a2c-6af4-4a0f-b13f-0dc131681bf7 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 770.049116] env[61957]: DEBUG nova.network.neutron [req-851144c6-fcd5-49bb-aa8e-5114c55b2304 req-21d0ad6d-ddb2-4c03-b6c6-31da5769b23b service nova] [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 770.175201] env[61957]: DEBUG nova.network.neutron [req-851144c6-fcd5-49bb-aa8e-5114c55b2304 req-21d0ad6d-ddb2-4c03-b6c6-31da5769b23b service nova] [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 770.211757] env[61957]: DEBUG nova.network.neutron [None req-838e8ad7-4b6c-4dbc-8d4b-8380b652e559 tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] [instance: 042425a0-5398-47a6-9842-380d5b3299c0] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 770.274398] env[61957]: DEBUG nova.network.neutron [None req-838e8ad7-4b6c-4dbc-8d4b-8380b652e559 tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] [instance: 042425a0-5398-47a6-9842-380d5b3299c0] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 770.552795] env[61957]: DEBUG oslo_vmware.api [None req-709951f5-5291-4ca3-b836-f8744d6c583b tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Task: {'id': task-1277386, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.112586} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 770.553104] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-709951f5-5291-4ca3-b836-f8744d6c583b tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Deleted the datastore file {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 770.553324] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-709951f5-5291-4ca3-b836-f8744d6c583b tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] [instance: 7f3850c0-a455-4e5b-a329-fb3b02f9725f] Deleted contents of the VM from datastore datastore2 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 770.553548] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-709951f5-5291-4ca3-b836-f8744d6c583b tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] [instance: 7f3850c0-a455-4e5b-a329-fb3b02f9725f] Instance destroyed {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 770.553883] env[61957]: INFO nova.compute.manager [None req-709951f5-5291-4ca3-b836-f8744d6c583b tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] [instance: 7f3850c0-a455-4e5b-a329-fb3b02f9725f] Took 1.09 seconds to destroy the instance on the hypervisor. [ 770.554042] env[61957]: DEBUG oslo.service.loopingcall [None req-709951f5-5291-4ca3-b836-f8744d6c583b tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 770.554265] env[61957]: DEBUG nova.compute.manager [-] [instance: 7f3850c0-a455-4e5b-a329-fb3b02f9725f] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 770.554371] env[61957]: DEBUG nova.network.neutron [-] [instance: 7f3850c0-a455-4e5b-a329-fb3b02f9725f] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 770.571533] env[61957]: DEBUG nova.network.neutron [-] [instance: 7f3850c0-a455-4e5b-a329-fb3b02f9725f] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 770.676567] env[61957]: DEBUG oslo_concurrency.lockutils [req-851144c6-fcd5-49bb-aa8e-5114c55b2304 req-21d0ad6d-ddb2-4c03-b6c6-31da5769b23b service nova] Releasing lock "refresh_cache-517eeb71-e9e3-4284-a144-dc4b63319d2d" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 770.677105] env[61957]: DEBUG oslo_concurrency.lockutils [None req-394366c7-3cf4-481b-bb9f-4da67892af40 tempest-ServersTestManualDisk-495750068 tempest-ServersTestManualDisk-495750068-project-member] Acquired lock "refresh_cache-517eeb71-e9e3-4284-a144-dc4b63319d2d" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 770.677336] env[61957]: DEBUG nova.network.neutron [None req-394366c7-3cf4-481b-bb9f-4da67892af40 tempest-ServersTestManualDisk-495750068 tempest-ServersTestManualDisk-495750068-project-member] [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 770.779607] env[61957]: DEBUG oslo_concurrency.lockutils [None req-838e8ad7-4b6c-4dbc-8d4b-8380b652e559 tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] Releasing lock "refresh_cache-042425a0-5398-47a6-9842-380d5b3299c0" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 770.779869] env[61957]: DEBUG nova.compute.manager [None req-838e8ad7-4b6c-4dbc-8d4b-8380b652e559 tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61957) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 770.780097] env[61957]: DEBUG nova.compute.manager [None req-838e8ad7-4b6c-4dbc-8d4b-8380b652e559 tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] [instance: 042425a0-5398-47a6-9842-380d5b3299c0] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 770.780868] env[61957]: DEBUG nova.network.neutron [None req-838e8ad7-4b6c-4dbc-8d4b-8380b652e559 tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] [instance: 042425a0-5398-47a6-9842-380d5b3299c0] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 770.801254] env[61957]: DEBUG nova.network.neutron [None req-838e8ad7-4b6c-4dbc-8d4b-8380b652e559 tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] [instance: 042425a0-5398-47a6-9842-380d5b3299c0] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 771.067814] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62c1cbe8-a1ca-4b55-b0e6-5c069e056340 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.074383] env[61957]: DEBUG nova.network.neutron [-] [instance: 7f3850c0-a455-4e5b-a329-fb3b02f9725f] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 771.075963] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd808e20-5a93-4984-adbb-5225cc55c68a {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.106143] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58cc4b7d-5efd-4910-865d-0105cf5b9019 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.114258] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81aa9991-2691-4d63-83e7-8880b3ce7216 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.127790] env[61957]: DEBUG nova.compute.provider_tree [None req-d6c9937c-2526-4768-bade-4ca2ced8b569 tempest-ServersTestFqdnHostnames-1308211839 tempest-ServersTestFqdnHostnames-1308211839-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 771.194497] env[61957]: DEBUG nova.network.neutron [None req-394366c7-3cf4-481b-bb9f-4da67892af40 tempest-ServersTestManualDisk-495750068 tempest-ServersTestManualDisk-495750068-project-member] [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 771.276591] env[61957]: DEBUG nova.network.neutron [None req-394366c7-3cf4-481b-bb9f-4da67892af40 tempest-ServersTestManualDisk-495750068 tempest-ServersTestManualDisk-495750068-project-member] [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 771.307029] env[61957]: DEBUG nova.network.neutron [None req-838e8ad7-4b6c-4dbc-8d4b-8380b652e559 tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] [instance: 042425a0-5398-47a6-9842-380d5b3299c0] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 771.582471] env[61957]: INFO nova.compute.manager [-] [instance: 7f3850c0-a455-4e5b-a329-fb3b02f9725f] Took 1.03 seconds to deallocate network for instance. [ 771.619839] env[61957]: DEBUG nova.compute.manager [req-7dfed62f-6bb2-46fa-b13f-33c9794cc917 req-7811467b-31c8-411d-b517-155e8c85b159 service nova] [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] Received event network-vif-deleted-da6abf85-4313-4087-bd58-112359201b6a {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 771.632487] env[61957]: DEBUG nova.scheduler.client.report [None req-d6c9937c-2526-4768-bade-4ca2ced8b569 tempest-ServersTestFqdnHostnames-1308211839 tempest-ServersTestFqdnHostnames-1308211839-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 771.782409] env[61957]: DEBUG oslo_concurrency.lockutils [None req-394366c7-3cf4-481b-bb9f-4da67892af40 tempest-ServersTestManualDisk-495750068 tempest-ServersTestManualDisk-495750068-project-member] Releasing lock "refresh_cache-517eeb71-e9e3-4284-a144-dc4b63319d2d" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 771.782409] env[61957]: DEBUG nova.compute.manager [None req-394366c7-3cf4-481b-bb9f-4da67892af40 tempest-ServersTestManualDisk-495750068 tempest-ServersTestManualDisk-495750068-project-member] [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] Start destroying the instance on the hypervisor. {{(pid=61957) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 771.782409] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-394366c7-3cf4-481b-bb9f-4da67892af40 tempest-ServersTestManualDisk-495750068 tempest-ServersTestManualDisk-495750068-project-member] [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] Destroying instance {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 771.782409] env[61957]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f1a0e1c3-252f-4cc9-8bd2-8235153ea82b {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.790315] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3ddffcf-9df9-4fd4-a9ca-98ce246cdf3f {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.810754] env[61957]: INFO nova.compute.manager [None req-838e8ad7-4b6c-4dbc-8d4b-8380b652e559 tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] [instance: 042425a0-5398-47a6-9842-380d5b3299c0] Took 1.03 seconds to deallocate network for instance. [ 771.814131] env[61957]: WARNING nova.virt.vmwareapi.vmops [None req-394366c7-3cf4-481b-bb9f-4da67892af40 tempest-ServersTestManualDisk-495750068 tempest-ServersTestManualDisk-495750068-project-member] [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 517eeb71-e9e3-4284-a144-dc4b63319d2d could not be found. [ 771.814349] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-394366c7-3cf4-481b-bb9f-4da67892af40 tempest-ServersTestManualDisk-495750068 tempest-ServersTestManualDisk-495750068-project-member] [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] Instance destroyed {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 771.814548] env[61957]: INFO nova.compute.manager [None req-394366c7-3cf4-481b-bb9f-4da67892af40 tempest-ServersTestManualDisk-495750068 tempest-ServersTestManualDisk-495750068-project-member] [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] Took 0.03 seconds to destroy the instance on the hypervisor. [ 771.814782] env[61957]: DEBUG oslo.service.loopingcall [None req-394366c7-3cf4-481b-bb9f-4da67892af40 tempest-ServersTestManualDisk-495750068 tempest-ServersTestManualDisk-495750068-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 771.815229] env[61957]: DEBUG nova.compute.manager [-] [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 771.815332] env[61957]: DEBUG nova.network.neutron [-] [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 771.830171] env[61957]: DEBUG nova.network.neutron [-] [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 772.091508] env[61957]: DEBUG oslo_concurrency.lockutils [None req-709951f5-5291-4ca3-b836-f8744d6c583b tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 772.136526] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d6c9937c-2526-4768-bade-4ca2ced8b569 tempest-ServersTestFqdnHostnames-1308211839 tempest-ServersTestFqdnHostnames-1308211839-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.455s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 772.137077] env[61957]: DEBUG nova.compute.manager [None req-d6c9937c-2526-4768-bade-4ca2ced8b569 tempest-ServersTestFqdnHostnames-1308211839 tempest-ServersTestFqdnHostnames-1308211839-project-member] [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] Start building networks asynchronously for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 772.139825] env[61957]: DEBUG oslo_concurrency.lockutils [None req-bb8cb37b-74c5-4607-9ba8-e5615134a1da tempest-InstanceActionsTestJSON-565478614 tempest-InstanceActionsTestJSON-565478614-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.766s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 772.141184] env[61957]: INFO nova.compute.claims [None req-bb8cb37b-74c5-4607-9ba8-e5615134a1da tempest-InstanceActionsTestJSON-565478614 tempest-InstanceActionsTestJSON-565478614-project-member] [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 772.335630] env[61957]: DEBUG nova.network.neutron [-] [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 772.645409] env[61957]: DEBUG nova.compute.utils [None req-d6c9937c-2526-4768-bade-4ca2ced8b569 tempest-ServersTestFqdnHostnames-1308211839 tempest-ServersTestFqdnHostnames-1308211839-project-member] Using /dev/sd instead of None {{(pid=61957) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 772.648470] env[61957]: DEBUG nova.compute.manager [None req-d6c9937c-2526-4768-bade-4ca2ced8b569 tempest-ServersTestFqdnHostnames-1308211839 tempest-ServersTestFqdnHostnames-1308211839-project-member] [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] Allocating IP information in the background. {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 772.648802] env[61957]: DEBUG nova.network.neutron [None req-d6c9937c-2526-4768-bade-4ca2ced8b569 tempest-ServersTestFqdnHostnames-1308211839 tempest-ServersTestFqdnHostnames-1308211839-project-member] [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] allocate_for_instance() {{(pid=61957) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 772.702405] env[61957]: DEBUG nova.policy [None req-d6c9937c-2526-4768-bade-4ca2ced8b569 tempest-ServersTestFqdnHostnames-1308211839 tempest-ServersTestFqdnHostnames-1308211839-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c10bbf10d15448cb90a071142b019f5c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3bc7ec9dab42411ab69361c587fae686', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61957) authorize /opt/stack/nova/nova/policy.py:203}} [ 772.838331] env[61957]: INFO nova.compute.manager [-] [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] Took 1.02 seconds to deallocate network for instance. [ 772.840920] env[61957]: DEBUG nova.compute.claims [None req-394366c7-3cf4-481b-bb9f-4da67892af40 tempest-ServersTestManualDisk-495750068 tempest-ServersTestManualDisk-495750068-project-member] [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] Aborting claim: {{(pid=61957) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 772.841225] env[61957]: DEBUG oslo_concurrency.lockutils [None req-394366c7-3cf4-481b-bb9f-4da67892af40 tempest-ServersTestManualDisk-495750068 tempest-ServersTestManualDisk-495750068-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 772.844251] env[61957]: INFO nova.scheduler.client.report [None req-838e8ad7-4b6c-4dbc-8d4b-8380b652e559 tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] Deleted allocations for instance 042425a0-5398-47a6-9842-380d5b3299c0 [ 772.988221] env[61957]: DEBUG nova.network.neutron [None req-d6c9937c-2526-4768-bade-4ca2ced8b569 tempest-ServersTestFqdnHostnames-1308211839 tempest-ServersTestFqdnHostnames-1308211839-project-member] [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] Successfully created port: ad48de80-9219-4f81-8c65-da8fb5df2729 {{(pid=61957) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 773.154196] env[61957]: DEBUG nova.compute.manager [None req-d6c9937c-2526-4768-bade-4ca2ced8b569 tempest-ServersTestFqdnHostnames-1308211839 tempest-ServersTestFqdnHostnames-1308211839-project-member] [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] Start building block device mappings for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 773.351528] env[61957]: DEBUG oslo_concurrency.lockutils [None req-838e8ad7-4b6c-4dbc-8d4b-8380b652e559 tempest-MigrationsAdminTest-557755407 tempest-MigrationsAdminTest-557755407-project-member] Lock "042425a0-5398-47a6-9842-380d5b3299c0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 139.162s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 773.463443] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cbce6dd-6e78-47b1-94f2-eac7217c15b1 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.470619] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-608a542a-5925-4458-9858-206c49d31f15 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.500584] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2f69054-0e42-4370-b531-78cfe1b7145a {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.507477] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7bafc8a-0a18-4a4e-af41-bc60507990c4 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.520174] env[61957]: DEBUG nova.compute.provider_tree [None req-bb8cb37b-74c5-4607-9ba8-e5615134a1da tempest-InstanceActionsTestJSON-565478614 tempest-InstanceActionsTestJSON-565478614-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 773.855242] env[61957]: DEBUG nova.compute.manager [None req-da68b092-4602-4484-b1b0-adcac4aae9f8 tempest-InstanceActionsV221TestJSON-1768084888 tempest-InstanceActionsV221TestJSON-1768084888-project-member] [instance: 6e7bd89c-2c2a-450a-9858-3526d96c28ab] Starting instance... {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 773.956278] env[61957]: DEBUG nova.compute.manager [req-a528ea5f-fad7-4bbf-b8fb-0d88738863c4 req-63af1110-e24c-414f-bd9f-8fd5d48f9787 service nova] [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] Received event network-changed-ad48de80-9219-4f81-8c65-da8fb5df2729 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 773.956466] env[61957]: DEBUG nova.compute.manager [req-a528ea5f-fad7-4bbf-b8fb-0d88738863c4 req-63af1110-e24c-414f-bd9f-8fd5d48f9787 service nova] [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] Refreshing instance network info cache due to event network-changed-ad48de80-9219-4f81-8c65-da8fb5df2729. {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 773.956682] env[61957]: DEBUG oslo_concurrency.lockutils [req-a528ea5f-fad7-4bbf-b8fb-0d88738863c4 req-63af1110-e24c-414f-bd9f-8fd5d48f9787 service nova] Acquiring lock "refresh_cache-5cb09b60-70b6-4d04-850e-049612ec4a89" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 773.956875] env[61957]: DEBUG oslo_concurrency.lockutils [req-a528ea5f-fad7-4bbf-b8fb-0d88738863c4 req-63af1110-e24c-414f-bd9f-8fd5d48f9787 service nova] Acquired lock "refresh_cache-5cb09b60-70b6-4d04-850e-049612ec4a89" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 773.956992] env[61957]: DEBUG nova.network.neutron [req-a528ea5f-fad7-4bbf-b8fb-0d88738863c4 req-63af1110-e24c-414f-bd9f-8fd5d48f9787 service nova] [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] Refreshing network info cache for port ad48de80-9219-4f81-8c65-da8fb5df2729 {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 773.989547] env[61957]: ERROR nova.compute.manager [None req-d6c9937c-2526-4768-bade-4ca2ced8b569 tempest-ServersTestFqdnHostnames-1308211839 tempest-ServersTestFqdnHostnames-1308211839-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port ad48de80-9219-4f81-8c65-da8fb5df2729, please check neutron logs for more information. [ 773.989547] env[61957]: ERROR nova.compute.manager Traceback (most recent call last): [ 773.989547] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 773.989547] env[61957]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 773.989547] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 773.989547] env[61957]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 773.989547] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 773.989547] env[61957]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 773.989547] env[61957]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 773.989547] env[61957]: ERROR nova.compute.manager self.force_reraise() [ 773.989547] env[61957]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 773.989547] env[61957]: ERROR nova.compute.manager raise self.value [ 773.989547] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 773.989547] env[61957]: ERROR nova.compute.manager updated_port = self._update_port( [ 773.989547] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 773.989547] env[61957]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 773.990087] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 773.990087] env[61957]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 773.990087] env[61957]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port ad48de80-9219-4f81-8c65-da8fb5df2729, please check neutron logs for more information. [ 773.990087] env[61957]: ERROR nova.compute.manager [ 773.990087] env[61957]: Traceback (most recent call last): [ 773.990087] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 773.990087] env[61957]: listener.cb(fileno) [ 773.990087] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 773.990087] env[61957]: result = function(*args, **kwargs) [ 773.990087] env[61957]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 773.990087] env[61957]: return func(*args, **kwargs) [ 773.990087] env[61957]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 773.990087] env[61957]: raise e [ 773.990087] env[61957]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 773.990087] env[61957]: nwinfo = self.network_api.allocate_for_instance( [ 773.990087] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 773.990087] env[61957]: created_port_ids = self._update_ports_for_instance( [ 773.990087] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 773.990087] env[61957]: with excutils.save_and_reraise_exception(): [ 773.990087] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 773.990087] env[61957]: self.force_reraise() [ 773.990087] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 773.990087] env[61957]: raise self.value [ 773.990087] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 773.990087] env[61957]: updated_port = self._update_port( [ 773.990087] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 773.990087] env[61957]: _ensure_no_port_binding_failure(port) [ 773.990087] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 773.990087] env[61957]: raise exception.PortBindingFailed(port_id=port['id']) [ 773.990965] env[61957]: nova.exception.PortBindingFailed: Binding failed for port ad48de80-9219-4f81-8c65-da8fb5df2729, please check neutron logs for more information. [ 773.990965] env[61957]: Removing descriptor: 16 [ 774.025092] env[61957]: DEBUG nova.scheduler.client.report [None req-bb8cb37b-74c5-4607-9ba8-e5615134a1da tempest-InstanceActionsTestJSON-565478614 tempest-InstanceActionsTestJSON-565478614-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 774.168335] env[61957]: DEBUG nova.compute.manager [None req-d6c9937c-2526-4768-bade-4ca2ced8b569 tempest-ServersTestFqdnHostnames-1308211839 tempest-ServersTestFqdnHostnames-1308211839-project-member] [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] Start spawning the instance on the hypervisor. {{(pid=61957) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 774.193522] env[61957]: DEBUG nova.virt.hardware [None req-d6c9937c-2526-4768-bade-4ca2ced8b569 tempest-ServersTestFqdnHostnames-1308211839 tempest-ServersTestFqdnHostnames-1308211839-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T17:22:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T17:21:50Z,direct_url=,disk_format='vmdk',id=11c76a2c-f705-470a-ba9d-4657858bab38,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='81d3096d0f094373913fc3dc8f5c3c6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T17:21:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 774.193853] env[61957]: DEBUG nova.virt.hardware [None req-d6c9937c-2526-4768-bade-4ca2ced8b569 tempest-ServersTestFqdnHostnames-1308211839 tempest-ServersTestFqdnHostnames-1308211839-project-member] Flavor limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 774.194185] env[61957]: DEBUG nova.virt.hardware [None req-d6c9937c-2526-4768-bade-4ca2ced8b569 tempest-ServersTestFqdnHostnames-1308211839 tempest-ServersTestFqdnHostnames-1308211839-project-member] Image limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 774.194346] env[61957]: DEBUG nova.virt.hardware [None req-d6c9937c-2526-4768-bade-4ca2ced8b569 tempest-ServersTestFqdnHostnames-1308211839 tempest-ServersTestFqdnHostnames-1308211839-project-member] Flavor pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 774.194514] env[61957]: DEBUG nova.virt.hardware [None req-d6c9937c-2526-4768-bade-4ca2ced8b569 tempest-ServersTestFqdnHostnames-1308211839 tempest-ServersTestFqdnHostnames-1308211839-project-member] Image pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 774.194686] env[61957]: DEBUG nova.virt.hardware [None req-d6c9937c-2526-4768-bade-4ca2ced8b569 tempest-ServersTestFqdnHostnames-1308211839 tempest-ServersTestFqdnHostnames-1308211839-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 774.194978] env[61957]: DEBUG nova.virt.hardware [None req-d6c9937c-2526-4768-bade-4ca2ced8b569 tempest-ServersTestFqdnHostnames-1308211839 tempest-ServersTestFqdnHostnames-1308211839-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 774.195213] env[61957]: DEBUG nova.virt.hardware [None req-d6c9937c-2526-4768-bade-4ca2ced8b569 tempest-ServersTestFqdnHostnames-1308211839 tempest-ServersTestFqdnHostnames-1308211839-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 774.195393] env[61957]: DEBUG nova.virt.hardware [None req-d6c9937c-2526-4768-bade-4ca2ced8b569 tempest-ServersTestFqdnHostnames-1308211839 tempest-ServersTestFqdnHostnames-1308211839-project-member] Got 1 possible topologies {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 774.195556] env[61957]: DEBUG nova.virt.hardware [None req-d6c9937c-2526-4768-bade-4ca2ced8b569 tempest-ServersTestFqdnHostnames-1308211839 tempest-ServersTestFqdnHostnames-1308211839-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 774.195727] env[61957]: DEBUG nova.virt.hardware [None req-d6c9937c-2526-4768-bade-4ca2ced8b569 tempest-ServersTestFqdnHostnames-1308211839 tempest-ServersTestFqdnHostnames-1308211839-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 774.196596] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54050b99-de92-4d57-b0cf-4aca4e1080ba {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.209025] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b95810e-3db9-4de4-a3ec-96f3de703a3c {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.232044] env[61957]: ERROR nova.compute.manager [None req-d6c9937c-2526-4768-bade-4ca2ced8b569 tempest-ServersTestFqdnHostnames-1308211839 tempest-ServersTestFqdnHostnames-1308211839-project-member] [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port ad48de80-9219-4f81-8c65-da8fb5df2729, please check neutron logs for more information. [ 774.232044] env[61957]: ERROR nova.compute.manager [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] Traceback (most recent call last): [ 774.232044] env[61957]: ERROR nova.compute.manager [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 774.232044] env[61957]: ERROR nova.compute.manager [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] yield resources [ 774.232044] env[61957]: ERROR nova.compute.manager [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 774.232044] env[61957]: ERROR nova.compute.manager [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] self.driver.spawn(context, instance, image_meta, [ 774.232044] env[61957]: ERROR nova.compute.manager [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 774.232044] env[61957]: ERROR nova.compute.manager [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] self._vmops.spawn(context, instance, image_meta, injected_files, [ 774.232044] env[61957]: ERROR nova.compute.manager [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 774.232044] env[61957]: ERROR nova.compute.manager [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] vm_ref = self.build_virtual_machine(instance, [ 774.232044] env[61957]: ERROR nova.compute.manager [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 774.232566] env[61957]: ERROR nova.compute.manager [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] vif_infos = vmwarevif.get_vif_info(self._session, [ 774.232566] env[61957]: ERROR nova.compute.manager [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 774.232566] env[61957]: ERROR nova.compute.manager [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] for vif in network_info: [ 774.232566] env[61957]: ERROR nova.compute.manager [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 774.232566] env[61957]: ERROR nova.compute.manager [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] return self._sync_wrapper(fn, *args, **kwargs) [ 774.232566] env[61957]: ERROR nova.compute.manager [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 774.232566] env[61957]: ERROR nova.compute.manager [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] self.wait() [ 774.232566] env[61957]: ERROR nova.compute.manager [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 774.232566] env[61957]: ERROR nova.compute.manager [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] self[:] = self._gt.wait() [ 774.232566] env[61957]: ERROR nova.compute.manager [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 774.232566] env[61957]: ERROR nova.compute.manager [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] return self._exit_event.wait() [ 774.232566] env[61957]: ERROR nova.compute.manager [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 774.232566] env[61957]: ERROR nova.compute.manager [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] current.throw(*self._exc) [ 774.233036] env[61957]: ERROR nova.compute.manager [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 774.233036] env[61957]: ERROR nova.compute.manager [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] result = function(*args, **kwargs) [ 774.233036] env[61957]: ERROR nova.compute.manager [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 774.233036] env[61957]: ERROR nova.compute.manager [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] return func(*args, **kwargs) [ 774.233036] env[61957]: ERROR nova.compute.manager [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 774.233036] env[61957]: ERROR nova.compute.manager [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] raise e [ 774.233036] env[61957]: ERROR nova.compute.manager [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 774.233036] env[61957]: ERROR nova.compute.manager [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] nwinfo = self.network_api.allocate_for_instance( [ 774.233036] env[61957]: ERROR nova.compute.manager [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 774.233036] env[61957]: ERROR nova.compute.manager [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] created_port_ids = self._update_ports_for_instance( [ 774.233036] env[61957]: ERROR nova.compute.manager [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 774.233036] env[61957]: ERROR nova.compute.manager [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] with excutils.save_and_reraise_exception(): [ 774.233036] env[61957]: ERROR nova.compute.manager [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 774.233550] env[61957]: ERROR nova.compute.manager [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] self.force_reraise() [ 774.233550] env[61957]: ERROR nova.compute.manager [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 774.233550] env[61957]: ERROR nova.compute.manager [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] raise self.value [ 774.233550] env[61957]: ERROR nova.compute.manager [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 774.233550] env[61957]: ERROR nova.compute.manager [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] updated_port = self._update_port( [ 774.233550] env[61957]: ERROR nova.compute.manager [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 774.233550] env[61957]: ERROR nova.compute.manager [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] _ensure_no_port_binding_failure(port) [ 774.233550] env[61957]: ERROR nova.compute.manager [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 774.233550] env[61957]: ERROR nova.compute.manager [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] raise exception.PortBindingFailed(port_id=port['id']) [ 774.233550] env[61957]: ERROR nova.compute.manager [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] nova.exception.PortBindingFailed: Binding failed for port ad48de80-9219-4f81-8c65-da8fb5df2729, please check neutron logs for more information. [ 774.233550] env[61957]: ERROR nova.compute.manager [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] [ 774.236220] env[61957]: INFO nova.compute.manager [None req-d6c9937c-2526-4768-bade-4ca2ced8b569 tempest-ServersTestFqdnHostnames-1308211839 tempest-ServersTestFqdnHostnames-1308211839-project-member] [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] Terminating instance [ 774.238462] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d6c9937c-2526-4768-bade-4ca2ced8b569 tempest-ServersTestFqdnHostnames-1308211839 tempest-ServersTestFqdnHostnames-1308211839-project-member] Acquiring lock "refresh_cache-5cb09b60-70b6-4d04-850e-049612ec4a89" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 774.377428] env[61957]: DEBUG oslo_concurrency.lockutils [None req-da68b092-4602-4484-b1b0-adcac4aae9f8 tempest-InstanceActionsV221TestJSON-1768084888 tempest-InstanceActionsV221TestJSON-1768084888-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 774.474994] env[61957]: DEBUG nova.network.neutron [req-a528ea5f-fad7-4bbf-b8fb-0d88738863c4 req-63af1110-e24c-414f-bd9f-8fd5d48f9787 service nova] [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 774.528809] env[61957]: DEBUG oslo_concurrency.lockutils [None req-bb8cb37b-74c5-4607-9ba8-e5615134a1da tempest-InstanceActionsTestJSON-565478614 tempest-InstanceActionsTestJSON-565478614-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.389s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 774.529167] env[61957]: DEBUG nova.compute.manager [None req-bb8cb37b-74c5-4607-9ba8-e5615134a1da tempest-InstanceActionsTestJSON-565478614 tempest-InstanceActionsTestJSON-565478614-project-member] [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] Start building networks asynchronously for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 774.531691] env[61957]: DEBUG oslo_concurrency.lockutils [None req-7860e034-eff8-489c-addf-33cf2776246c tempest-ServerActionsTestOtherA-1610268163 tempest-ServerActionsTestOtherA-1610268163-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.960s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 774.533131] env[61957]: INFO nova.compute.claims [None req-7860e034-eff8-489c-addf-33cf2776246c tempest-ServerActionsTestOtherA-1610268163 tempest-ServerActionsTestOtherA-1610268163-project-member] [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 774.564642] env[61957]: DEBUG nova.network.neutron [req-a528ea5f-fad7-4bbf-b8fb-0d88738863c4 req-63af1110-e24c-414f-bd9f-8fd5d48f9787 service nova] [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 775.037523] env[61957]: DEBUG nova.compute.utils [None req-bb8cb37b-74c5-4607-9ba8-e5615134a1da tempest-InstanceActionsTestJSON-565478614 tempest-InstanceActionsTestJSON-565478614-project-member] Using /dev/sd instead of None {{(pid=61957) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 775.038877] env[61957]: DEBUG nova.compute.manager [None req-bb8cb37b-74c5-4607-9ba8-e5615134a1da tempest-InstanceActionsTestJSON-565478614 tempest-InstanceActionsTestJSON-565478614-project-member] [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] Allocating IP information in the background. {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 775.039116] env[61957]: DEBUG nova.network.neutron [None req-bb8cb37b-74c5-4607-9ba8-e5615134a1da tempest-InstanceActionsTestJSON-565478614 tempest-InstanceActionsTestJSON-565478614-project-member] [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] allocate_for_instance() {{(pid=61957) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 775.069867] env[61957]: DEBUG oslo_concurrency.lockutils [req-a528ea5f-fad7-4bbf-b8fb-0d88738863c4 req-63af1110-e24c-414f-bd9f-8fd5d48f9787 service nova] Releasing lock "refresh_cache-5cb09b60-70b6-4d04-850e-049612ec4a89" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 775.070312] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d6c9937c-2526-4768-bade-4ca2ced8b569 tempest-ServersTestFqdnHostnames-1308211839 tempest-ServersTestFqdnHostnames-1308211839-project-member] Acquired lock "refresh_cache-5cb09b60-70b6-4d04-850e-049612ec4a89" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 775.070499] env[61957]: DEBUG nova.network.neutron [None req-d6c9937c-2526-4768-bade-4ca2ced8b569 tempest-ServersTestFqdnHostnames-1308211839 tempest-ServersTestFqdnHostnames-1308211839-project-member] [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 775.089928] env[61957]: DEBUG nova.policy [None req-bb8cb37b-74c5-4607-9ba8-e5615134a1da tempest-InstanceActionsTestJSON-565478614 tempest-InstanceActionsTestJSON-565478614-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '56a9dc738be04bda8ba7c8efa95ca34b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '48e08880dc8f473384a8a9f00809708d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61957) authorize /opt/stack/nova/nova/policy.py:203}} [ 775.373470] env[61957]: DEBUG nova.network.neutron [None req-bb8cb37b-74c5-4607-9ba8-e5615134a1da tempest-InstanceActionsTestJSON-565478614 tempest-InstanceActionsTestJSON-565478614-project-member] [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] Successfully created port: 5ec2282a-722a-45ab-9124-906367749fe7 {{(pid=61957) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 775.544512] env[61957]: DEBUG nova.compute.manager [None req-bb8cb37b-74c5-4607-9ba8-e5615134a1da tempest-InstanceActionsTestJSON-565478614 tempest-InstanceActionsTestJSON-565478614-project-member] [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] Start building block device mappings for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 775.597539] env[61957]: DEBUG nova.network.neutron [None req-d6c9937c-2526-4768-bade-4ca2ced8b569 tempest-ServersTestFqdnHostnames-1308211839 tempest-ServersTestFqdnHostnames-1308211839-project-member] [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 775.717175] env[61957]: DEBUG nova.network.neutron [None req-d6c9937c-2526-4768-bade-4ca2ced8b569 tempest-ServersTestFqdnHostnames-1308211839 tempest-ServersTestFqdnHostnames-1308211839-project-member] [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 775.918818] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b0513d5-2c57-420d-8fb2-ba3bb1b12e34 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.926852] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9d22545-45a5-4dc7-82b2-b3f2149ea3c4 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.958491] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27ac2224-9b42-4b75-8678-1c24c2fa96b2 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.967234] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3296bcd-ca3a-429e-8a96-af6ebf4a2284 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.983056] env[61957]: DEBUG nova.compute.provider_tree [None req-7860e034-eff8-489c-addf-33cf2776246c tempest-ServerActionsTestOtherA-1610268163 tempest-ServerActionsTestOtherA-1610268163-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 775.987421] env[61957]: DEBUG nova.compute.manager [req-c6d741a5-687c-45ef-8a98-efec9d9c7355 req-c5566d48-1d5b-4cfd-be35-a35cf249263e service nova] [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] Received event network-vif-deleted-ad48de80-9219-4f81-8c65-da8fb5df2729 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 776.220511] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d6c9937c-2526-4768-bade-4ca2ced8b569 tempest-ServersTestFqdnHostnames-1308211839 tempest-ServersTestFqdnHostnames-1308211839-project-member] Releasing lock "refresh_cache-5cb09b60-70b6-4d04-850e-049612ec4a89" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 776.221209] env[61957]: DEBUG nova.compute.manager [None req-d6c9937c-2526-4768-bade-4ca2ced8b569 tempest-ServersTestFqdnHostnames-1308211839 tempest-ServersTestFqdnHostnames-1308211839-project-member] [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] Start destroying the instance on the hypervisor. {{(pid=61957) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 776.221397] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-d6c9937c-2526-4768-bade-4ca2ced8b569 tempest-ServersTestFqdnHostnames-1308211839 tempest-ServersTestFqdnHostnames-1308211839-project-member] [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] Destroying instance {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 776.221768] env[61957]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-38f0f24b-9afd-4de8-9f45-23255e2e4851 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.232945] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39e795e7-4010-4ccf-80ab-a2350555f598 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.254999] env[61957]: WARNING nova.virt.vmwareapi.vmops [None req-d6c9937c-2526-4768-bade-4ca2ced8b569 tempest-ServersTestFqdnHostnames-1308211839 tempest-ServersTestFqdnHostnames-1308211839-project-member] [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 5cb09b60-70b6-4d04-850e-049612ec4a89 could not be found. [ 776.255467] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-d6c9937c-2526-4768-bade-4ca2ced8b569 tempest-ServersTestFqdnHostnames-1308211839 tempest-ServersTestFqdnHostnames-1308211839-project-member] [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] Instance destroyed {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 776.255467] env[61957]: INFO nova.compute.manager [None req-d6c9937c-2526-4768-bade-4ca2ced8b569 tempest-ServersTestFqdnHostnames-1308211839 tempest-ServersTestFqdnHostnames-1308211839-project-member] [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] Took 0.03 seconds to destroy the instance on the hypervisor. [ 776.255655] env[61957]: DEBUG oslo.service.loopingcall [None req-d6c9937c-2526-4768-bade-4ca2ced8b569 tempest-ServersTestFqdnHostnames-1308211839 tempest-ServersTestFqdnHostnames-1308211839-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 776.255871] env[61957]: DEBUG nova.compute.manager [-] [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 776.255967] env[61957]: DEBUG nova.network.neutron [-] [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 776.271674] env[61957]: DEBUG nova.network.neutron [-] [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 776.345851] env[61957]: ERROR nova.compute.manager [None req-bb8cb37b-74c5-4607-9ba8-e5615134a1da tempest-InstanceActionsTestJSON-565478614 tempest-InstanceActionsTestJSON-565478614-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 5ec2282a-722a-45ab-9124-906367749fe7, please check neutron logs for more information. [ 776.345851] env[61957]: ERROR nova.compute.manager Traceback (most recent call last): [ 776.345851] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 776.345851] env[61957]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 776.345851] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 776.345851] env[61957]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 776.345851] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 776.345851] env[61957]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 776.345851] env[61957]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 776.345851] env[61957]: ERROR nova.compute.manager self.force_reraise() [ 776.345851] env[61957]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 776.345851] env[61957]: ERROR nova.compute.manager raise self.value [ 776.345851] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 776.345851] env[61957]: ERROR nova.compute.manager updated_port = self._update_port( [ 776.345851] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 776.345851] env[61957]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 776.346431] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 776.346431] env[61957]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 776.346431] env[61957]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 5ec2282a-722a-45ab-9124-906367749fe7, please check neutron logs for more information. [ 776.346431] env[61957]: ERROR nova.compute.manager [ 776.346431] env[61957]: Traceback (most recent call last): [ 776.346431] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 776.346431] env[61957]: listener.cb(fileno) [ 776.346431] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 776.346431] env[61957]: result = function(*args, **kwargs) [ 776.346431] env[61957]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 776.346431] env[61957]: return func(*args, **kwargs) [ 776.346431] env[61957]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 776.346431] env[61957]: raise e [ 776.346431] env[61957]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 776.346431] env[61957]: nwinfo = self.network_api.allocate_for_instance( [ 776.346431] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 776.346431] env[61957]: created_port_ids = self._update_ports_for_instance( [ 776.346431] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 776.346431] env[61957]: with excutils.save_and_reraise_exception(): [ 776.346431] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 776.346431] env[61957]: self.force_reraise() [ 776.346431] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 776.346431] env[61957]: raise self.value [ 776.346431] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 776.346431] env[61957]: updated_port = self._update_port( [ 776.346431] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 776.346431] env[61957]: _ensure_no_port_binding_failure(port) [ 776.346431] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 776.346431] env[61957]: raise exception.PortBindingFailed(port_id=port['id']) [ 776.347436] env[61957]: nova.exception.PortBindingFailed: Binding failed for port 5ec2282a-722a-45ab-9124-906367749fe7, please check neutron logs for more information. [ 776.347436] env[61957]: Removing descriptor: 16 [ 776.489491] env[61957]: DEBUG nova.scheduler.client.report [None req-7860e034-eff8-489c-addf-33cf2776246c tempest-ServerActionsTestOtherA-1610268163 tempest-ServerActionsTestOtherA-1610268163-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 776.556814] env[61957]: DEBUG nova.compute.manager [None req-bb8cb37b-74c5-4607-9ba8-e5615134a1da tempest-InstanceActionsTestJSON-565478614 tempest-InstanceActionsTestJSON-565478614-project-member] [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] Start spawning the instance on the hypervisor. {{(pid=61957) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 776.584051] env[61957]: DEBUG nova.virt.hardware [None req-bb8cb37b-74c5-4607-9ba8-e5615134a1da tempest-InstanceActionsTestJSON-565478614 tempest-InstanceActionsTestJSON-565478614-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T17:22:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T17:21:50Z,direct_url=,disk_format='vmdk',id=11c76a2c-f705-470a-ba9d-4657858bab38,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='81d3096d0f094373913fc3dc8f5c3c6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T17:21:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 776.584364] env[61957]: DEBUG nova.virt.hardware [None req-bb8cb37b-74c5-4607-9ba8-e5615134a1da tempest-InstanceActionsTestJSON-565478614 tempest-InstanceActionsTestJSON-565478614-project-member] Flavor limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 776.584457] env[61957]: DEBUG nova.virt.hardware [None req-bb8cb37b-74c5-4607-9ba8-e5615134a1da tempest-InstanceActionsTestJSON-565478614 tempest-InstanceActionsTestJSON-565478614-project-member] Image limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 776.584627] env[61957]: DEBUG nova.virt.hardware [None req-bb8cb37b-74c5-4607-9ba8-e5615134a1da tempest-InstanceActionsTestJSON-565478614 tempest-InstanceActionsTestJSON-565478614-project-member] Flavor pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 776.584768] env[61957]: DEBUG nova.virt.hardware [None req-bb8cb37b-74c5-4607-9ba8-e5615134a1da tempest-InstanceActionsTestJSON-565478614 tempest-InstanceActionsTestJSON-565478614-project-member] Image pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 776.584913] env[61957]: DEBUG nova.virt.hardware [None req-bb8cb37b-74c5-4607-9ba8-e5615134a1da tempest-InstanceActionsTestJSON-565478614 tempest-InstanceActionsTestJSON-565478614-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 776.585205] env[61957]: DEBUG nova.virt.hardware [None req-bb8cb37b-74c5-4607-9ba8-e5615134a1da tempest-InstanceActionsTestJSON-565478614 tempest-InstanceActionsTestJSON-565478614-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 776.585393] env[61957]: DEBUG nova.virt.hardware [None req-bb8cb37b-74c5-4607-9ba8-e5615134a1da tempest-InstanceActionsTestJSON-565478614 tempest-InstanceActionsTestJSON-565478614-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 776.585450] env[61957]: DEBUG nova.virt.hardware [None req-bb8cb37b-74c5-4607-9ba8-e5615134a1da tempest-InstanceActionsTestJSON-565478614 tempest-InstanceActionsTestJSON-565478614-project-member] Got 1 possible topologies {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 776.585613] env[61957]: DEBUG nova.virt.hardware [None req-bb8cb37b-74c5-4607-9ba8-e5615134a1da tempest-InstanceActionsTestJSON-565478614 tempest-InstanceActionsTestJSON-565478614-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 776.585781] env[61957]: DEBUG nova.virt.hardware [None req-bb8cb37b-74c5-4607-9ba8-e5615134a1da tempest-InstanceActionsTestJSON-565478614 tempest-InstanceActionsTestJSON-565478614-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 776.586629] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-874693cd-6fb0-45ed-8b88-afbdf845efff {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.595052] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-515931f9-9496-4882-9c82-6bc4a3167773 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.609408] env[61957]: ERROR nova.compute.manager [None req-bb8cb37b-74c5-4607-9ba8-e5615134a1da tempest-InstanceActionsTestJSON-565478614 tempest-InstanceActionsTestJSON-565478614-project-member] [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 5ec2282a-722a-45ab-9124-906367749fe7, please check neutron logs for more information. [ 776.609408] env[61957]: ERROR nova.compute.manager [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] Traceback (most recent call last): [ 776.609408] env[61957]: ERROR nova.compute.manager [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 776.609408] env[61957]: ERROR nova.compute.manager [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] yield resources [ 776.609408] env[61957]: ERROR nova.compute.manager [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 776.609408] env[61957]: ERROR nova.compute.manager [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] self.driver.spawn(context, instance, image_meta, [ 776.609408] env[61957]: ERROR nova.compute.manager [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 776.609408] env[61957]: ERROR nova.compute.manager [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 776.609408] env[61957]: ERROR nova.compute.manager [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 776.609408] env[61957]: ERROR nova.compute.manager [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] vm_ref = self.build_virtual_machine(instance, [ 776.609408] env[61957]: ERROR nova.compute.manager [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 776.609859] env[61957]: ERROR nova.compute.manager [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] vif_infos = vmwarevif.get_vif_info(self._session, [ 776.609859] env[61957]: ERROR nova.compute.manager [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 776.609859] env[61957]: ERROR nova.compute.manager [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] for vif in network_info: [ 776.609859] env[61957]: ERROR nova.compute.manager [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 776.609859] env[61957]: ERROR nova.compute.manager [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] return self._sync_wrapper(fn, *args, **kwargs) [ 776.609859] env[61957]: ERROR nova.compute.manager [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 776.609859] env[61957]: ERROR nova.compute.manager [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] self.wait() [ 776.609859] env[61957]: ERROR nova.compute.manager [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 776.609859] env[61957]: ERROR nova.compute.manager [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] self[:] = self._gt.wait() [ 776.609859] env[61957]: ERROR nova.compute.manager [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 776.609859] env[61957]: ERROR nova.compute.manager [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] return self._exit_event.wait() [ 776.609859] env[61957]: ERROR nova.compute.manager [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 776.609859] env[61957]: ERROR nova.compute.manager [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] current.throw(*self._exc) [ 776.610334] env[61957]: ERROR nova.compute.manager [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 776.610334] env[61957]: ERROR nova.compute.manager [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] result = function(*args, **kwargs) [ 776.610334] env[61957]: ERROR nova.compute.manager [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 776.610334] env[61957]: ERROR nova.compute.manager [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] return func(*args, **kwargs) [ 776.610334] env[61957]: ERROR nova.compute.manager [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 776.610334] env[61957]: ERROR nova.compute.manager [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] raise e [ 776.610334] env[61957]: ERROR nova.compute.manager [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 776.610334] env[61957]: ERROR nova.compute.manager [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] nwinfo = self.network_api.allocate_for_instance( [ 776.610334] env[61957]: ERROR nova.compute.manager [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 776.610334] env[61957]: ERROR nova.compute.manager [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] created_port_ids = self._update_ports_for_instance( [ 776.610334] env[61957]: ERROR nova.compute.manager [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 776.610334] env[61957]: ERROR nova.compute.manager [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] with excutils.save_and_reraise_exception(): [ 776.610334] env[61957]: ERROR nova.compute.manager [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 776.610783] env[61957]: ERROR nova.compute.manager [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] self.force_reraise() [ 776.610783] env[61957]: ERROR nova.compute.manager [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 776.610783] env[61957]: ERROR nova.compute.manager [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] raise self.value [ 776.610783] env[61957]: ERROR nova.compute.manager [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 776.610783] env[61957]: ERROR nova.compute.manager [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] updated_port = self._update_port( [ 776.610783] env[61957]: ERROR nova.compute.manager [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 776.610783] env[61957]: ERROR nova.compute.manager [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] _ensure_no_port_binding_failure(port) [ 776.610783] env[61957]: ERROR nova.compute.manager [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 776.610783] env[61957]: ERROR nova.compute.manager [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] raise exception.PortBindingFailed(port_id=port['id']) [ 776.610783] env[61957]: ERROR nova.compute.manager [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] nova.exception.PortBindingFailed: Binding failed for port 5ec2282a-722a-45ab-9124-906367749fe7, please check neutron logs for more information. [ 776.610783] env[61957]: ERROR nova.compute.manager [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] [ 776.610783] env[61957]: INFO nova.compute.manager [None req-bb8cb37b-74c5-4607-9ba8-e5615134a1da tempest-InstanceActionsTestJSON-565478614 tempest-InstanceActionsTestJSON-565478614-project-member] [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] Terminating instance [ 776.611870] env[61957]: DEBUG oslo_concurrency.lockutils [None req-bb8cb37b-74c5-4607-9ba8-e5615134a1da tempest-InstanceActionsTestJSON-565478614 tempest-InstanceActionsTestJSON-565478614-project-member] Acquiring lock "refresh_cache-2f84d2df-00f0-4860-acb4-4e10ecf8ce4e" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 776.612041] env[61957]: DEBUG oslo_concurrency.lockutils [None req-bb8cb37b-74c5-4607-9ba8-e5615134a1da tempest-InstanceActionsTestJSON-565478614 tempest-InstanceActionsTestJSON-565478614-project-member] Acquired lock "refresh_cache-2f84d2df-00f0-4860-acb4-4e10ecf8ce4e" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 776.612210] env[61957]: DEBUG nova.network.neutron [None req-bb8cb37b-74c5-4607-9ba8-e5615134a1da tempest-InstanceActionsTestJSON-565478614 tempest-InstanceActionsTestJSON-565478614-project-member] [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 776.774009] env[61957]: DEBUG nova.network.neutron [-] [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 776.994726] env[61957]: DEBUG oslo_concurrency.lockutils [None req-7860e034-eff8-489c-addf-33cf2776246c tempest-ServerActionsTestOtherA-1610268163 tempest-ServerActionsTestOtherA-1610268163-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.463s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 776.995312] env[61957]: DEBUG nova.compute.manager [None req-7860e034-eff8-489c-addf-33cf2776246c tempest-ServerActionsTestOtherA-1610268163 tempest-ServerActionsTestOtherA-1610268163-project-member] [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] Start building networks asynchronously for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 776.998230] env[61957]: DEBUG oslo_concurrency.lockutils [None req-09f1eb26-3ef8-431d-92bc-9c62b712a196 tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.025s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 777.130022] env[61957]: DEBUG nova.network.neutron [None req-bb8cb37b-74c5-4607-9ba8-e5615134a1da tempest-InstanceActionsTestJSON-565478614 tempest-InstanceActionsTestJSON-565478614-project-member] [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 777.214116] env[61957]: DEBUG nova.network.neutron [None req-bb8cb37b-74c5-4607-9ba8-e5615134a1da tempest-InstanceActionsTestJSON-565478614 tempest-InstanceActionsTestJSON-565478614-project-member] [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 777.276289] env[61957]: INFO nova.compute.manager [-] [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] Took 1.02 seconds to deallocate network for instance. [ 777.280717] env[61957]: DEBUG nova.compute.claims [None req-d6c9937c-2526-4768-bade-4ca2ced8b569 tempest-ServersTestFqdnHostnames-1308211839 tempest-ServersTestFqdnHostnames-1308211839-project-member] [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] Aborting claim: {{(pid=61957) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 777.280933] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d6c9937c-2526-4768-bade-4ca2ced8b569 tempest-ServersTestFqdnHostnames-1308211839 tempest-ServersTestFqdnHostnames-1308211839-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 777.503574] env[61957]: DEBUG nova.compute.utils [None req-7860e034-eff8-489c-addf-33cf2776246c tempest-ServerActionsTestOtherA-1610268163 tempest-ServerActionsTestOtherA-1610268163-project-member] Using /dev/sd instead of None {{(pid=61957) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 777.508364] env[61957]: DEBUG nova.compute.manager [None req-7860e034-eff8-489c-addf-33cf2776246c tempest-ServerActionsTestOtherA-1610268163 tempest-ServerActionsTestOtherA-1610268163-project-member] [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] Allocating IP information in the background. {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 777.508364] env[61957]: DEBUG nova.network.neutron [None req-7860e034-eff8-489c-addf-33cf2776246c tempest-ServerActionsTestOtherA-1610268163 tempest-ServerActionsTestOtherA-1610268163-project-member] [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] allocate_for_instance() {{(pid=61957) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 777.571145] env[61957]: DEBUG nova.policy [None req-7860e034-eff8-489c-addf-33cf2776246c tempest-ServerActionsTestOtherA-1610268163 tempest-ServerActionsTestOtherA-1610268163-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '06afb43cdb314dcb8c606a9fb030e03a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3cd1fa82dbbd462f8db2d8471ece77ba', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61957) authorize /opt/stack/nova/nova/policy.py:203}} [ 777.716842] env[61957]: DEBUG oslo_concurrency.lockutils [None req-bb8cb37b-74c5-4607-9ba8-e5615134a1da tempest-InstanceActionsTestJSON-565478614 tempest-InstanceActionsTestJSON-565478614-project-member] Releasing lock "refresh_cache-2f84d2df-00f0-4860-acb4-4e10ecf8ce4e" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 777.717436] env[61957]: DEBUG nova.compute.manager [None req-bb8cb37b-74c5-4607-9ba8-e5615134a1da tempest-InstanceActionsTestJSON-565478614 tempest-InstanceActionsTestJSON-565478614-project-member] [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] Start destroying the instance on the hypervisor. {{(pid=61957) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 777.717847] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-bb8cb37b-74c5-4607-9ba8-e5615134a1da tempest-InstanceActionsTestJSON-565478614 tempest-InstanceActionsTestJSON-565478614-project-member] [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] Destroying instance {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 777.717847] env[61957]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-bb5bfbf0-81ab-463f-a755-090ac9cd4f66 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.726696] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e849f89e-9282-40c8-99ff-692050258e33 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.751205] env[61957]: WARNING nova.virt.vmwareapi.vmops [None req-bb8cb37b-74c5-4607-9ba8-e5615134a1da tempest-InstanceActionsTestJSON-565478614 tempest-InstanceActionsTestJSON-565478614-project-member] [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e could not be found. [ 777.751205] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-bb8cb37b-74c5-4607-9ba8-e5615134a1da tempest-InstanceActionsTestJSON-565478614 tempest-InstanceActionsTestJSON-565478614-project-member] [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] Instance destroyed {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 777.751205] env[61957]: INFO nova.compute.manager [None req-bb8cb37b-74c5-4607-9ba8-e5615134a1da tempest-InstanceActionsTestJSON-565478614 tempest-InstanceActionsTestJSON-565478614-project-member] [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] Took 0.03 seconds to destroy the instance on the hypervisor. [ 777.751205] env[61957]: DEBUG oslo.service.loopingcall [None req-bb8cb37b-74c5-4607-9ba8-e5615134a1da tempest-InstanceActionsTestJSON-565478614 tempest-InstanceActionsTestJSON-565478614-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 777.753326] env[61957]: DEBUG nova.compute.manager [-] [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 777.753433] env[61957]: DEBUG nova.network.neutron [-] [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 777.774614] env[61957]: DEBUG nova.network.neutron [-] [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 777.886444] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a52e18f-80ed-4aa4-961c-17146a4fe1be {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.893524] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77795a90-c5d1-4a3f-aaaa-4c2eebe6830f {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.926040] env[61957]: DEBUG nova.network.neutron [None req-7860e034-eff8-489c-addf-33cf2776246c tempest-ServerActionsTestOtherA-1610268163 tempest-ServerActionsTestOtherA-1610268163-project-member] [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] Successfully created port: 76ccb95a-3bfc-4be2-a607-679432fcf74d {{(pid=61957) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 777.927072] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba74846d-f2b8-42bf-a5b9-da5b970fe90f {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.934813] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-726cf136-ec2b-4853-bad2-0b2ae0e22393 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.950162] env[61957]: DEBUG nova.compute.provider_tree [None req-09f1eb26-3ef8-431d-92bc-9c62b712a196 tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 778.008670] env[61957]: DEBUG nova.compute.manager [None req-7860e034-eff8-489c-addf-33cf2776246c tempest-ServerActionsTestOtherA-1610268163 tempest-ServerActionsTestOtherA-1610268163-project-member] [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] Start building block device mappings for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 778.013098] env[61957]: DEBUG nova.compute.manager [req-3d556a74-167d-4258-990e-81b2bf33b271 req-4a84d6be-c486-448b-9aa3-0986bf966a5a service nova] [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] Received event network-changed-5ec2282a-722a-45ab-9124-906367749fe7 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 778.013287] env[61957]: DEBUG nova.compute.manager [req-3d556a74-167d-4258-990e-81b2bf33b271 req-4a84d6be-c486-448b-9aa3-0986bf966a5a service nova] [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] Refreshing instance network info cache due to event network-changed-5ec2282a-722a-45ab-9124-906367749fe7. {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 778.013518] env[61957]: DEBUG oslo_concurrency.lockutils [req-3d556a74-167d-4258-990e-81b2bf33b271 req-4a84d6be-c486-448b-9aa3-0986bf966a5a service nova] Acquiring lock "refresh_cache-2f84d2df-00f0-4860-acb4-4e10ecf8ce4e" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 778.013629] env[61957]: DEBUG oslo_concurrency.lockutils [req-3d556a74-167d-4258-990e-81b2bf33b271 req-4a84d6be-c486-448b-9aa3-0986bf966a5a service nova] Acquired lock "refresh_cache-2f84d2df-00f0-4860-acb4-4e10ecf8ce4e" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 778.013781] env[61957]: DEBUG nova.network.neutron [req-3d556a74-167d-4258-990e-81b2bf33b271 req-4a84d6be-c486-448b-9aa3-0986bf966a5a service nova] [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] Refreshing network info cache for port 5ec2282a-722a-45ab-9124-906367749fe7 {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 778.277099] env[61957]: DEBUG nova.network.neutron [-] [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 778.453635] env[61957]: DEBUG nova.scheduler.client.report [None req-09f1eb26-3ef8-431d-92bc-9c62b712a196 tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 778.541997] env[61957]: DEBUG nova.network.neutron [req-3d556a74-167d-4258-990e-81b2bf33b271 req-4a84d6be-c486-448b-9aa3-0986bf966a5a service nova] [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 778.676288] env[61957]: DEBUG nova.network.neutron [req-3d556a74-167d-4258-990e-81b2bf33b271 req-4a84d6be-c486-448b-9aa3-0986bf966a5a service nova] [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 778.779800] env[61957]: INFO nova.compute.manager [-] [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] Took 1.03 seconds to deallocate network for instance. [ 778.782356] env[61957]: DEBUG nova.compute.claims [None req-bb8cb37b-74c5-4607-9ba8-e5615134a1da tempest-InstanceActionsTestJSON-565478614 tempest-InstanceActionsTestJSON-565478614-project-member] [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] Aborting claim: {{(pid=61957) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 778.782543] env[61957]: DEBUG oslo_concurrency.lockutils [None req-bb8cb37b-74c5-4607-9ba8-e5615134a1da tempest-InstanceActionsTestJSON-565478614 tempest-InstanceActionsTestJSON-565478614-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 778.857109] env[61957]: ERROR nova.compute.manager [None req-7860e034-eff8-489c-addf-33cf2776246c tempest-ServerActionsTestOtherA-1610268163 tempest-ServerActionsTestOtherA-1610268163-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 76ccb95a-3bfc-4be2-a607-679432fcf74d, please check neutron logs for more information. [ 778.857109] env[61957]: ERROR nova.compute.manager Traceback (most recent call last): [ 778.857109] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 778.857109] env[61957]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 778.857109] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 778.857109] env[61957]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 778.857109] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 778.857109] env[61957]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 778.857109] env[61957]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 778.857109] env[61957]: ERROR nova.compute.manager self.force_reraise() [ 778.857109] env[61957]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 778.857109] env[61957]: ERROR nova.compute.manager raise self.value [ 778.857109] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 778.857109] env[61957]: ERROR nova.compute.manager updated_port = self._update_port( [ 778.857109] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 778.857109] env[61957]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 778.857625] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 778.857625] env[61957]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 778.857625] env[61957]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 76ccb95a-3bfc-4be2-a607-679432fcf74d, please check neutron logs for more information. [ 778.857625] env[61957]: ERROR nova.compute.manager [ 778.857625] env[61957]: Traceback (most recent call last): [ 778.857625] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 778.857625] env[61957]: listener.cb(fileno) [ 778.857625] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 778.857625] env[61957]: result = function(*args, **kwargs) [ 778.857625] env[61957]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 778.857625] env[61957]: return func(*args, **kwargs) [ 778.857625] env[61957]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 778.857625] env[61957]: raise e [ 778.857625] env[61957]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 778.857625] env[61957]: nwinfo = self.network_api.allocate_for_instance( [ 778.857625] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 778.857625] env[61957]: created_port_ids = self._update_ports_for_instance( [ 778.857625] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 778.857625] env[61957]: with excutils.save_and_reraise_exception(): [ 778.857625] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 778.857625] env[61957]: self.force_reraise() [ 778.857625] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 778.857625] env[61957]: raise self.value [ 778.857625] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 778.857625] env[61957]: updated_port = self._update_port( [ 778.857625] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 778.857625] env[61957]: _ensure_no_port_binding_failure(port) [ 778.857625] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 778.857625] env[61957]: raise exception.PortBindingFailed(port_id=port['id']) [ 778.858697] env[61957]: nova.exception.PortBindingFailed: Binding failed for port 76ccb95a-3bfc-4be2-a607-679432fcf74d, please check neutron logs for more information. [ 778.858697] env[61957]: Removing descriptor: 16 [ 778.960825] env[61957]: DEBUG oslo_concurrency.lockutils [None req-09f1eb26-3ef8-431d-92bc-9c62b712a196 tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.963s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 778.961476] env[61957]: ERROR nova.compute.manager [None req-09f1eb26-3ef8-431d-92bc-9c62b712a196 tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] [instance: 60ec4271-2614-470d-bf40-47b9955f544c] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port de4fd670-560d-4cad-ba00-671731d57410, please check neutron logs for more information. [ 778.961476] env[61957]: ERROR nova.compute.manager [instance: 60ec4271-2614-470d-bf40-47b9955f544c] Traceback (most recent call last): [ 778.961476] env[61957]: ERROR nova.compute.manager [instance: 60ec4271-2614-470d-bf40-47b9955f544c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 778.961476] env[61957]: ERROR nova.compute.manager [instance: 60ec4271-2614-470d-bf40-47b9955f544c] self.driver.spawn(context, instance, image_meta, [ 778.961476] env[61957]: ERROR nova.compute.manager [instance: 60ec4271-2614-470d-bf40-47b9955f544c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 778.961476] env[61957]: ERROR nova.compute.manager [instance: 60ec4271-2614-470d-bf40-47b9955f544c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 778.961476] env[61957]: ERROR nova.compute.manager [instance: 60ec4271-2614-470d-bf40-47b9955f544c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 778.961476] env[61957]: ERROR nova.compute.manager [instance: 60ec4271-2614-470d-bf40-47b9955f544c] vm_ref = self.build_virtual_machine(instance, [ 778.961476] env[61957]: ERROR nova.compute.manager [instance: 60ec4271-2614-470d-bf40-47b9955f544c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 778.961476] env[61957]: ERROR nova.compute.manager [instance: 60ec4271-2614-470d-bf40-47b9955f544c] vif_infos = vmwarevif.get_vif_info(self._session, [ 778.961476] env[61957]: ERROR nova.compute.manager [instance: 60ec4271-2614-470d-bf40-47b9955f544c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 778.961874] env[61957]: ERROR nova.compute.manager [instance: 60ec4271-2614-470d-bf40-47b9955f544c] for vif in network_info: [ 778.961874] env[61957]: ERROR nova.compute.manager [instance: 60ec4271-2614-470d-bf40-47b9955f544c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 778.961874] env[61957]: ERROR nova.compute.manager [instance: 60ec4271-2614-470d-bf40-47b9955f544c] return self._sync_wrapper(fn, *args, **kwargs) [ 778.961874] env[61957]: ERROR nova.compute.manager [instance: 60ec4271-2614-470d-bf40-47b9955f544c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 778.961874] env[61957]: ERROR nova.compute.manager [instance: 60ec4271-2614-470d-bf40-47b9955f544c] self.wait() [ 778.961874] env[61957]: ERROR nova.compute.manager [instance: 60ec4271-2614-470d-bf40-47b9955f544c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 778.961874] env[61957]: ERROR nova.compute.manager [instance: 60ec4271-2614-470d-bf40-47b9955f544c] self[:] = self._gt.wait() [ 778.961874] env[61957]: ERROR nova.compute.manager [instance: 60ec4271-2614-470d-bf40-47b9955f544c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 778.961874] env[61957]: ERROR nova.compute.manager [instance: 60ec4271-2614-470d-bf40-47b9955f544c] return self._exit_event.wait() [ 778.961874] env[61957]: ERROR nova.compute.manager [instance: 60ec4271-2614-470d-bf40-47b9955f544c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 778.961874] env[61957]: ERROR nova.compute.manager [instance: 60ec4271-2614-470d-bf40-47b9955f544c] current.throw(*self._exc) [ 778.961874] env[61957]: ERROR nova.compute.manager [instance: 60ec4271-2614-470d-bf40-47b9955f544c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 778.961874] env[61957]: ERROR nova.compute.manager [instance: 60ec4271-2614-470d-bf40-47b9955f544c] result = function(*args, **kwargs) [ 778.962322] env[61957]: ERROR nova.compute.manager [instance: 60ec4271-2614-470d-bf40-47b9955f544c] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 778.962322] env[61957]: ERROR nova.compute.manager [instance: 60ec4271-2614-470d-bf40-47b9955f544c] return func(*args, **kwargs) [ 778.962322] env[61957]: ERROR nova.compute.manager [instance: 60ec4271-2614-470d-bf40-47b9955f544c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 778.962322] env[61957]: ERROR nova.compute.manager [instance: 60ec4271-2614-470d-bf40-47b9955f544c] raise e [ 778.962322] env[61957]: ERROR nova.compute.manager [instance: 60ec4271-2614-470d-bf40-47b9955f544c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 778.962322] env[61957]: ERROR nova.compute.manager [instance: 60ec4271-2614-470d-bf40-47b9955f544c] nwinfo = self.network_api.allocate_for_instance( [ 778.962322] env[61957]: ERROR nova.compute.manager [instance: 60ec4271-2614-470d-bf40-47b9955f544c] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 778.962322] env[61957]: ERROR nova.compute.manager [instance: 60ec4271-2614-470d-bf40-47b9955f544c] created_port_ids = self._update_ports_for_instance( [ 778.962322] env[61957]: ERROR nova.compute.manager [instance: 60ec4271-2614-470d-bf40-47b9955f544c] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 778.962322] env[61957]: ERROR nova.compute.manager [instance: 60ec4271-2614-470d-bf40-47b9955f544c] with excutils.save_and_reraise_exception(): [ 778.962322] env[61957]: ERROR nova.compute.manager [instance: 60ec4271-2614-470d-bf40-47b9955f544c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 778.962322] env[61957]: ERROR nova.compute.manager [instance: 60ec4271-2614-470d-bf40-47b9955f544c] self.force_reraise() [ 778.962322] env[61957]: ERROR nova.compute.manager [instance: 60ec4271-2614-470d-bf40-47b9955f544c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 778.962756] env[61957]: ERROR nova.compute.manager [instance: 60ec4271-2614-470d-bf40-47b9955f544c] raise self.value [ 778.962756] env[61957]: ERROR nova.compute.manager [instance: 60ec4271-2614-470d-bf40-47b9955f544c] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 778.962756] env[61957]: ERROR nova.compute.manager [instance: 60ec4271-2614-470d-bf40-47b9955f544c] updated_port = self._update_port( [ 778.962756] env[61957]: ERROR nova.compute.manager [instance: 60ec4271-2614-470d-bf40-47b9955f544c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 778.962756] env[61957]: ERROR nova.compute.manager [instance: 60ec4271-2614-470d-bf40-47b9955f544c] _ensure_no_port_binding_failure(port) [ 778.962756] env[61957]: ERROR nova.compute.manager [instance: 60ec4271-2614-470d-bf40-47b9955f544c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 778.962756] env[61957]: ERROR nova.compute.manager [instance: 60ec4271-2614-470d-bf40-47b9955f544c] raise exception.PortBindingFailed(port_id=port['id']) [ 778.962756] env[61957]: ERROR nova.compute.manager [instance: 60ec4271-2614-470d-bf40-47b9955f544c] nova.exception.PortBindingFailed: Binding failed for port de4fd670-560d-4cad-ba00-671731d57410, please check neutron logs for more information. [ 778.962756] env[61957]: ERROR nova.compute.manager [instance: 60ec4271-2614-470d-bf40-47b9955f544c] [ 778.962756] env[61957]: DEBUG nova.compute.utils [None req-09f1eb26-3ef8-431d-92bc-9c62b712a196 tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] [instance: 60ec4271-2614-470d-bf40-47b9955f544c] Binding failed for port de4fd670-560d-4cad-ba00-671731d57410, please check neutron logs for more information. {{(pid=61957) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 778.963396] env[61957]: DEBUG oslo_concurrency.lockutils [None req-6b91bce8-d0f8-4c40-855e-ed34f3eb7e64 tempest-ServerActionsTestJSON-1567241756 tempest-ServerActionsTestJSON-1567241756-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.264s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 778.966916] env[61957]: DEBUG nova.compute.manager [None req-09f1eb26-3ef8-431d-92bc-9c62b712a196 tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] [instance: 60ec4271-2614-470d-bf40-47b9955f544c] Build of instance 60ec4271-2614-470d-bf40-47b9955f544c was re-scheduled: Binding failed for port de4fd670-560d-4cad-ba00-671731d57410, please check neutron logs for more information. {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 778.967337] env[61957]: DEBUG nova.compute.manager [None req-09f1eb26-3ef8-431d-92bc-9c62b712a196 tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] [instance: 60ec4271-2614-470d-bf40-47b9955f544c] Unplugging VIFs for instance {{(pid=61957) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 778.967617] env[61957]: DEBUG oslo_concurrency.lockutils [None req-09f1eb26-3ef8-431d-92bc-9c62b712a196 tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] Acquiring lock "refresh_cache-60ec4271-2614-470d-bf40-47b9955f544c" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 778.967781] env[61957]: DEBUG oslo_concurrency.lockutils [None req-09f1eb26-3ef8-431d-92bc-9c62b712a196 tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] Acquired lock "refresh_cache-60ec4271-2614-470d-bf40-47b9955f544c" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 778.968764] env[61957]: DEBUG nova.network.neutron [None req-09f1eb26-3ef8-431d-92bc-9c62b712a196 tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] [instance: 60ec4271-2614-470d-bf40-47b9955f544c] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 779.024690] env[61957]: DEBUG nova.compute.manager [None req-7860e034-eff8-489c-addf-33cf2776246c tempest-ServerActionsTestOtherA-1610268163 tempest-ServerActionsTestOtherA-1610268163-project-member] [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] Start spawning the instance on the hypervisor. {{(pid=61957) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 779.054887] env[61957]: DEBUG nova.virt.hardware [None req-7860e034-eff8-489c-addf-33cf2776246c tempest-ServerActionsTestOtherA-1610268163 tempest-ServerActionsTestOtherA-1610268163-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T17:22:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T17:21:50Z,direct_url=,disk_format='vmdk',id=11c76a2c-f705-470a-ba9d-4657858bab38,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='81d3096d0f094373913fc3dc8f5c3c6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T17:21:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 779.055317] env[61957]: DEBUG nova.virt.hardware [None req-7860e034-eff8-489c-addf-33cf2776246c tempest-ServerActionsTestOtherA-1610268163 tempest-ServerActionsTestOtherA-1610268163-project-member] Flavor limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 779.055578] env[61957]: DEBUG nova.virt.hardware [None req-7860e034-eff8-489c-addf-33cf2776246c tempest-ServerActionsTestOtherA-1610268163 tempest-ServerActionsTestOtherA-1610268163-project-member] Image limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 779.055919] env[61957]: DEBUG nova.virt.hardware [None req-7860e034-eff8-489c-addf-33cf2776246c tempest-ServerActionsTestOtherA-1610268163 tempest-ServerActionsTestOtherA-1610268163-project-member] Flavor pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 779.056043] env[61957]: DEBUG nova.virt.hardware [None req-7860e034-eff8-489c-addf-33cf2776246c tempest-ServerActionsTestOtherA-1610268163 tempest-ServerActionsTestOtherA-1610268163-project-member] Image pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 779.056270] env[61957]: DEBUG nova.virt.hardware [None req-7860e034-eff8-489c-addf-33cf2776246c tempest-ServerActionsTestOtherA-1610268163 tempest-ServerActionsTestOtherA-1610268163-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 779.056563] env[61957]: DEBUG nova.virt.hardware [None req-7860e034-eff8-489c-addf-33cf2776246c tempest-ServerActionsTestOtherA-1610268163 tempest-ServerActionsTestOtherA-1610268163-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 779.056871] env[61957]: DEBUG nova.virt.hardware [None req-7860e034-eff8-489c-addf-33cf2776246c tempest-ServerActionsTestOtherA-1610268163 tempest-ServerActionsTestOtherA-1610268163-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 779.057049] env[61957]: DEBUG nova.virt.hardware [None req-7860e034-eff8-489c-addf-33cf2776246c tempest-ServerActionsTestOtherA-1610268163 tempest-ServerActionsTestOtherA-1610268163-project-member] Got 1 possible topologies {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 779.057283] env[61957]: DEBUG nova.virt.hardware [None req-7860e034-eff8-489c-addf-33cf2776246c tempest-ServerActionsTestOtherA-1610268163 tempest-ServerActionsTestOtherA-1610268163-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 779.057609] env[61957]: DEBUG nova.virt.hardware [None req-7860e034-eff8-489c-addf-33cf2776246c tempest-ServerActionsTestOtherA-1610268163 tempest-ServerActionsTestOtherA-1610268163-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 779.058512] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdc462df-eb7b-4459-bb08-004d9510a06b {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.066951] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-154edc4c-7be1-49f6-a5f3-d538d33bc8fa {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.081253] env[61957]: ERROR nova.compute.manager [None req-7860e034-eff8-489c-addf-33cf2776246c tempest-ServerActionsTestOtherA-1610268163 tempest-ServerActionsTestOtherA-1610268163-project-member] [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 76ccb95a-3bfc-4be2-a607-679432fcf74d, please check neutron logs for more information. [ 779.081253] env[61957]: ERROR nova.compute.manager [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] Traceback (most recent call last): [ 779.081253] env[61957]: ERROR nova.compute.manager [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 779.081253] env[61957]: ERROR nova.compute.manager [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] yield resources [ 779.081253] env[61957]: ERROR nova.compute.manager [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 779.081253] env[61957]: ERROR nova.compute.manager [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] self.driver.spawn(context, instance, image_meta, [ 779.081253] env[61957]: ERROR nova.compute.manager [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 779.081253] env[61957]: ERROR nova.compute.manager [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] self._vmops.spawn(context, instance, image_meta, injected_files, [ 779.081253] env[61957]: ERROR nova.compute.manager [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 779.081253] env[61957]: ERROR nova.compute.manager [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] vm_ref = self.build_virtual_machine(instance, [ 779.081253] env[61957]: ERROR nova.compute.manager [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 779.081686] env[61957]: ERROR nova.compute.manager [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] vif_infos = vmwarevif.get_vif_info(self._session, [ 779.081686] env[61957]: ERROR nova.compute.manager [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 779.081686] env[61957]: ERROR nova.compute.manager [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] for vif in network_info: [ 779.081686] env[61957]: ERROR nova.compute.manager [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 779.081686] env[61957]: ERROR nova.compute.manager [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] return self._sync_wrapper(fn, *args, **kwargs) [ 779.081686] env[61957]: ERROR nova.compute.manager [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 779.081686] env[61957]: ERROR nova.compute.manager [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] self.wait() [ 779.081686] env[61957]: ERROR nova.compute.manager [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 779.081686] env[61957]: ERROR nova.compute.manager [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] self[:] = self._gt.wait() [ 779.081686] env[61957]: ERROR nova.compute.manager [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 779.081686] env[61957]: ERROR nova.compute.manager [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] return self._exit_event.wait() [ 779.081686] env[61957]: ERROR nova.compute.manager [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 779.081686] env[61957]: ERROR nova.compute.manager [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] current.throw(*self._exc) [ 779.082077] env[61957]: ERROR nova.compute.manager [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 779.082077] env[61957]: ERROR nova.compute.manager [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] result = function(*args, **kwargs) [ 779.082077] env[61957]: ERROR nova.compute.manager [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 779.082077] env[61957]: ERROR nova.compute.manager [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] return func(*args, **kwargs) [ 779.082077] env[61957]: ERROR nova.compute.manager [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 779.082077] env[61957]: ERROR nova.compute.manager [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] raise e [ 779.082077] env[61957]: ERROR nova.compute.manager [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 779.082077] env[61957]: ERROR nova.compute.manager [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] nwinfo = self.network_api.allocate_for_instance( [ 779.082077] env[61957]: ERROR nova.compute.manager [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 779.082077] env[61957]: ERROR nova.compute.manager [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] created_port_ids = self._update_ports_for_instance( [ 779.082077] env[61957]: ERROR nova.compute.manager [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 779.082077] env[61957]: ERROR nova.compute.manager [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] with excutils.save_and_reraise_exception(): [ 779.082077] env[61957]: ERROR nova.compute.manager [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 779.082481] env[61957]: ERROR nova.compute.manager [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] self.force_reraise() [ 779.082481] env[61957]: ERROR nova.compute.manager [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 779.082481] env[61957]: ERROR nova.compute.manager [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] raise self.value [ 779.082481] env[61957]: ERROR nova.compute.manager [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 779.082481] env[61957]: ERROR nova.compute.manager [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] updated_port = self._update_port( [ 779.082481] env[61957]: ERROR nova.compute.manager [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 779.082481] env[61957]: ERROR nova.compute.manager [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] _ensure_no_port_binding_failure(port) [ 779.082481] env[61957]: ERROR nova.compute.manager [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 779.082481] env[61957]: ERROR nova.compute.manager [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] raise exception.PortBindingFailed(port_id=port['id']) [ 779.082481] env[61957]: ERROR nova.compute.manager [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] nova.exception.PortBindingFailed: Binding failed for port 76ccb95a-3bfc-4be2-a607-679432fcf74d, please check neutron logs for more information. [ 779.082481] env[61957]: ERROR nova.compute.manager [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] [ 779.082481] env[61957]: INFO nova.compute.manager [None req-7860e034-eff8-489c-addf-33cf2776246c tempest-ServerActionsTestOtherA-1610268163 tempest-ServerActionsTestOtherA-1610268163-project-member] [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] Terminating instance [ 779.084245] env[61957]: DEBUG oslo_concurrency.lockutils [None req-7860e034-eff8-489c-addf-33cf2776246c tempest-ServerActionsTestOtherA-1610268163 tempest-ServerActionsTestOtherA-1610268163-project-member] Acquiring lock "refresh_cache-6372a384-61a8-43e8-8f6c-376e2a735045" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 779.084447] env[61957]: DEBUG oslo_concurrency.lockutils [None req-7860e034-eff8-489c-addf-33cf2776246c tempest-ServerActionsTestOtherA-1610268163 tempest-ServerActionsTestOtherA-1610268163-project-member] Acquired lock "refresh_cache-6372a384-61a8-43e8-8f6c-376e2a735045" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 779.084650] env[61957]: DEBUG nova.network.neutron [None req-7860e034-eff8-489c-addf-33cf2776246c tempest-ServerActionsTestOtherA-1610268163 tempest-ServerActionsTestOtherA-1610268163-project-member] [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 779.181438] env[61957]: DEBUG oslo_concurrency.lockutils [req-3d556a74-167d-4258-990e-81b2bf33b271 req-4a84d6be-c486-448b-9aa3-0986bf966a5a service nova] Releasing lock "refresh_cache-2f84d2df-00f0-4860-acb4-4e10ecf8ce4e" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 779.181858] env[61957]: DEBUG nova.compute.manager [req-3d556a74-167d-4258-990e-81b2bf33b271 req-4a84d6be-c486-448b-9aa3-0986bf966a5a service nova] [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] Received event network-vif-deleted-5ec2282a-722a-45ab-9124-906367749fe7 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 779.495617] env[61957]: DEBUG nova.network.neutron [None req-09f1eb26-3ef8-431d-92bc-9c62b712a196 tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] [instance: 60ec4271-2614-470d-bf40-47b9955f544c] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 779.610446] env[61957]: DEBUG nova.network.neutron [None req-7860e034-eff8-489c-addf-33cf2776246c tempest-ServerActionsTestOtherA-1610268163 tempest-ServerActionsTestOtherA-1610268163-project-member] [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 779.624428] env[61957]: DEBUG nova.network.neutron [None req-09f1eb26-3ef8-431d-92bc-9c62b712a196 tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] [instance: 60ec4271-2614-470d-bf40-47b9955f544c] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 779.708208] env[61957]: DEBUG nova.network.neutron [None req-7860e034-eff8-489c-addf-33cf2776246c tempest-ServerActionsTestOtherA-1610268163 tempest-ServerActionsTestOtherA-1610268163-project-member] [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 779.883576] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc0c5d85-1848-4f29-ae56-67a195a6c30c {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.894899] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-406a76e9-cf0d-434d-a123-c4c8ca900dd0 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.929821] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a79467d2-5de5-4d11-9cac-6818acbf7d94 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.938646] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e9b490c-cbb7-4637-9e99-04d3fc2e06e2 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.952431] env[61957]: DEBUG nova.compute.provider_tree [None req-6b91bce8-d0f8-4c40-855e-ed34f3eb7e64 tempest-ServerActionsTestJSON-1567241756 tempest-ServerActionsTestJSON-1567241756-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 780.053417] env[61957]: DEBUG nova.compute.manager [req-9827a022-d24d-4319-bc93-5488304c0d3e req-478a7dba-fd94-4e32-b3df-87ee1d8f8556 service nova] [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] Received event network-changed-76ccb95a-3bfc-4be2-a607-679432fcf74d {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 780.053590] env[61957]: DEBUG nova.compute.manager [req-9827a022-d24d-4319-bc93-5488304c0d3e req-478a7dba-fd94-4e32-b3df-87ee1d8f8556 service nova] [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] Refreshing instance network info cache due to event network-changed-76ccb95a-3bfc-4be2-a607-679432fcf74d. {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 780.053781] env[61957]: DEBUG oslo_concurrency.lockutils [req-9827a022-d24d-4319-bc93-5488304c0d3e req-478a7dba-fd94-4e32-b3df-87ee1d8f8556 service nova] Acquiring lock "refresh_cache-6372a384-61a8-43e8-8f6c-376e2a735045" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 780.126951] env[61957]: DEBUG oslo_concurrency.lockutils [None req-09f1eb26-3ef8-431d-92bc-9c62b712a196 tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] Releasing lock "refresh_cache-60ec4271-2614-470d-bf40-47b9955f544c" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 780.127258] env[61957]: DEBUG nova.compute.manager [None req-09f1eb26-3ef8-431d-92bc-9c62b712a196 tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61957) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 780.127411] env[61957]: DEBUG nova.compute.manager [None req-09f1eb26-3ef8-431d-92bc-9c62b712a196 tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] [instance: 60ec4271-2614-470d-bf40-47b9955f544c] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 780.127578] env[61957]: DEBUG nova.network.neutron [None req-09f1eb26-3ef8-431d-92bc-9c62b712a196 tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] [instance: 60ec4271-2614-470d-bf40-47b9955f544c] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 780.145739] env[61957]: DEBUG nova.network.neutron [None req-09f1eb26-3ef8-431d-92bc-9c62b712a196 tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] [instance: 60ec4271-2614-470d-bf40-47b9955f544c] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 780.211928] env[61957]: DEBUG oslo_concurrency.lockutils [None req-7860e034-eff8-489c-addf-33cf2776246c tempest-ServerActionsTestOtherA-1610268163 tempest-ServerActionsTestOtherA-1610268163-project-member] Releasing lock "refresh_cache-6372a384-61a8-43e8-8f6c-376e2a735045" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 780.212405] env[61957]: DEBUG nova.compute.manager [None req-7860e034-eff8-489c-addf-33cf2776246c tempest-ServerActionsTestOtherA-1610268163 tempest-ServerActionsTestOtherA-1610268163-project-member] [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] Start destroying the instance on the hypervisor. {{(pid=61957) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 780.212598] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-7860e034-eff8-489c-addf-33cf2776246c tempest-ServerActionsTestOtherA-1610268163 tempest-ServerActionsTestOtherA-1610268163-project-member] [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] Destroying instance {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 780.212920] env[61957]: DEBUG oslo_concurrency.lockutils [req-9827a022-d24d-4319-bc93-5488304c0d3e req-478a7dba-fd94-4e32-b3df-87ee1d8f8556 service nova] Acquired lock "refresh_cache-6372a384-61a8-43e8-8f6c-376e2a735045" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 780.213114] env[61957]: DEBUG nova.network.neutron [req-9827a022-d24d-4319-bc93-5488304c0d3e req-478a7dba-fd94-4e32-b3df-87ee1d8f8556 service nova] [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] Refreshing network info cache for port 76ccb95a-3bfc-4be2-a607-679432fcf74d {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 780.214219] env[61957]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0595c93c-11f4-4a71-8b95-56b1ea1f5067 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.226090] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54a1b834-53b8-4bfb-8bc1-cdbe79713619 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.251745] env[61957]: WARNING nova.virt.vmwareapi.vmops [None req-7860e034-eff8-489c-addf-33cf2776246c tempest-ServerActionsTestOtherA-1610268163 tempest-ServerActionsTestOtherA-1610268163-project-member] [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 6372a384-61a8-43e8-8f6c-376e2a735045 could not be found. [ 780.251973] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-7860e034-eff8-489c-addf-33cf2776246c tempest-ServerActionsTestOtherA-1610268163 tempest-ServerActionsTestOtherA-1610268163-project-member] [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] Instance destroyed {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 780.252328] env[61957]: INFO nova.compute.manager [None req-7860e034-eff8-489c-addf-33cf2776246c tempest-ServerActionsTestOtherA-1610268163 tempest-ServerActionsTestOtherA-1610268163-project-member] [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] Took 0.04 seconds to destroy the instance on the hypervisor. [ 780.252579] env[61957]: DEBUG oslo.service.loopingcall [None req-7860e034-eff8-489c-addf-33cf2776246c tempest-ServerActionsTestOtherA-1610268163 tempest-ServerActionsTestOtherA-1610268163-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 780.252792] env[61957]: DEBUG nova.compute.manager [-] [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 780.252883] env[61957]: DEBUG nova.network.neutron [-] [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 780.274317] env[61957]: DEBUG nova.network.neutron [-] [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 780.455647] env[61957]: DEBUG nova.scheduler.client.report [None req-6b91bce8-d0f8-4c40-855e-ed34f3eb7e64 tempest-ServerActionsTestJSON-1567241756 tempest-ServerActionsTestJSON-1567241756-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 780.648893] env[61957]: DEBUG nova.network.neutron [None req-09f1eb26-3ef8-431d-92bc-9c62b712a196 tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] [instance: 60ec4271-2614-470d-bf40-47b9955f544c] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 780.739378] env[61957]: DEBUG nova.network.neutron [req-9827a022-d24d-4319-bc93-5488304c0d3e req-478a7dba-fd94-4e32-b3df-87ee1d8f8556 service nova] [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 780.777711] env[61957]: DEBUG nova.network.neutron [-] [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 780.823887] env[61957]: DEBUG nova.network.neutron [req-9827a022-d24d-4319-bc93-5488304c0d3e req-478a7dba-fd94-4e32-b3df-87ee1d8f8556 service nova] [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 780.964717] env[61957]: DEBUG oslo_concurrency.lockutils [None req-6b91bce8-d0f8-4c40-855e-ed34f3eb7e64 tempest-ServerActionsTestJSON-1567241756 tempest-ServerActionsTestJSON-1567241756-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.001s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 780.965369] env[61957]: ERROR nova.compute.manager [None req-6b91bce8-d0f8-4c40-855e-ed34f3eb7e64 tempest-ServerActionsTestJSON-1567241756 tempest-ServerActionsTestJSON-1567241756-project-member] [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 3233fc75-1990-45ba-992a-b932f2d5eafc, please check neutron logs for more information. [ 780.965369] env[61957]: ERROR nova.compute.manager [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] Traceback (most recent call last): [ 780.965369] env[61957]: ERROR nova.compute.manager [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 780.965369] env[61957]: ERROR nova.compute.manager [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] self.driver.spawn(context, instance, image_meta, [ 780.965369] env[61957]: ERROR nova.compute.manager [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 780.965369] env[61957]: ERROR nova.compute.manager [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 780.965369] env[61957]: ERROR nova.compute.manager [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 780.965369] env[61957]: ERROR nova.compute.manager [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] vm_ref = self.build_virtual_machine(instance, [ 780.965369] env[61957]: ERROR nova.compute.manager [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 780.965369] env[61957]: ERROR nova.compute.manager [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] vif_infos = vmwarevif.get_vif_info(self._session, [ 780.965369] env[61957]: ERROR nova.compute.manager [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 780.965742] env[61957]: ERROR nova.compute.manager [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] for vif in network_info: [ 780.965742] env[61957]: ERROR nova.compute.manager [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 780.965742] env[61957]: ERROR nova.compute.manager [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] return self._sync_wrapper(fn, *args, **kwargs) [ 780.965742] env[61957]: ERROR nova.compute.manager [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 780.965742] env[61957]: ERROR nova.compute.manager [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] self.wait() [ 780.965742] env[61957]: ERROR nova.compute.manager [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 780.965742] env[61957]: ERROR nova.compute.manager [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] self[:] = self._gt.wait() [ 780.965742] env[61957]: ERROR nova.compute.manager [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 780.965742] env[61957]: ERROR nova.compute.manager [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] return self._exit_event.wait() [ 780.965742] env[61957]: ERROR nova.compute.manager [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 780.965742] env[61957]: ERROR nova.compute.manager [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] current.throw(*self._exc) [ 780.965742] env[61957]: ERROR nova.compute.manager [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 780.965742] env[61957]: ERROR nova.compute.manager [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] result = function(*args, **kwargs) [ 780.966159] env[61957]: ERROR nova.compute.manager [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 780.966159] env[61957]: ERROR nova.compute.manager [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] return func(*args, **kwargs) [ 780.966159] env[61957]: ERROR nova.compute.manager [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 780.966159] env[61957]: ERROR nova.compute.manager [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] raise e [ 780.966159] env[61957]: ERROR nova.compute.manager [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 780.966159] env[61957]: ERROR nova.compute.manager [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] nwinfo = self.network_api.allocate_for_instance( [ 780.966159] env[61957]: ERROR nova.compute.manager [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 780.966159] env[61957]: ERROR nova.compute.manager [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] created_port_ids = self._update_ports_for_instance( [ 780.966159] env[61957]: ERROR nova.compute.manager [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 780.966159] env[61957]: ERROR nova.compute.manager [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] with excutils.save_and_reraise_exception(): [ 780.966159] env[61957]: ERROR nova.compute.manager [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 780.966159] env[61957]: ERROR nova.compute.manager [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] self.force_reraise() [ 780.966159] env[61957]: ERROR nova.compute.manager [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 780.966559] env[61957]: ERROR nova.compute.manager [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] raise self.value [ 780.966559] env[61957]: ERROR nova.compute.manager [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 780.966559] env[61957]: ERROR nova.compute.manager [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] updated_port = self._update_port( [ 780.966559] env[61957]: ERROR nova.compute.manager [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 780.966559] env[61957]: ERROR nova.compute.manager [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] _ensure_no_port_binding_failure(port) [ 780.966559] env[61957]: ERROR nova.compute.manager [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 780.966559] env[61957]: ERROR nova.compute.manager [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] raise exception.PortBindingFailed(port_id=port['id']) [ 780.966559] env[61957]: ERROR nova.compute.manager [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] nova.exception.PortBindingFailed: Binding failed for port 3233fc75-1990-45ba-992a-b932f2d5eafc, please check neutron logs for more information. [ 780.966559] env[61957]: ERROR nova.compute.manager [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] [ 780.966559] env[61957]: DEBUG nova.compute.utils [None req-6b91bce8-d0f8-4c40-855e-ed34f3eb7e64 tempest-ServerActionsTestJSON-1567241756 tempest-ServerActionsTestJSON-1567241756-project-member] [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] Binding failed for port 3233fc75-1990-45ba-992a-b932f2d5eafc, please check neutron logs for more information. {{(pid=61957) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 780.967723] env[61957]: DEBUG oslo_concurrency.lockutils [None req-660095f0-9d5a-4797-b621-e0585ffddbba tempest-ServerActionsTestOtherB-303558607 tempest-ServerActionsTestOtherB-303558607-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.816s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 780.971012] env[61957]: DEBUG nova.compute.manager [None req-6b91bce8-d0f8-4c40-855e-ed34f3eb7e64 tempest-ServerActionsTestJSON-1567241756 tempest-ServerActionsTestJSON-1567241756-project-member] [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] Build of instance e69a7833-97b1-4063-a6e8-c4558980ee1e was re-scheduled: Binding failed for port 3233fc75-1990-45ba-992a-b932f2d5eafc, please check neutron logs for more information. {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 780.971426] env[61957]: DEBUG nova.compute.manager [None req-6b91bce8-d0f8-4c40-855e-ed34f3eb7e64 tempest-ServerActionsTestJSON-1567241756 tempest-ServerActionsTestJSON-1567241756-project-member] [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] Unplugging VIFs for instance {{(pid=61957) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 780.971643] env[61957]: DEBUG oslo_concurrency.lockutils [None req-6b91bce8-d0f8-4c40-855e-ed34f3eb7e64 tempest-ServerActionsTestJSON-1567241756 tempest-ServerActionsTestJSON-1567241756-project-member] Acquiring lock "refresh_cache-e69a7833-97b1-4063-a6e8-c4558980ee1e" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 780.971784] env[61957]: DEBUG oslo_concurrency.lockutils [None req-6b91bce8-d0f8-4c40-855e-ed34f3eb7e64 tempest-ServerActionsTestJSON-1567241756 tempest-ServerActionsTestJSON-1567241756-project-member] Acquired lock "refresh_cache-e69a7833-97b1-4063-a6e8-c4558980ee1e" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 780.971938] env[61957]: DEBUG nova.network.neutron [None req-6b91bce8-d0f8-4c40-855e-ed34f3eb7e64 tempest-ServerActionsTestJSON-1567241756 tempest-ServerActionsTestJSON-1567241756-project-member] [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 781.159642] env[61957]: INFO nova.compute.manager [None req-09f1eb26-3ef8-431d-92bc-9c62b712a196 tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] [instance: 60ec4271-2614-470d-bf40-47b9955f544c] Took 1.03 seconds to deallocate network for instance. [ 781.279604] env[61957]: INFO nova.compute.manager [-] [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] Took 1.03 seconds to deallocate network for instance. [ 781.282960] env[61957]: DEBUG nova.compute.claims [None req-7860e034-eff8-489c-addf-33cf2776246c tempest-ServerActionsTestOtherA-1610268163 tempest-ServerActionsTestOtherA-1610268163-project-member] [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] Aborting claim: {{(pid=61957) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 781.283168] env[61957]: DEBUG oslo_concurrency.lockutils [None req-7860e034-eff8-489c-addf-33cf2776246c tempest-ServerActionsTestOtherA-1610268163 tempest-ServerActionsTestOtherA-1610268163-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 781.329085] env[61957]: DEBUG oslo_concurrency.lockutils [req-9827a022-d24d-4319-bc93-5488304c0d3e req-478a7dba-fd94-4e32-b3df-87ee1d8f8556 service nova] Releasing lock "refresh_cache-6372a384-61a8-43e8-8f6c-376e2a735045" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 781.329398] env[61957]: DEBUG nova.compute.manager [req-9827a022-d24d-4319-bc93-5488304c0d3e req-478a7dba-fd94-4e32-b3df-87ee1d8f8556 service nova] [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] Received event network-vif-deleted-76ccb95a-3bfc-4be2-a607-679432fcf74d {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 781.491452] env[61957]: DEBUG nova.network.neutron [None req-6b91bce8-d0f8-4c40-855e-ed34f3eb7e64 tempest-ServerActionsTestJSON-1567241756 tempest-ServerActionsTestJSON-1567241756-project-member] [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 781.586122] env[61957]: DEBUG nova.network.neutron [None req-6b91bce8-d0f8-4c40-855e-ed34f3eb7e64 tempest-ServerActionsTestJSON-1567241756 tempest-ServerActionsTestJSON-1567241756-project-member] [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 781.814126] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b02e47ae-9e52-4d75-8f9f-3ae457dd5aa5 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.822269] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ad41261-09c7-47cb-955c-0786219e7c26 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.858411] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35e56453-bf42-4e5d-a3ec-59b14d5471fd {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.867235] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f0acb49-fc3f-4b17-8a61-d9d7576a292f {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.884450] env[61957]: DEBUG oslo_concurrency.lockutils [None req-09e8eb69-1030-4228-8cb0-915d5cee1a52 tempest-AttachInterfacesV270Test-2032541732 tempest-AttachInterfacesV270Test-2032541732-project-member] Acquiring lock "281b9644-0e7d-48b9-a7b6-45fd6102d558" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 781.887860] env[61957]: DEBUG oslo_concurrency.lockutils [None req-09e8eb69-1030-4228-8cb0-915d5cee1a52 tempest-AttachInterfacesV270Test-2032541732 tempest-AttachInterfacesV270Test-2032541732-project-member] Lock "281b9644-0e7d-48b9-a7b6-45fd6102d558" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 781.887860] env[61957]: DEBUG nova.compute.provider_tree [None req-660095f0-9d5a-4797-b621-e0585ffddbba tempest-ServerActionsTestOtherB-303558607 tempest-ServerActionsTestOtherB-303558607-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 782.088664] env[61957]: DEBUG oslo_concurrency.lockutils [None req-6b91bce8-d0f8-4c40-855e-ed34f3eb7e64 tempest-ServerActionsTestJSON-1567241756 tempest-ServerActionsTestJSON-1567241756-project-member] Releasing lock "refresh_cache-e69a7833-97b1-4063-a6e8-c4558980ee1e" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 782.088944] env[61957]: DEBUG nova.compute.manager [None req-6b91bce8-d0f8-4c40-855e-ed34f3eb7e64 tempest-ServerActionsTestJSON-1567241756 tempest-ServerActionsTestJSON-1567241756-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61957) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 782.089191] env[61957]: DEBUG nova.compute.manager [None req-6b91bce8-d0f8-4c40-855e-ed34f3eb7e64 tempest-ServerActionsTestJSON-1567241756 tempest-ServerActionsTestJSON-1567241756-project-member] [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 782.089372] env[61957]: DEBUG nova.network.neutron [None req-6b91bce8-d0f8-4c40-855e-ed34f3eb7e64 tempest-ServerActionsTestJSON-1567241756 tempest-ServerActionsTestJSON-1567241756-project-member] [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 782.104101] env[61957]: DEBUG nova.network.neutron [None req-6b91bce8-d0f8-4c40-855e-ed34f3eb7e64 tempest-ServerActionsTestJSON-1567241756 tempest-ServerActionsTestJSON-1567241756-project-member] [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 782.190196] env[61957]: INFO nova.scheduler.client.report [None req-09f1eb26-3ef8-431d-92bc-9c62b712a196 tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] Deleted allocations for instance 60ec4271-2614-470d-bf40-47b9955f544c [ 782.388214] env[61957]: DEBUG nova.scheduler.client.report [None req-660095f0-9d5a-4797-b621-e0585ffddbba tempest-ServerActionsTestOtherB-303558607 tempest-ServerActionsTestOtherB-303558607-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 782.607254] env[61957]: DEBUG nova.network.neutron [None req-6b91bce8-d0f8-4c40-855e-ed34f3eb7e64 tempest-ServerActionsTestJSON-1567241756 tempest-ServerActionsTestJSON-1567241756-project-member] [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 782.700146] env[61957]: DEBUG oslo_concurrency.lockutils [None req-09f1eb26-3ef8-431d-92bc-9c62b712a196 tempest-SecurityGroupsTestJSON-86884773 tempest-SecurityGroupsTestJSON-86884773-project-member] Lock "60ec4271-2614-470d-bf40-47b9955f544c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 147.832s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 782.895290] env[61957]: DEBUG oslo_concurrency.lockutils [None req-660095f0-9d5a-4797-b621-e0585ffddbba tempest-ServerActionsTestOtherB-303558607 tempest-ServerActionsTestOtherB-303558607-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.928s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 782.895927] env[61957]: ERROR nova.compute.manager [None req-660095f0-9d5a-4797-b621-e0585ffddbba tempest-ServerActionsTestOtherB-303558607 tempest-ServerActionsTestOtherB-303558607-project-member] [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 5d6fb25b-5ed5-4828-acbe-5d639e8419fa, please check neutron logs for more information. [ 782.895927] env[61957]: ERROR nova.compute.manager [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] Traceback (most recent call last): [ 782.895927] env[61957]: ERROR nova.compute.manager [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 782.895927] env[61957]: ERROR nova.compute.manager [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] self.driver.spawn(context, instance, image_meta, [ 782.895927] env[61957]: ERROR nova.compute.manager [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 782.895927] env[61957]: ERROR nova.compute.manager [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] self._vmops.spawn(context, instance, image_meta, injected_files, [ 782.895927] env[61957]: ERROR nova.compute.manager [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 782.895927] env[61957]: ERROR nova.compute.manager [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] vm_ref = self.build_virtual_machine(instance, [ 782.895927] env[61957]: ERROR nova.compute.manager [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 782.895927] env[61957]: ERROR nova.compute.manager [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] vif_infos = vmwarevif.get_vif_info(self._session, [ 782.895927] env[61957]: ERROR nova.compute.manager [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 782.896556] env[61957]: ERROR nova.compute.manager [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] for vif in network_info: [ 782.896556] env[61957]: ERROR nova.compute.manager [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 782.896556] env[61957]: ERROR nova.compute.manager [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] return self._sync_wrapper(fn, *args, **kwargs) [ 782.896556] env[61957]: ERROR nova.compute.manager [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 782.896556] env[61957]: ERROR nova.compute.manager [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] self.wait() [ 782.896556] env[61957]: ERROR nova.compute.manager [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 782.896556] env[61957]: ERROR nova.compute.manager [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] self[:] = self._gt.wait() [ 782.896556] env[61957]: ERROR nova.compute.manager [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 782.896556] env[61957]: ERROR nova.compute.manager [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] return self._exit_event.wait() [ 782.896556] env[61957]: ERROR nova.compute.manager [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 782.896556] env[61957]: ERROR nova.compute.manager [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] current.throw(*self._exc) [ 782.896556] env[61957]: ERROR nova.compute.manager [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 782.896556] env[61957]: ERROR nova.compute.manager [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] result = function(*args, **kwargs) [ 782.898507] env[61957]: ERROR nova.compute.manager [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 782.898507] env[61957]: ERROR nova.compute.manager [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] return func(*args, **kwargs) [ 782.898507] env[61957]: ERROR nova.compute.manager [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 782.898507] env[61957]: ERROR nova.compute.manager [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] raise e [ 782.898507] env[61957]: ERROR nova.compute.manager [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 782.898507] env[61957]: ERROR nova.compute.manager [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] nwinfo = self.network_api.allocate_for_instance( [ 782.898507] env[61957]: ERROR nova.compute.manager [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 782.898507] env[61957]: ERROR nova.compute.manager [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] created_port_ids = self._update_ports_for_instance( [ 782.898507] env[61957]: ERROR nova.compute.manager [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 782.898507] env[61957]: ERROR nova.compute.manager [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] with excutils.save_and_reraise_exception(): [ 782.898507] env[61957]: ERROR nova.compute.manager [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 782.898507] env[61957]: ERROR nova.compute.manager [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] self.force_reraise() [ 782.898507] env[61957]: ERROR nova.compute.manager [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 782.899274] env[61957]: ERROR nova.compute.manager [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] raise self.value [ 782.899274] env[61957]: ERROR nova.compute.manager [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 782.899274] env[61957]: ERROR nova.compute.manager [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] updated_port = self._update_port( [ 782.899274] env[61957]: ERROR nova.compute.manager [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 782.899274] env[61957]: ERROR nova.compute.manager [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] _ensure_no_port_binding_failure(port) [ 782.899274] env[61957]: ERROR nova.compute.manager [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 782.899274] env[61957]: ERROR nova.compute.manager [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] raise exception.PortBindingFailed(port_id=port['id']) [ 782.899274] env[61957]: ERROR nova.compute.manager [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] nova.exception.PortBindingFailed: Binding failed for port 5d6fb25b-5ed5-4828-acbe-5d639e8419fa, please check neutron logs for more information. [ 782.899274] env[61957]: ERROR nova.compute.manager [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] [ 782.899274] env[61957]: DEBUG nova.compute.utils [None req-660095f0-9d5a-4797-b621-e0585ffddbba tempest-ServerActionsTestOtherB-303558607 tempest-ServerActionsTestOtherB-303558607-project-member] [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] Binding failed for port 5d6fb25b-5ed5-4828-acbe-5d639e8419fa, please check neutron logs for more information. {{(pid=61957) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 782.899746] env[61957]: DEBUG oslo_concurrency.lockutils [None req-3617f05d-8cb2-4824-b3d0-fbcde866b375 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 15.071s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 782.899746] env[61957]: DEBUG nova.objects.instance [None req-3617f05d-8cb2-4824-b3d0-fbcde866b375 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] [instance: 7f3850c0-a455-4e5b-a329-fb3b02f9725f] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61957) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 782.902029] env[61957]: DEBUG nova.compute.manager [None req-660095f0-9d5a-4797-b621-e0585ffddbba tempest-ServerActionsTestOtherB-303558607 tempest-ServerActionsTestOtherB-303558607-project-member] [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] Build of instance d77276ba-6d37-440b-a113-c4f62b2e7946 was re-scheduled: Binding failed for port 5d6fb25b-5ed5-4828-acbe-5d639e8419fa, please check neutron logs for more information. {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 782.902482] env[61957]: DEBUG nova.compute.manager [None req-660095f0-9d5a-4797-b621-e0585ffddbba tempest-ServerActionsTestOtherB-303558607 tempest-ServerActionsTestOtherB-303558607-project-member] [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] Unplugging VIFs for instance {{(pid=61957) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 782.903860] env[61957]: DEBUG oslo_concurrency.lockutils [None req-660095f0-9d5a-4797-b621-e0585ffddbba tempest-ServerActionsTestOtherB-303558607 tempest-ServerActionsTestOtherB-303558607-project-member] Acquiring lock "refresh_cache-d77276ba-6d37-440b-a113-c4f62b2e7946" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 782.903860] env[61957]: DEBUG oslo_concurrency.lockutils [None req-660095f0-9d5a-4797-b621-e0585ffddbba tempest-ServerActionsTestOtherB-303558607 tempest-ServerActionsTestOtherB-303558607-project-member] Acquired lock "refresh_cache-d77276ba-6d37-440b-a113-c4f62b2e7946" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 782.903860] env[61957]: DEBUG nova.network.neutron [None req-660095f0-9d5a-4797-b621-e0585ffddbba tempest-ServerActionsTestOtherB-303558607 tempest-ServerActionsTestOtherB-303558607-project-member] [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 783.110252] env[61957]: INFO nova.compute.manager [None req-6b91bce8-d0f8-4c40-855e-ed34f3eb7e64 tempest-ServerActionsTestJSON-1567241756 tempest-ServerActionsTestJSON-1567241756-project-member] [instance: e69a7833-97b1-4063-a6e8-c4558980ee1e] Took 1.02 seconds to deallocate network for instance. [ 783.204987] env[61957]: DEBUG nova.compute.manager [None req-182a8e01-a1e1-4ecf-8692-2abcfc9944ba tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] [instance: e0249936-d616-4ffb-8f77-d8107633c42a] Starting instance... {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 783.424102] env[61957]: DEBUG nova.network.neutron [None req-660095f0-9d5a-4797-b621-e0585ffddbba tempest-ServerActionsTestOtherB-303558607 tempest-ServerActionsTestOtherB-303558607-project-member] [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 783.510477] env[61957]: DEBUG nova.network.neutron [None req-660095f0-9d5a-4797-b621-e0585ffddbba tempest-ServerActionsTestOtherB-303558607 tempest-ServerActionsTestOtherB-303558607-project-member] [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 783.726271] env[61957]: DEBUG oslo_concurrency.lockutils [None req-182a8e01-a1e1-4ecf-8692-2abcfc9944ba tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 783.911268] env[61957]: DEBUG oslo_concurrency.lockutils [None req-3617f05d-8cb2-4824-b3d0-fbcde866b375 tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.013s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 783.912359] env[61957]: DEBUG oslo_concurrency.lockutils [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 14.156s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 784.013526] env[61957]: DEBUG oslo_concurrency.lockutils [None req-660095f0-9d5a-4797-b621-e0585ffddbba tempest-ServerActionsTestOtherB-303558607 tempest-ServerActionsTestOtherB-303558607-project-member] Releasing lock "refresh_cache-d77276ba-6d37-440b-a113-c4f62b2e7946" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 784.013771] env[61957]: DEBUG nova.compute.manager [None req-660095f0-9d5a-4797-b621-e0585ffddbba tempest-ServerActionsTestOtherB-303558607 tempest-ServerActionsTestOtherB-303558607-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61957) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 784.013954] env[61957]: DEBUG nova.compute.manager [None req-660095f0-9d5a-4797-b621-e0585ffddbba tempest-ServerActionsTestOtherB-303558607 tempest-ServerActionsTestOtherB-303558607-project-member] [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 784.014141] env[61957]: DEBUG nova.network.neutron [None req-660095f0-9d5a-4797-b621-e0585ffddbba tempest-ServerActionsTestOtherB-303558607 tempest-ServerActionsTestOtherB-303558607-project-member] [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 784.029663] env[61957]: DEBUG nova.network.neutron [None req-660095f0-9d5a-4797-b621-e0585ffddbba tempest-ServerActionsTestOtherB-303558607 tempest-ServerActionsTestOtherB-303558607-project-member] [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 784.142406] env[61957]: INFO nova.scheduler.client.report [None req-6b91bce8-d0f8-4c40-855e-ed34f3eb7e64 tempest-ServerActionsTestJSON-1567241756 tempest-ServerActionsTestJSON-1567241756-project-member] Deleted allocations for instance e69a7833-97b1-4063-a6e8-c4558980ee1e [ 784.532603] env[61957]: DEBUG nova.network.neutron [None req-660095f0-9d5a-4797-b621-e0585ffddbba tempest-ServerActionsTestOtherB-303558607 tempest-ServerActionsTestOtherB-303558607-project-member] [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 784.653564] env[61957]: DEBUG oslo_concurrency.lockutils [None req-6b91bce8-d0f8-4c40-855e-ed34f3eb7e64 tempest-ServerActionsTestJSON-1567241756 tempest-ServerActionsTestJSON-1567241756-project-member] Lock "e69a7833-97b1-4063-a6e8-c4558980ee1e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 149.707s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 784.946042] env[61957]: WARNING nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Instance 7f3850c0-a455-4e5b-a329-fb3b02f9725f is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 785.035306] env[61957]: INFO nova.compute.manager [None req-660095f0-9d5a-4797-b621-e0585ffddbba tempest-ServerActionsTestOtherB-303558607 tempest-ServerActionsTestOtherB-303558607-project-member] [instance: d77276ba-6d37-440b-a113-c4f62b2e7946] Took 1.02 seconds to deallocate network for instance. [ 785.157163] env[61957]: DEBUG nova.compute.manager [None req-bc6cc00e-6b7d-4787-8382-d5cf71a4537c tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] [instance: 615dbbed-2b02-4351-9e03-8c13f424a133] Starting instance... {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 785.448532] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Instance d77276ba-6d37-440b-a113-c4f62b2e7946 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61957) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 785.448902] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Instance 5dc993e9-5970-43ae-90be-f3a53d1f090a actively managed on this compute host and has allocations in placement: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61957) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 785.449145] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Instance 517eeb71-e9e3-4284-a144-dc4b63319d2d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61957) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 785.449834] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Instance 5cb09b60-70b6-4d04-850e-049612ec4a89 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61957) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 785.450210] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Instance 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61957) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 785.450966] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Instance 6372a384-61a8-43e8-8f6c-376e2a735045 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61957) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 785.686123] env[61957]: DEBUG oslo_concurrency.lockutils [None req-bc6cc00e-6b7d-4787-8382-d5cf71a4537c tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 785.955078] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Instance b0af6294-f30c-4266-ae46-6fb03dc0cbeb has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61957) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 786.076074] env[61957]: INFO nova.scheduler.client.report [None req-660095f0-9d5a-4797-b621-e0585ffddbba tempest-ServerActionsTestOtherB-303558607 tempest-ServerActionsTestOtherB-303558607-project-member] Deleted allocations for instance d77276ba-6d37-440b-a113-c4f62b2e7946 [ 786.458749] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Instance 6e7bd89c-2c2a-450a-9858-3526d96c28ab has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61957) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 786.591634] env[61957]: DEBUG oslo_concurrency.lockutils [None req-660095f0-9d5a-4797-b621-e0585ffddbba tempest-ServerActionsTestOtherB-303558607 tempest-ServerActionsTestOtherB-303558607-project-member] Lock "d77276ba-6d37-440b-a113-c4f62b2e7946" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 148.941s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 786.964921] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Instance e0249936-d616-4ffb-8f77-d8107633c42a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61957) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 787.095020] env[61957]: DEBUG nova.compute.manager [None req-dce171c8-b166-439d-a4ea-36b63ebfc3eb tempest-ServerGroupTestJSON-950728618 tempest-ServerGroupTestJSON-950728618-project-member] [instance: e9ac118f-08b3-430b-848c-461c2b2e3e02] Starting instance... {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 787.469461] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Instance 615dbbed-2b02-4351-9e03-8c13f424a133 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61957) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 787.619605] env[61957]: DEBUG oslo_concurrency.lockutils [None req-dce171c8-b166-439d-a4ea-36b63ebfc3eb tempest-ServerGroupTestJSON-950728618 tempest-ServerGroupTestJSON-950728618-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 787.972872] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Instance e9ac118f-08b3-430b-848c-461c2b2e3e02 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61957) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 788.476134] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Instance 14ba830d-4e2c-4e9a-a059-3c86209f0127 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61957) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 788.982689] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Instance 17f00b20-9d3b-45e6-919d-6fab9999ec77 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61957) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 789.491971] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Instance 4951132e-7247-4772-8f88-3664c6a7e61e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61957) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 789.995905] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Instance ead6aae1-36b5-4f57-9129-3bb02cf103ce has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61957) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 790.504176] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Instance 37ce46a1-8f9a-4d15-bd81-e40845a0e48a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61957) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 790.635826] env[61957]: DEBUG oslo_concurrency.lockutils [None req-902199c7-5b38-4a0e-a88f-80539210637b tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Acquiring lock "526ff179-62a6-4763-ab25-797617c4ed57" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 790.636069] env[61957]: DEBUG oslo_concurrency.lockutils [None req-902199c7-5b38-4a0e-a88f-80539210637b tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Lock "526ff179-62a6-4763-ab25-797617c4ed57" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 791.006835] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Instance 37a9e1c1-9f6c-4047-9ff5-e141d0ca383e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61957) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 791.509396] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Instance 0adae8e1-8c2f-4110-805b-1f286debc833 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61957) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 792.012240] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Instance 6438fe40-046c-45d5-9986-8f182ecde49f has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61957) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 792.515572] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Instance f66db265-887e-4d61-b848-c609e5c884cb has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61957) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 793.019173] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Instance c23141ee-0cbb-4d1b-8390-c3073fe354f1 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61957) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 793.522873] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Instance 281b9644-0e7d-48b9-a7b6-45fd6102d558 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61957) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 793.523194] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Total usable vcpus: 48, total allocated vcpus: 5 {{(pid=61957) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 793.523351] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1472MB phys_disk=200GB used_disk=4GB total_vcpus=48 used_vcpus=5 pci_stats=[] {{(pid=61957) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 793.773200] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7de35f31-a66f-493e-b5f0-5b14e3dadd07 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.780964] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6221583-54c6-44fb-9d65-717ab5884cad {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.813191] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5799deff-4b26-46ee-b7e3-093bb62ccc3c {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.821315] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-335d453f-89bc-469d-9798-68cff2d36022 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.836226] env[61957]: DEBUG nova.compute.provider_tree [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 794.341099] env[61957]: DEBUG nova.scheduler.client.report [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 794.846396] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61957) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 794.846742] env[61957]: DEBUG oslo_concurrency.lockutils [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 10.934s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 794.847428] env[61957]: DEBUG oslo_concurrency.lockutils [None req-3d8587e3-7ad5-40a6-88af-e8ed44c49baa tempest-ServerActionsV293TestJSON-1269989829 tempest-ServerActionsV293TestJSON-1269989829-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 24.942s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 795.603037] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff4dbd80-43d2-4e85-8b24-188b864465a3 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.609942] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f459346f-6c11-4e97-b343-34d5719b241f {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.638650] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3c4ccf6-fe48-4036-9c23-d6ac0282340f {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.645986] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c044bd13-5207-4905-8eb8-19815bcf4696 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.665489] env[61957]: DEBUG nova.compute.provider_tree [None req-3d8587e3-7ad5-40a6-88af-e8ed44c49baa tempest-ServerActionsV293TestJSON-1269989829 tempest-ServerActionsV293TestJSON-1269989829-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 796.169052] env[61957]: DEBUG nova.scheduler.client.report [None req-3d8587e3-7ad5-40a6-88af-e8ed44c49baa tempest-ServerActionsV293TestJSON-1269989829 tempest-ServerActionsV293TestJSON-1269989829-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 796.674513] env[61957]: DEBUG oslo_concurrency.lockutils [None req-3d8587e3-7ad5-40a6-88af-e8ed44c49baa tempest-ServerActionsV293TestJSON-1269989829 tempest-ServerActionsV293TestJSON-1269989829-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.827s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 796.675288] env[61957]: ERROR nova.compute.manager [None req-3d8587e3-7ad5-40a6-88af-e8ed44c49baa tempest-ServerActionsV293TestJSON-1269989829 tempest-ServerActionsV293TestJSON-1269989829-project-member] [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 37cebbf5-42fc-4387-989c-dd71a77adabe, please check neutron logs for more information. [ 796.675288] env[61957]: ERROR nova.compute.manager [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] Traceback (most recent call last): [ 796.675288] env[61957]: ERROR nova.compute.manager [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 796.675288] env[61957]: ERROR nova.compute.manager [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] self.driver.spawn(context, instance, image_meta, [ 796.675288] env[61957]: ERROR nova.compute.manager [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 796.675288] env[61957]: ERROR nova.compute.manager [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 796.675288] env[61957]: ERROR nova.compute.manager [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 796.675288] env[61957]: ERROR nova.compute.manager [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] vm_ref = self.build_virtual_machine(instance, [ 796.675288] env[61957]: ERROR nova.compute.manager [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 796.675288] env[61957]: ERROR nova.compute.manager [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] vif_infos = vmwarevif.get_vif_info(self._session, [ 796.675288] env[61957]: ERROR nova.compute.manager [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 796.675646] env[61957]: ERROR nova.compute.manager [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] for vif in network_info: [ 796.675646] env[61957]: ERROR nova.compute.manager [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 796.675646] env[61957]: ERROR nova.compute.manager [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] return self._sync_wrapper(fn, *args, **kwargs) [ 796.675646] env[61957]: ERROR nova.compute.manager [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 796.675646] env[61957]: ERROR nova.compute.manager [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] self.wait() [ 796.675646] env[61957]: ERROR nova.compute.manager [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 796.675646] env[61957]: ERROR nova.compute.manager [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] self[:] = self._gt.wait() [ 796.675646] env[61957]: ERROR nova.compute.manager [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 796.675646] env[61957]: ERROR nova.compute.manager [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] return self._exit_event.wait() [ 796.675646] env[61957]: ERROR nova.compute.manager [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 796.675646] env[61957]: ERROR nova.compute.manager [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] current.throw(*self._exc) [ 796.675646] env[61957]: ERROR nova.compute.manager [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 796.675646] env[61957]: ERROR nova.compute.manager [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] result = function(*args, **kwargs) [ 796.676023] env[61957]: ERROR nova.compute.manager [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 796.676023] env[61957]: ERROR nova.compute.manager [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] return func(*args, **kwargs) [ 796.676023] env[61957]: ERROR nova.compute.manager [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 796.676023] env[61957]: ERROR nova.compute.manager [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] raise e [ 796.676023] env[61957]: ERROR nova.compute.manager [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 796.676023] env[61957]: ERROR nova.compute.manager [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] nwinfo = self.network_api.allocate_for_instance( [ 796.676023] env[61957]: ERROR nova.compute.manager [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 796.676023] env[61957]: ERROR nova.compute.manager [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] created_port_ids = self._update_ports_for_instance( [ 796.676023] env[61957]: ERROR nova.compute.manager [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 796.676023] env[61957]: ERROR nova.compute.manager [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] with excutils.save_and_reraise_exception(): [ 796.676023] env[61957]: ERROR nova.compute.manager [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 796.676023] env[61957]: ERROR nova.compute.manager [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] self.force_reraise() [ 796.676023] env[61957]: ERROR nova.compute.manager [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 796.676430] env[61957]: ERROR nova.compute.manager [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] raise self.value [ 796.676430] env[61957]: ERROR nova.compute.manager [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 796.676430] env[61957]: ERROR nova.compute.manager [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] updated_port = self._update_port( [ 796.676430] env[61957]: ERROR nova.compute.manager [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 796.676430] env[61957]: ERROR nova.compute.manager [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] _ensure_no_port_binding_failure(port) [ 796.676430] env[61957]: ERROR nova.compute.manager [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 796.676430] env[61957]: ERROR nova.compute.manager [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] raise exception.PortBindingFailed(port_id=port['id']) [ 796.676430] env[61957]: ERROR nova.compute.manager [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] nova.exception.PortBindingFailed: Binding failed for port 37cebbf5-42fc-4387-989c-dd71a77adabe, please check neutron logs for more information. [ 796.676430] env[61957]: ERROR nova.compute.manager [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] [ 796.676430] env[61957]: DEBUG nova.compute.utils [None req-3d8587e3-7ad5-40a6-88af-e8ed44c49baa tempest-ServerActionsV293TestJSON-1269989829 tempest-ServerActionsV293TestJSON-1269989829-project-member] [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] Binding failed for port 37cebbf5-42fc-4387-989c-dd71a77adabe, please check neutron logs for more information. {{(pid=61957) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 796.677353] env[61957]: DEBUG oslo_concurrency.lockutils [None req-79f56a2c-6af4-4a0f-b13f-0dc131681bf7 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.629s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 796.680255] env[61957]: INFO nova.compute.claims [None req-79f56a2c-6af4-4a0f-b13f-0dc131681bf7 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 796.682948] env[61957]: DEBUG nova.compute.manager [None req-3d8587e3-7ad5-40a6-88af-e8ed44c49baa tempest-ServerActionsV293TestJSON-1269989829 tempest-ServerActionsV293TestJSON-1269989829-project-member] [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] Build of instance 5dc993e9-5970-43ae-90be-f3a53d1f090a was re-scheduled: Binding failed for port 37cebbf5-42fc-4387-989c-dd71a77adabe, please check neutron logs for more information. {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 796.683384] env[61957]: DEBUG nova.compute.manager [None req-3d8587e3-7ad5-40a6-88af-e8ed44c49baa tempest-ServerActionsV293TestJSON-1269989829 tempest-ServerActionsV293TestJSON-1269989829-project-member] [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] Unplugging VIFs for instance {{(pid=61957) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 796.684042] env[61957]: DEBUG oslo_concurrency.lockutils [None req-3d8587e3-7ad5-40a6-88af-e8ed44c49baa tempest-ServerActionsV293TestJSON-1269989829 tempest-ServerActionsV293TestJSON-1269989829-project-member] Acquiring lock "refresh_cache-5dc993e9-5970-43ae-90be-f3a53d1f090a" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 796.684042] env[61957]: DEBUG oslo_concurrency.lockutils [None req-3d8587e3-7ad5-40a6-88af-e8ed44c49baa tempest-ServerActionsV293TestJSON-1269989829 tempest-ServerActionsV293TestJSON-1269989829-project-member] Acquired lock "refresh_cache-5dc993e9-5970-43ae-90be-f3a53d1f090a" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 796.684042] env[61957]: DEBUG nova.network.neutron [None req-3d8587e3-7ad5-40a6-88af-e8ed44c49baa tempest-ServerActionsV293TestJSON-1269989829 tempest-ServerActionsV293TestJSON-1269989829-project-member] [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 797.204371] env[61957]: DEBUG nova.network.neutron [None req-3d8587e3-7ad5-40a6-88af-e8ed44c49baa tempest-ServerActionsV293TestJSON-1269989829 tempest-ServerActionsV293TestJSON-1269989829-project-member] [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 797.284728] env[61957]: DEBUG nova.network.neutron [None req-3d8587e3-7ad5-40a6-88af-e8ed44c49baa tempest-ServerActionsV293TestJSON-1269989829 tempest-ServerActionsV293TestJSON-1269989829-project-member] [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 797.659033] env[61957]: DEBUG oslo_service.periodic_task [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61957) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 797.659381] env[61957]: DEBUG oslo_service.periodic_task [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61957) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 797.787078] env[61957]: DEBUG oslo_concurrency.lockutils [None req-3d8587e3-7ad5-40a6-88af-e8ed44c49baa tempest-ServerActionsV293TestJSON-1269989829 tempest-ServerActionsV293TestJSON-1269989829-project-member] Releasing lock "refresh_cache-5dc993e9-5970-43ae-90be-f3a53d1f090a" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 797.787078] env[61957]: DEBUG nova.compute.manager [None req-3d8587e3-7ad5-40a6-88af-e8ed44c49baa tempest-ServerActionsV293TestJSON-1269989829 tempest-ServerActionsV293TestJSON-1269989829-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61957) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 797.787573] env[61957]: DEBUG nova.compute.manager [None req-3d8587e3-7ad5-40a6-88af-e8ed44c49baa tempest-ServerActionsV293TestJSON-1269989829 tempest-ServerActionsV293TestJSON-1269989829-project-member] [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 797.787573] env[61957]: DEBUG nova.network.neutron [None req-3d8587e3-7ad5-40a6-88af-e8ed44c49baa tempest-ServerActionsV293TestJSON-1269989829 tempest-ServerActionsV293TestJSON-1269989829-project-member] [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 797.802567] env[61957]: DEBUG nova.network.neutron [None req-3d8587e3-7ad5-40a6-88af-e8ed44c49baa tempest-ServerActionsV293TestJSON-1269989829 tempest-ServerActionsV293TestJSON-1269989829-project-member] [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 797.942872] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4885a04-fc1f-45ca-a3d2-cef940dabd4d {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.950785] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a04912ff-bd15-402b-a7b2-d00d8dafcd85 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.979316] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f6e86aa-6914-43e6-ac12-9fb5b687a0aa {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.986330] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8928df5-cc77-4dc2-b829-202810854e1c {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.999877] env[61957]: DEBUG nova.compute.provider_tree [None req-79f56a2c-6af4-4a0f-b13f-0dc131681bf7 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 798.165400] env[61957]: DEBUG oslo_service.periodic_task [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61957) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 798.165540] env[61957]: DEBUG oslo_service.periodic_task [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61957) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 798.166013] env[61957]: DEBUG oslo_service.periodic_task [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61957) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 798.166013] env[61957]: DEBUG oslo_service.periodic_task [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61957) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 798.166013] env[61957]: DEBUG oslo_service.periodic_task [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61957) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 798.306028] env[61957]: DEBUG nova.network.neutron [None req-3d8587e3-7ad5-40a6-88af-e8ed44c49baa tempest-ServerActionsV293TestJSON-1269989829 tempest-ServerActionsV293TestJSON-1269989829-project-member] [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 798.502666] env[61957]: DEBUG nova.scheduler.client.report [None req-79f56a2c-6af4-4a0f-b13f-0dc131681bf7 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 798.668460] env[61957]: DEBUG oslo_concurrency.lockutils [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 798.808932] env[61957]: INFO nova.compute.manager [None req-3d8587e3-7ad5-40a6-88af-e8ed44c49baa tempest-ServerActionsV293TestJSON-1269989829 tempest-ServerActionsV293TestJSON-1269989829-project-member] [instance: 5dc993e9-5970-43ae-90be-f3a53d1f090a] Took 1.02 seconds to deallocate network for instance. [ 799.007939] env[61957]: DEBUG oslo_concurrency.lockutils [None req-79f56a2c-6af4-4a0f-b13f-0dc131681bf7 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.330s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 799.008515] env[61957]: DEBUG nova.compute.manager [None req-79f56a2c-6af4-4a0f-b13f-0dc131681bf7 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] Start building networks asynchronously for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 799.011653] env[61957]: DEBUG oslo_concurrency.lockutils [None req-709951f5-5291-4ca3-b836-f8744d6c583b tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 26.920s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 799.011827] env[61957]: DEBUG oslo_concurrency.lockutils [None req-709951f5-5291-4ca3-b836-f8744d6c583b tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 799.013805] env[61957]: DEBUG oslo_concurrency.lockutils [None req-394366c7-3cf4-481b-bb9f-4da67892af40 tempest-ServersTestManualDisk-495750068 tempest-ServersTestManualDisk-495750068-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 26.173s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 799.036890] env[61957]: INFO nova.scheduler.client.report [None req-709951f5-5291-4ca3-b836-f8744d6c583b tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Deleted allocations for instance 7f3850c0-a455-4e5b-a329-fb3b02f9725f [ 799.519120] env[61957]: DEBUG nova.compute.utils [None req-79f56a2c-6af4-4a0f-b13f-0dc131681bf7 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Using /dev/sd instead of None {{(pid=61957) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 799.523864] env[61957]: DEBUG nova.compute.manager [None req-79f56a2c-6af4-4a0f-b13f-0dc131681bf7 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] Allocating IP information in the background. {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 799.524092] env[61957]: DEBUG nova.network.neutron [None req-79f56a2c-6af4-4a0f-b13f-0dc131681bf7 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] allocate_for_instance() {{(pid=61957) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 799.544690] env[61957]: DEBUG oslo_concurrency.lockutils [None req-709951f5-5291-4ca3-b836-f8744d6c583b tempest-ServerShowV257Test-250381700 tempest-ServerShowV257Test-250381700-project-member] Lock "7f3850c0-a455-4e5b-a329-fb3b02f9725f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 31.188s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 799.630916] env[61957]: DEBUG nova.policy [None req-79f56a2c-6af4-4a0f-b13f-0dc131681bf7 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8a534d0b5fc24c3babe7a24cc6b6d941', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '55742b1d4fbb401ea61f990ecaa2def6', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61957) authorize /opt/stack/nova/nova/policy.py:203}} [ 799.855319] env[61957]: INFO nova.scheduler.client.report [None req-3d8587e3-7ad5-40a6-88af-e8ed44c49baa tempest-ServerActionsV293TestJSON-1269989829 tempest-ServerActionsV293TestJSON-1269989829-project-member] Deleted allocations for instance 5dc993e9-5970-43ae-90be-f3a53d1f090a [ 799.894286] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4c49d46-56f8-42c1-b2e0-4a39f91a049e {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.906663] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf765c4d-d260-4472-9dcb-8def1f197f6e {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.939269] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cded040-25a5-471e-b7df-73654cc1923d {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.949387] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad394cd7-2e08-4de1-9619-0844d3674b14 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.964027] env[61957]: DEBUG nova.compute.provider_tree [None req-394366c7-3cf4-481b-bb9f-4da67892af40 tempest-ServersTestManualDisk-495750068 tempest-ServersTestManualDisk-495750068-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 800.024865] env[61957]: DEBUG nova.network.neutron [None req-79f56a2c-6af4-4a0f-b13f-0dc131681bf7 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] Successfully created port: 0584a64b-4127-4f93-80bc-48deadb5339c {{(pid=61957) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 800.027094] env[61957]: DEBUG nova.compute.manager [None req-79f56a2c-6af4-4a0f-b13f-0dc131681bf7 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] Start building block device mappings for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 800.365189] env[61957]: DEBUG oslo_concurrency.lockutils [None req-3d8587e3-7ad5-40a6-88af-e8ed44c49baa tempest-ServerActionsV293TestJSON-1269989829 tempest-ServerActionsV293TestJSON-1269989829-project-member] Lock "5dc993e9-5970-43ae-90be-f3a53d1f090a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 157.947s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 800.466622] env[61957]: DEBUG nova.scheduler.client.report [None req-394366c7-3cf4-481b-bb9f-4da67892af40 tempest-ServersTestManualDisk-495750068 tempest-ServersTestManualDisk-495750068-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 800.720168] env[61957]: DEBUG nova.compute.manager [req-7685b171-c8b8-421a-8e25-b09be935e28f req-866da798-cd33-4543-bab0-06417964df50 service nova] [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] Received event network-changed-0584a64b-4127-4f93-80bc-48deadb5339c {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 800.720366] env[61957]: DEBUG nova.compute.manager [req-7685b171-c8b8-421a-8e25-b09be935e28f req-866da798-cd33-4543-bab0-06417964df50 service nova] [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] Refreshing instance network info cache due to event network-changed-0584a64b-4127-4f93-80bc-48deadb5339c. {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 800.720609] env[61957]: DEBUG oslo_concurrency.lockutils [req-7685b171-c8b8-421a-8e25-b09be935e28f req-866da798-cd33-4543-bab0-06417964df50 service nova] Acquiring lock "refresh_cache-b0af6294-f30c-4266-ae46-6fb03dc0cbeb" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 800.720760] env[61957]: DEBUG oslo_concurrency.lockutils [req-7685b171-c8b8-421a-8e25-b09be935e28f req-866da798-cd33-4543-bab0-06417964df50 service nova] Acquired lock "refresh_cache-b0af6294-f30c-4266-ae46-6fb03dc0cbeb" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 800.720910] env[61957]: DEBUG nova.network.neutron [req-7685b171-c8b8-421a-8e25-b09be935e28f req-866da798-cd33-4543-bab0-06417964df50 service nova] [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] Refreshing network info cache for port 0584a64b-4127-4f93-80bc-48deadb5339c {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 800.867896] env[61957]: DEBUG nova.compute.manager [None req-36e797a2-06a5-481b-9dd9-66f4ec9dadf3 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 14ba830d-4e2c-4e9a-a059-3c86209f0127] Starting instance... {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 800.935918] env[61957]: ERROR nova.compute.manager [None req-79f56a2c-6af4-4a0f-b13f-0dc131681bf7 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 0584a64b-4127-4f93-80bc-48deadb5339c, please check neutron logs for more information. [ 800.935918] env[61957]: ERROR nova.compute.manager Traceback (most recent call last): [ 800.935918] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 800.935918] env[61957]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 800.935918] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 800.935918] env[61957]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 800.935918] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 800.935918] env[61957]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 800.935918] env[61957]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 800.935918] env[61957]: ERROR nova.compute.manager self.force_reraise() [ 800.935918] env[61957]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 800.935918] env[61957]: ERROR nova.compute.manager raise self.value [ 800.935918] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 800.935918] env[61957]: ERROR nova.compute.manager updated_port = self._update_port( [ 800.935918] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 800.935918] env[61957]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 800.936505] env[61957]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 800.936505] env[61957]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 800.936505] env[61957]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 0584a64b-4127-4f93-80bc-48deadb5339c, please check neutron logs for more information. [ 800.936505] env[61957]: ERROR nova.compute.manager [ 800.936505] env[61957]: Traceback (most recent call last): [ 800.936505] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 800.936505] env[61957]: listener.cb(fileno) [ 800.936505] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 800.936505] env[61957]: result = function(*args, **kwargs) [ 800.936505] env[61957]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 800.936505] env[61957]: return func(*args, **kwargs) [ 800.936505] env[61957]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 800.936505] env[61957]: raise e [ 800.936505] env[61957]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 800.936505] env[61957]: nwinfo = self.network_api.allocate_for_instance( [ 800.936505] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 800.936505] env[61957]: created_port_ids = self._update_ports_for_instance( [ 800.936505] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 800.936505] env[61957]: with excutils.save_and_reraise_exception(): [ 800.936505] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 800.936505] env[61957]: self.force_reraise() [ 800.936505] env[61957]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 800.936505] env[61957]: raise self.value [ 800.936505] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 800.936505] env[61957]: updated_port = self._update_port( [ 800.936505] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 800.936505] env[61957]: _ensure_no_port_binding_failure(port) [ 800.936505] env[61957]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 800.936505] env[61957]: raise exception.PortBindingFailed(port_id=port['id']) [ 800.937423] env[61957]: nova.exception.PortBindingFailed: Binding failed for port 0584a64b-4127-4f93-80bc-48deadb5339c, please check neutron logs for more information. [ 800.937423] env[61957]: Removing descriptor: 17 [ 800.973437] env[61957]: DEBUG oslo_concurrency.lockutils [None req-394366c7-3cf4-481b-bb9f-4da67892af40 tempest-ServersTestManualDisk-495750068 tempest-ServersTestManualDisk-495750068-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.959s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 800.973653] env[61957]: ERROR nova.compute.manager [None req-394366c7-3cf4-481b-bb9f-4da67892af40 tempest-ServersTestManualDisk-495750068 tempest-ServersTestManualDisk-495750068-project-member] [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port da6abf85-4313-4087-bd58-112359201b6a, please check neutron logs for more information. [ 800.973653] env[61957]: ERROR nova.compute.manager [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] Traceback (most recent call last): [ 800.973653] env[61957]: ERROR nova.compute.manager [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 800.973653] env[61957]: ERROR nova.compute.manager [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] self.driver.spawn(context, instance, image_meta, [ 800.973653] env[61957]: ERROR nova.compute.manager [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 800.973653] env[61957]: ERROR nova.compute.manager [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 800.973653] env[61957]: ERROR nova.compute.manager [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 800.973653] env[61957]: ERROR nova.compute.manager [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] vm_ref = self.build_virtual_machine(instance, [ 800.973653] env[61957]: ERROR nova.compute.manager [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 800.973653] env[61957]: ERROR nova.compute.manager [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] vif_infos = vmwarevif.get_vif_info(self._session, [ 800.973653] env[61957]: ERROR nova.compute.manager [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 800.974048] env[61957]: ERROR nova.compute.manager [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] for vif in network_info: [ 800.974048] env[61957]: ERROR nova.compute.manager [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 800.974048] env[61957]: ERROR nova.compute.manager [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] return self._sync_wrapper(fn, *args, **kwargs) [ 800.974048] env[61957]: ERROR nova.compute.manager [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 800.974048] env[61957]: ERROR nova.compute.manager [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] self.wait() [ 800.974048] env[61957]: ERROR nova.compute.manager [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 800.974048] env[61957]: ERROR nova.compute.manager [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] self[:] = self._gt.wait() [ 800.974048] env[61957]: ERROR nova.compute.manager [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 800.974048] env[61957]: ERROR nova.compute.manager [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] return self._exit_event.wait() [ 800.974048] env[61957]: ERROR nova.compute.manager [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 800.974048] env[61957]: ERROR nova.compute.manager [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] result = hub.switch() [ 800.974048] env[61957]: ERROR nova.compute.manager [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 800.974048] env[61957]: ERROR nova.compute.manager [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] return self.greenlet.switch() [ 800.974447] env[61957]: ERROR nova.compute.manager [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 800.974447] env[61957]: ERROR nova.compute.manager [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] result = function(*args, **kwargs) [ 800.974447] env[61957]: ERROR nova.compute.manager [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 800.974447] env[61957]: ERROR nova.compute.manager [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] return func(*args, **kwargs) [ 800.974447] env[61957]: ERROR nova.compute.manager [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 800.974447] env[61957]: ERROR nova.compute.manager [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] raise e [ 800.974447] env[61957]: ERROR nova.compute.manager [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 800.974447] env[61957]: ERROR nova.compute.manager [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] nwinfo = self.network_api.allocate_for_instance( [ 800.974447] env[61957]: ERROR nova.compute.manager [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 800.974447] env[61957]: ERROR nova.compute.manager [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] created_port_ids = self._update_ports_for_instance( [ 800.974447] env[61957]: ERROR nova.compute.manager [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 800.974447] env[61957]: ERROR nova.compute.manager [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] with excutils.save_and_reraise_exception(): [ 800.974447] env[61957]: ERROR nova.compute.manager [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 800.974840] env[61957]: ERROR nova.compute.manager [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] self.force_reraise() [ 800.974840] env[61957]: ERROR nova.compute.manager [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 800.974840] env[61957]: ERROR nova.compute.manager [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] raise self.value [ 800.974840] env[61957]: ERROR nova.compute.manager [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 800.974840] env[61957]: ERROR nova.compute.manager [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] updated_port = self._update_port( [ 800.974840] env[61957]: ERROR nova.compute.manager [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 800.974840] env[61957]: ERROR nova.compute.manager [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] _ensure_no_port_binding_failure(port) [ 800.974840] env[61957]: ERROR nova.compute.manager [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 800.974840] env[61957]: ERROR nova.compute.manager [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] raise exception.PortBindingFailed(port_id=port['id']) [ 800.974840] env[61957]: ERROR nova.compute.manager [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] nova.exception.PortBindingFailed: Binding failed for port da6abf85-4313-4087-bd58-112359201b6a, please check neutron logs for more information. [ 800.974840] env[61957]: ERROR nova.compute.manager [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] [ 800.975280] env[61957]: DEBUG nova.compute.utils [None req-394366c7-3cf4-481b-bb9f-4da67892af40 tempest-ServersTestManualDisk-495750068 tempest-ServersTestManualDisk-495750068-project-member] [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] Binding failed for port da6abf85-4313-4087-bd58-112359201b6a, please check neutron logs for more information. {{(pid=61957) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 800.976509] env[61957]: DEBUG oslo_concurrency.lockutils [None req-da68b092-4602-4484-b1b0-adcac4aae9f8 tempest-InstanceActionsV221TestJSON-1768084888 tempest-InstanceActionsV221TestJSON-1768084888-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.599s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 800.978366] env[61957]: INFO nova.compute.claims [None req-da68b092-4602-4484-b1b0-adcac4aae9f8 tempest-InstanceActionsV221TestJSON-1768084888 tempest-InstanceActionsV221TestJSON-1768084888-project-member] [instance: 6e7bd89c-2c2a-450a-9858-3526d96c28ab] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 800.981562] env[61957]: DEBUG nova.compute.manager [None req-394366c7-3cf4-481b-bb9f-4da67892af40 tempest-ServersTestManualDisk-495750068 tempest-ServersTestManualDisk-495750068-project-member] [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] Build of instance 517eeb71-e9e3-4284-a144-dc4b63319d2d was re-scheduled: Binding failed for port da6abf85-4313-4087-bd58-112359201b6a, please check neutron logs for more information. {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 800.981562] env[61957]: DEBUG nova.compute.manager [None req-394366c7-3cf4-481b-bb9f-4da67892af40 tempest-ServersTestManualDisk-495750068 tempest-ServersTestManualDisk-495750068-project-member] [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] Unplugging VIFs for instance {{(pid=61957) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 800.981784] env[61957]: DEBUG oslo_concurrency.lockutils [None req-394366c7-3cf4-481b-bb9f-4da67892af40 tempest-ServersTestManualDisk-495750068 tempest-ServersTestManualDisk-495750068-project-member] Acquiring lock "refresh_cache-517eeb71-e9e3-4284-a144-dc4b63319d2d" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 800.981950] env[61957]: DEBUG oslo_concurrency.lockutils [None req-394366c7-3cf4-481b-bb9f-4da67892af40 tempest-ServersTestManualDisk-495750068 tempest-ServersTestManualDisk-495750068-project-member] Acquired lock "refresh_cache-517eeb71-e9e3-4284-a144-dc4b63319d2d" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 800.982956] env[61957]: DEBUG nova.network.neutron [None req-394366c7-3cf4-481b-bb9f-4da67892af40 tempest-ServersTestManualDisk-495750068 tempest-ServersTestManualDisk-495750068-project-member] [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 801.038400] env[61957]: DEBUG nova.compute.manager [None req-79f56a2c-6af4-4a0f-b13f-0dc131681bf7 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] Start spawning the instance on the hypervisor. {{(pid=61957) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 801.066267] env[61957]: DEBUG nova.virt.hardware [None req-79f56a2c-6af4-4a0f-b13f-0dc131681bf7 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T17:22:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T17:21:50Z,direct_url=,disk_format='vmdk',id=11c76a2c-f705-470a-ba9d-4657858bab38,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='81d3096d0f094373913fc3dc8f5c3c6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T17:21:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 801.066535] env[61957]: DEBUG nova.virt.hardware [None req-79f56a2c-6af4-4a0f-b13f-0dc131681bf7 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Flavor limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 801.066692] env[61957]: DEBUG nova.virt.hardware [None req-79f56a2c-6af4-4a0f-b13f-0dc131681bf7 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Image limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 801.066871] env[61957]: DEBUG nova.virt.hardware [None req-79f56a2c-6af4-4a0f-b13f-0dc131681bf7 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Flavor pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 801.067030] env[61957]: DEBUG nova.virt.hardware [None req-79f56a2c-6af4-4a0f-b13f-0dc131681bf7 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Image pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 801.067181] env[61957]: DEBUG nova.virt.hardware [None req-79f56a2c-6af4-4a0f-b13f-0dc131681bf7 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 801.067410] env[61957]: DEBUG nova.virt.hardware [None req-79f56a2c-6af4-4a0f-b13f-0dc131681bf7 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 801.067579] env[61957]: DEBUG nova.virt.hardware [None req-79f56a2c-6af4-4a0f-b13f-0dc131681bf7 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 801.067744] env[61957]: DEBUG nova.virt.hardware [None req-79f56a2c-6af4-4a0f-b13f-0dc131681bf7 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Got 1 possible topologies {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 801.067904] env[61957]: DEBUG nova.virt.hardware [None req-79f56a2c-6af4-4a0f-b13f-0dc131681bf7 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 801.068413] env[61957]: DEBUG nova.virt.hardware [None req-79f56a2c-6af4-4a0f-b13f-0dc131681bf7 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 801.071582] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-168aa9ab-c3f0-4187-96cb-fdfad4553229 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.080632] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa178c92-7ea2-427b-a341-848f706b256f {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.095867] env[61957]: ERROR nova.compute.manager [None req-79f56a2c-6af4-4a0f-b13f-0dc131681bf7 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 0584a64b-4127-4f93-80bc-48deadb5339c, please check neutron logs for more information. [ 801.095867] env[61957]: ERROR nova.compute.manager [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] Traceback (most recent call last): [ 801.095867] env[61957]: ERROR nova.compute.manager [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 801.095867] env[61957]: ERROR nova.compute.manager [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] yield resources [ 801.095867] env[61957]: ERROR nova.compute.manager [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 801.095867] env[61957]: ERROR nova.compute.manager [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] self.driver.spawn(context, instance, image_meta, [ 801.095867] env[61957]: ERROR nova.compute.manager [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 801.095867] env[61957]: ERROR nova.compute.manager [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] self._vmops.spawn(context, instance, image_meta, injected_files, [ 801.095867] env[61957]: ERROR nova.compute.manager [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 801.095867] env[61957]: ERROR nova.compute.manager [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] vm_ref = self.build_virtual_machine(instance, [ 801.095867] env[61957]: ERROR nova.compute.manager [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 801.096306] env[61957]: ERROR nova.compute.manager [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] vif_infos = vmwarevif.get_vif_info(self._session, [ 801.096306] env[61957]: ERROR nova.compute.manager [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 801.096306] env[61957]: ERROR nova.compute.manager [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] for vif in network_info: [ 801.096306] env[61957]: ERROR nova.compute.manager [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 801.096306] env[61957]: ERROR nova.compute.manager [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] return self._sync_wrapper(fn, *args, **kwargs) [ 801.096306] env[61957]: ERROR nova.compute.manager [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 801.096306] env[61957]: ERROR nova.compute.manager [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] self.wait() [ 801.096306] env[61957]: ERROR nova.compute.manager [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 801.096306] env[61957]: ERROR nova.compute.manager [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] self[:] = self._gt.wait() [ 801.096306] env[61957]: ERROR nova.compute.manager [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 801.096306] env[61957]: ERROR nova.compute.manager [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] return self._exit_event.wait() [ 801.096306] env[61957]: ERROR nova.compute.manager [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 801.096306] env[61957]: ERROR nova.compute.manager [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] current.throw(*self._exc) [ 801.096781] env[61957]: ERROR nova.compute.manager [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 801.096781] env[61957]: ERROR nova.compute.manager [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] result = function(*args, **kwargs) [ 801.096781] env[61957]: ERROR nova.compute.manager [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 801.096781] env[61957]: ERROR nova.compute.manager [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] return func(*args, **kwargs) [ 801.096781] env[61957]: ERROR nova.compute.manager [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 801.096781] env[61957]: ERROR nova.compute.manager [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] raise e [ 801.096781] env[61957]: ERROR nova.compute.manager [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 801.096781] env[61957]: ERROR nova.compute.manager [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] nwinfo = self.network_api.allocate_for_instance( [ 801.096781] env[61957]: ERROR nova.compute.manager [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 801.096781] env[61957]: ERROR nova.compute.manager [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] created_port_ids = self._update_ports_for_instance( [ 801.096781] env[61957]: ERROR nova.compute.manager [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 801.096781] env[61957]: ERROR nova.compute.manager [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] with excutils.save_and_reraise_exception(): [ 801.096781] env[61957]: ERROR nova.compute.manager [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 801.097238] env[61957]: ERROR nova.compute.manager [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] self.force_reraise() [ 801.097238] env[61957]: ERROR nova.compute.manager [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 801.097238] env[61957]: ERROR nova.compute.manager [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] raise self.value [ 801.097238] env[61957]: ERROR nova.compute.manager [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 801.097238] env[61957]: ERROR nova.compute.manager [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] updated_port = self._update_port( [ 801.097238] env[61957]: ERROR nova.compute.manager [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 801.097238] env[61957]: ERROR nova.compute.manager [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] _ensure_no_port_binding_failure(port) [ 801.097238] env[61957]: ERROR nova.compute.manager [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 801.097238] env[61957]: ERROR nova.compute.manager [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] raise exception.PortBindingFailed(port_id=port['id']) [ 801.097238] env[61957]: ERROR nova.compute.manager [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] nova.exception.PortBindingFailed: Binding failed for port 0584a64b-4127-4f93-80bc-48deadb5339c, please check neutron logs for more information. [ 801.097238] env[61957]: ERROR nova.compute.manager [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] [ 801.097238] env[61957]: INFO nova.compute.manager [None req-79f56a2c-6af4-4a0f-b13f-0dc131681bf7 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] Terminating instance [ 801.101052] env[61957]: DEBUG oslo_concurrency.lockutils [None req-79f56a2c-6af4-4a0f-b13f-0dc131681bf7 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Acquiring lock "refresh_cache-b0af6294-f30c-4266-ae46-6fb03dc0cbeb" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 801.239661] env[61957]: DEBUG nova.network.neutron [req-7685b171-c8b8-421a-8e25-b09be935e28f req-866da798-cd33-4543-bab0-06417964df50 service nova] [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 801.333429] env[61957]: DEBUG nova.network.neutron [req-7685b171-c8b8-421a-8e25-b09be935e28f req-866da798-cd33-4543-bab0-06417964df50 service nova] [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 801.399067] env[61957]: DEBUG oslo_concurrency.lockutils [None req-36e797a2-06a5-481b-9dd9-66f4ec9dadf3 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 801.508107] env[61957]: DEBUG nova.network.neutron [None req-394366c7-3cf4-481b-bb9f-4da67892af40 tempest-ServersTestManualDisk-495750068 tempest-ServersTestManualDisk-495750068-project-member] [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 801.592303] env[61957]: DEBUG nova.network.neutron [None req-394366c7-3cf4-481b-bb9f-4da67892af40 tempest-ServersTestManualDisk-495750068 tempest-ServersTestManualDisk-495750068-project-member] [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 801.836351] env[61957]: DEBUG oslo_concurrency.lockutils [req-7685b171-c8b8-421a-8e25-b09be935e28f req-866da798-cd33-4543-bab0-06417964df50 service nova] Releasing lock "refresh_cache-b0af6294-f30c-4266-ae46-6fb03dc0cbeb" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 801.836747] env[61957]: DEBUG oslo_concurrency.lockutils [None req-79f56a2c-6af4-4a0f-b13f-0dc131681bf7 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Acquired lock "refresh_cache-b0af6294-f30c-4266-ae46-6fb03dc0cbeb" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 801.837409] env[61957]: DEBUG nova.network.neutron [None req-79f56a2c-6af4-4a0f-b13f-0dc131681bf7 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 802.096415] env[61957]: DEBUG oslo_concurrency.lockutils [None req-394366c7-3cf4-481b-bb9f-4da67892af40 tempest-ServersTestManualDisk-495750068 tempest-ServersTestManualDisk-495750068-project-member] Releasing lock "refresh_cache-517eeb71-e9e3-4284-a144-dc4b63319d2d" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 802.096704] env[61957]: DEBUG nova.compute.manager [None req-394366c7-3cf4-481b-bb9f-4da67892af40 tempest-ServersTestManualDisk-495750068 tempest-ServersTestManualDisk-495750068-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61957) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 802.097127] env[61957]: DEBUG nova.compute.manager [None req-394366c7-3cf4-481b-bb9f-4da67892af40 tempest-ServersTestManualDisk-495750068 tempest-ServersTestManualDisk-495750068-project-member] [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 802.097127] env[61957]: DEBUG nova.network.neutron [None req-394366c7-3cf4-481b-bb9f-4da67892af40 tempest-ServersTestManualDisk-495750068 tempest-ServersTestManualDisk-495750068-project-member] [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 802.115646] env[61957]: DEBUG nova.network.neutron [None req-394366c7-3cf4-481b-bb9f-4da67892af40 tempest-ServersTestManualDisk-495750068 tempest-ServersTestManualDisk-495750068-project-member] [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 802.269224] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5a9733a-8a29-44da-9415-2a7d72d2bc9f {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.277831] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52b26e6a-7d13-480d-8496-0018a34e67c5 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.309699] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac48f071-9cd6-4dc5-a414-6a761f873e24 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.317116] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30cd644f-2c13-493d-8dbf-85e8b7811693 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.333444] env[61957]: DEBUG nova.compute.provider_tree [None req-da68b092-4602-4484-b1b0-adcac4aae9f8 tempest-InstanceActionsV221TestJSON-1768084888 tempest-InstanceActionsV221TestJSON-1768084888-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 802.360820] env[61957]: DEBUG nova.network.neutron [None req-79f56a2c-6af4-4a0f-b13f-0dc131681bf7 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 802.504601] env[61957]: DEBUG nova.network.neutron [None req-79f56a2c-6af4-4a0f-b13f-0dc131681bf7 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 802.620841] env[61957]: DEBUG nova.network.neutron [None req-394366c7-3cf4-481b-bb9f-4da67892af40 tempest-ServersTestManualDisk-495750068 tempest-ServersTestManualDisk-495750068-project-member] [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 802.746930] env[61957]: DEBUG nova.compute.manager [req-cbf562ff-83d7-46d7-a373-f01280d0fba8 req-2b1823a6-070b-472f-87a2-1c7ab21cde36 service nova] [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] Received event network-vif-deleted-0584a64b-4127-4f93-80bc-48deadb5339c {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 802.837107] env[61957]: DEBUG nova.scheduler.client.report [None req-da68b092-4602-4484-b1b0-adcac4aae9f8 tempest-InstanceActionsV221TestJSON-1768084888 tempest-InstanceActionsV221TestJSON-1768084888-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 803.006679] env[61957]: DEBUG oslo_concurrency.lockutils [None req-79f56a2c-6af4-4a0f-b13f-0dc131681bf7 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Releasing lock "refresh_cache-b0af6294-f30c-4266-ae46-6fb03dc0cbeb" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 803.007123] env[61957]: DEBUG nova.compute.manager [None req-79f56a2c-6af4-4a0f-b13f-0dc131681bf7 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] Start destroying the instance on the hypervisor. {{(pid=61957) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 803.007339] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-79f56a2c-6af4-4a0f-b13f-0dc131681bf7 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] Destroying instance {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 803.007642] env[61957]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b57e80e1-6656-4d18-96c3-c5d4cd9a868c {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.018400] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75108cb3-8b4e-4dae-a796-bf7d3d4e3bae {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.041526] env[61957]: WARNING nova.virt.vmwareapi.vmops [None req-79f56a2c-6af4-4a0f-b13f-0dc131681bf7 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance b0af6294-f30c-4266-ae46-6fb03dc0cbeb could not be found. [ 803.041526] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-79f56a2c-6af4-4a0f-b13f-0dc131681bf7 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] Instance destroyed {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 803.041526] env[61957]: INFO nova.compute.manager [None req-79f56a2c-6af4-4a0f-b13f-0dc131681bf7 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] Took 0.03 seconds to destroy the instance on the hypervisor. [ 803.041526] env[61957]: DEBUG oslo.service.loopingcall [None req-79f56a2c-6af4-4a0f-b13f-0dc131681bf7 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 803.041526] env[61957]: DEBUG nova.compute.manager [-] [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 803.041526] env[61957]: DEBUG nova.network.neutron [-] [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 803.064178] env[61957]: DEBUG nova.network.neutron [-] [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 803.124024] env[61957]: INFO nova.compute.manager [None req-394366c7-3cf4-481b-bb9f-4da67892af40 tempest-ServersTestManualDisk-495750068 tempest-ServersTestManualDisk-495750068-project-member] [instance: 517eeb71-e9e3-4284-a144-dc4b63319d2d] Took 1.03 seconds to deallocate network for instance. [ 803.344186] env[61957]: DEBUG oslo_concurrency.lockutils [None req-da68b092-4602-4484-b1b0-adcac4aae9f8 tempest-InstanceActionsV221TestJSON-1768084888 tempest-InstanceActionsV221TestJSON-1768084888-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.368s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 803.344716] env[61957]: DEBUG nova.compute.manager [None req-da68b092-4602-4484-b1b0-adcac4aae9f8 tempest-InstanceActionsV221TestJSON-1768084888 tempest-InstanceActionsV221TestJSON-1768084888-project-member] [instance: 6e7bd89c-2c2a-450a-9858-3526d96c28ab] Start building networks asynchronously for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 803.348969] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d6c9937c-2526-4768-bade-4ca2ced8b569 tempest-ServersTestFqdnHostnames-1308211839 tempest-ServersTestFqdnHostnames-1308211839-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 26.066s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 803.565813] env[61957]: DEBUG nova.network.neutron [-] [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 803.854047] env[61957]: DEBUG nova.compute.utils [None req-da68b092-4602-4484-b1b0-adcac4aae9f8 tempest-InstanceActionsV221TestJSON-1768084888 tempest-InstanceActionsV221TestJSON-1768084888-project-member] Using /dev/sd instead of None {{(pid=61957) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 803.859885] env[61957]: DEBUG nova.compute.manager [None req-da68b092-4602-4484-b1b0-adcac4aae9f8 tempest-InstanceActionsV221TestJSON-1768084888 tempest-InstanceActionsV221TestJSON-1768084888-project-member] [instance: 6e7bd89c-2c2a-450a-9858-3526d96c28ab] Allocating IP information in the background. {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 803.860109] env[61957]: DEBUG nova.network.neutron [None req-da68b092-4602-4484-b1b0-adcac4aae9f8 tempest-InstanceActionsV221TestJSON-1768084888 tempest-InstanceActionsV221TestJSON-1768084888-project-member] [instance: 6e7bd89c-2c2a-450a-9858-3526d96c28ab] allocate_for_instance() {{(pid=61957) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 803.906801] env[61957]: DEBUG nova.policy [None req-da68b092-4602-4484-b1b0-adcac4aae9f8 tempest-InstanceActionsV221TestJSON-1768084888 tempest-InstanceActionsV221TestJSON-1768084888-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '126e5dd7d3c64b9593263f87bd0fa7a1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '645818a1252d49fcbea063f2215923e5', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61957) authorize /opt/stack/nova/nova/policy.py:203}} [ 804.071980] env[61957]: INFO nova.compute.manager [-] [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] Took 1.03 seconds to deallocate network for instance. [ 804.074222] env[61957]: DEBUG nova.compute.claims [None req-79f56a2c-6af4-4a0f-b13f-0dc131681bf7 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] Aborting claim: {{(pid=61957) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 804.074502] env[61957]: DEBUG oslo_concurrency.lockutils [None req-79f56a2c-6af4-4a0f-b13f-0dc131681bf7 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 804.111158] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-406a0bbf-2f45-43e8-8537-1aab53c900f9 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.119048] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45deb0c4-c1a8-4fca-818f-e4d2405e2e96 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.154300] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c22e85d-a936-4ea2-b807-2a453f9926cd {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.164397] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9497d21f-4a7e-4953-b15a-3d5a690f285a {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.168884] env[61957]: DEBUG nova.network.neutron [None req-da68b092-4602-4484-b1b0-adcac4aae9f8 tempest-InstanceActionsV221TestJSON-1768084888 tempest-InstanceActionsV221TestJSON-1768084888-project-member] [instance: 6e7bd89c-2c2a-450a-9858-3526d96c28ab] Successfully created port: 54e92ab2-af65-4141-994f-2ff5ddb01f84 {{(pid=61957) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 804.180087] env[61957]: DEBUG nova.compute.provider_tree [None req-d6c9937c-2526-4768-bade-4ca2ced8b569 tempest-ServersTestFqdnHostnames-1308211839 tempest-ServersTestFqdnHostnames-1308211839-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 804.182093] env[61957]: INFO nova.scheduler.client.report [None req-394366c7-3cf4-481b-bb9f-4da67892af40 tempest-ServersTestManualDisk-495750068 tempest-ServersTestManualDisk-495750068-project-member] Deleted allocations for instance 517eeb71-e9e3-4284-a144-dc4b63319d2d [ 804.362313] env[61957]: DEBUG nova.compute.manager [None req-da68b092-4602-4484-b1b0-adcac4aae9f8 tempest-InstanceActionsV221TestJSON-1768084888 tempest-InstanceActionsV221TestJSON-1768084888-project-member] [instance: 6e7bd89c-2c2a-450a-9858-3526d96c28ab] Start building block device mappings for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 804.691634] env[61957]: DEBUG nova.scheduler.client.report [None req-d6c9937c-2526-4768-bade-4ca2ced8b569 tempest-ServersTestFqdnHostnames-1308211839 tempest-ServersTestFqdnHostnames-1308211839-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 804.694405] env[61957]: DEBUG oslo_concurrency.lockutils [None req-394366c7-3cf4-481b-bb9f-4da67892af40 tempest-ServersTestManualDisk-495750068 tempest-ServersTestManualDisk-495750068-project-member] Lock "517eeb71-e9e3-4284-a144-dc4b63319d2d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 161.163s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 805.195564] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d6c9937c-2526-4768-bade-4ca2ced8b569 tempest-ServersTestFqdnHostnames-1308211839 tempest-ServersTestFqdnHostnames-1308211839-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.848s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 805.196395] env[61957]: ERROR nova.compute.manager [None req-d6c9937c-2526-4768-bade-4ca2ced8b569 tempest-ServersTestFqdnHostnames-1308211839 tempest-ServersTestFqdnHostnames-1308211839-project-member] [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port ad48de80-9219-4f81-8c65-da8fb5df2729, please check neutron logs for more information. [ 805.196395] env[61957]: ERROR nova.compute.manager [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] Traceback (most recent call last): [ 805.196395] env[61957]: ERROR nova.compute.manager [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 805.196395] env[61957]: ERROR nova.compute.manager [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] self.driver.spawn(context, instance, image_meta, [ 805.196395] env[61957]: ERROR nova.compute.manager [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 805.196395] env[61957]: ERROR nova.compute.manager [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] self._vmops.spawn(context, instance, image_meta, injected_files, [ 805.196395] env[61957]: ERROR nova.compute.manager [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 805.196395] env[61957]: ERROR nova.compute.manager [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] vm_ref = self.build_virtual_machine(instance, [ 805.196395] env[61957]: ERROR nova.compute.manager [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 805.196395] env[61957]: ERROR nova.compute.manager [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] vif_infos = vmwarevif.get_vif_info(self._session, [ 805.196395] env[61957]: ERROR nova.compute.manager [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 805.196918] env[61957]: ERROR nova.compute.manager [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] for vif in network_info: [ 805.196918] env[61957]: ERROR nova.compute.manager [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 805.196918] env[61957]: ERROR nova.compute.manager [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] return self._sync_wrapper(fn, *args, **kwargs) [ 805.196918] env[61957]: ERROR nova.compute.manager [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 805.196918] env[61957]: ERROR nova.compute.manager [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] self.wait() [ 805.196918] env[61957]: ERROR nova.compute.manager [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 805.196918] env[61957]: ERROR nova.compute.manager [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] self[:] = self._gt.wait() [ 805.196918] env[61957]: ERROR nova.compute.manager [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 805.196918] env[61957]: ERROR nova.compute.manager [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] return self._exit_event.wait() [ 805.196918] env[61957]: ERROR nova.compute.manager [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 805.196918] env[61957]: ERROR nova.compute.manager [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] current.throw(*self._exc) [ 805.196918] env[61957]: ERROR nova.compute.manager [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 805.196918] env[61957]: ERROR nova.compute.manager [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] result = function(*args, **kwargs) [ 805.197390] env[61957]: ERROR nova.compute.manager [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 805.197390] env[61957]: ERROR nova.compute.manager [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] return func(*args, **kwargs) [ 805.197390] env[61957]: ERROR nova.compute.manager [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 805.197390] env[61957]: ERROR nova.compute.manager [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] raise e [ 805.197390] env[61957]: ERROR nova.compute.manager [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 805.197390] env[61957]: ERROR nova.compute.manager [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] nwinfo = self.network_api.allocate_for_instance( [ 805.197390] env[61957]: ERROR nova.compute.manager [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 805.197390] env[61957]: ERROR nova.compute.manager [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] created_port_ids = self._update_ports_for_instance( [ 805.197390] env[61957]: ERROR nova.compute.manager [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 805.197390] env[61957]: ERROR nova.compute.manager [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] with excutils.save_and_reraise_exception(): [ 805.197390] env[61957]: ERROR nova.compute.manager [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 805.197390] env[61957]: ERROR nova.compute.manager [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] self.force_reraise() [ 805.197390] env[61957]: ERROR nova.compute.manager [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 805.197835] env[61957]: ERROR nova.compute.manager [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] raise self.value [ 805.197835] env[61957]: ERROR nova.compute.manager [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 805.197835] env[61957]: ERROR nova.compute.manager [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] updated_port = self._update_port( [ 805.197835] env[61957]: ERROR nova.compute.manager [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 805.197835] env[61957]: ERROR nova.compute.manager [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] _ensure_no_port_binding_failure(port) [ 805.197835] env[61957]: ERROR nova.compute.manager [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 805.197835] env[61957]: ERROR nova.compute.manager [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] raise exception.PortBindingFailed(port_id=port['id']) [ 805.197835] env[61957]: ERROR nova.compute.manager [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] nova.exception.PortBindingFailed: Binding failed for port ad48de80-9219-4f81-8c65-da8fb5df2729, please check neutron logs for more information. [ 805.197835] env[61957]: ERROR nova.compute.manager [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] [ 805.198315] env[61957]: DEBUG nova.compute.utils [None req-d6c9937c-2526-4768-bade-4ca2ced8b569 tempest-ServersTestFqdnHostnames-1308211839 tempest-ServersTestFqdnHostnames-1308211839-project-member] [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] Binding failed for port ad48de80-9219-4f81-8c65-da8fb5df2729, please check neutron logs for more information. {{(pid=61957) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 805.200506] env[61957]: DEBUG oslo_concurrency.lockutils [None req-bb8cb37b-74c5-4607-9ba8-e5615134a1da tempest-InstanceActionsTestJSON-565478614 tempest-InstanceActionsTestJSON-565478614-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 26.417s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 805.205104] env[61957]: DEBUG nova.compute.manager [None req-9387ee6a-989e-4086-be03-e20e994ba789 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] [instance: 17f00b20-9d3b-45e6-919d-6fab9999ec77] Starting instance... {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 805.205842] env[61957]: DEBUG nova.compute.manager [None req-d6c9937c-2526-4768-bade-4ca2ced8b569 tempest-ServersTestFqdnHostnames-1308211839 tempest-ServersTestFqdnHostnames-1308211839-project-member] [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] Build of instance 5cb09b60-70b6-4d04-850e-049612ec4a89 was re-scheduled: Binding failed for port ad48de80-9219-4f81-8c65-da8fb5df2729, please check neutron logs for more information. {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 805.206421] env[61957]: DEBUG nova.compute.manager [None req-d6c9937c-2526-4768-bade-4ca2ced8b569 tempest-ServersTestFqdnHostnames-1308211839 tempest-ServersTestFqdnHostnames-1308211839-project-member] [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] Unplugging VIFs for instance {{(pid=61957) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 805.206701] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d6c9937c-2526-4768-bade-4ca2ced8b569 tempest-ServersTestFqdnHostnames-1308211839 tempest-ServersTestFqdnHostnames-1308211839-project-member] Acquiring lock "refresh_cache-5cb09b60-70b6-4d04-850e-049612ec4a89" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 805.206908] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d6c9937c-2526-4768-bade-4ca2ced8b569 tempest-ServersTestFqdnHostnames-1308211839 tempest-ServersTestFqdnHostnames-1308211839-project-member] Acquired lock "refresh_cache-5cb09b60-70b6-4d04-850e-049612ec4a89" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 805.207180] env[61957]: DEBUG nova.network.neutron [None req-d6c9937c-2526-4768-bade-4ca2ced8b569 tempest-ServersTestFqdnHostnames-1308211839 tempest-ServersTestFqdnHostnames-1308211839-project-member] [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 805.381676] env[61957]: DEBUG nova.compute.manager [None req-da68b092-4602-4484-b1b0-adcac4aae9f8 tempest-InstanceActionsV221TestJSON-1768084888 tempest-InstanceActionsV221TestJSON-1768084888-project-member] [instance: 6e7bd89c-2c2a-450a-9858-3526d96c28ab] Start spawning the instance on the hypervisor. {{(pid=61957) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 805.409038] env[61957]: DEBUG nova.virt.hardware [None req-da68b092-4602-4484-b1b0-adcac4aae9f8 tempest-InstanceActionsV221TestJSON-1768084888 tempest-InstanceActionsV221TestJSON-1768084888-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T17:22:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T17:21:50Z,direct_url=,disk_format='vmdk',id=11c76a2c-f705-470a-ba9d-4657858bab38,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='81d3096d0f094373913fc3dc8f5c3c6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T17:21:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 805.409292] env[61957]: DEBUG nova.virt.hardware [None req-da68b092-4602-4484-b1b0-adcac4aae9f8 tempest-InstanceActionsV221TestJSON-1768084888 tempest-InstanceActionsV221TestJSON-1768084888-project-member] Flavor limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 805.409476] env[61957]: DEBUG nova.virt.hardware [None req-da68b092-4602-4484-b1b0-adcac4aae9f8 tempest-InstanceActionsV221TestJSON-1768084888 tempest-InstanceActionsV221TestJSON-1768084888-project-member] Image limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 805.409663] env[61957]: DEBUG nova.virt.hardware [None req-da68b092-4602-4484-b1b0-adcac4aae9f8 tempest-InstanceActionsV221TestJSON-1768084888 tempest-InstanceActionsV221TestJSON-1768084888-project-member] Flavor pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 805.409800] env[61957]: DEBUG nova.virt.hardware [None req-da68b092-4602-4484-b1b0-adcac4aae9f8 tempest-InstanceActionsV221TestJSON-1768084888 tempest-InstanceActionsV221TestJSON-1768084888-project-member] Image pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 805.409939] env[61957]: DEBUG nova.virt.hardware [None req-da68b092-4602-4484-b1b0-adcac4aae9f8 tempest-InstanceActionsV221TestJSON-1768084888 tempest-InstanceActionsV221TestJSON-1768084888-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 805.410152] env[61957]: DEBUG nova.virt.hardware [None req-da68b092-4602-4484-b1b0-adcac4aae9f8 tempest-InstanceActionsV221TestJSON-1768084888 tempest-InstanceActionsV221TestJSON-1768084888-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 805.410301] env[61957]: DEBUG nova.virt.hardware [None req-da68b092-4602-4484-b1b0-adcac4aae9f8 tempest-InstanceActionsV221TestJSON-1768084888 tempest-InstanceActionsV221TestJSON-1768084888-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 805.410958] env[61957]: DEBUG nova.virt.hardware [None req-da68b092-4602-4484-b1b0-adcac4aae9f8 tempest-InstanceActionsV221TestJSON-1768084888 tempest-InstanceActionsV221TestJSON-1768084888-project-member] Got 1 possible topologies {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 805.410958] env[61957]: DEBUG nova.virt.hardware [None req-da68b092-4602-4484-b1b0-adcac4aae9f8 tempest-InstanceActionsV221TestJSON-1768084888 tempest-InstanceActionsV221TestJSON-1768084888-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 805.410958] env[61957]: DEBUG nova.virt.hardware [None req-da68b092-4602-4484-b1b0-adcac4aae9f8 tempest-InstanceActionsV221TestJSON-1768084888 tempest-InstanceActionsV221TestJSON-1768084888-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 805.411690] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-623449f2-64c7-4dc4-879e-cb8036cd119b {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.423740] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c51aa599-3a5e-442f-9735-7c7281437034 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.745784] env[61957]: DEBUG nova.network.neutron [None req-d6c9937c-2526-4768-bade-4ca2ced8b569 tempest-ServersTestFqdnHostnames-1308211839 tempest-ServersTestFqdnHostnames-1308211839-project-member] [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 805.748650] env[61957]: DEBUG oslo_concurrency.lockutils [None req-9387ee6a-989e-4086-be03-e20e994ba789 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 805.941735] env[61957]: DEBUG nova.network.neutron [None req-d6c9937c-2526-4768-bade-4ca2ced8b569 tempest-ServersTestFqdnHostnames-1308211839 tempest-ServersTestFqdnHostnames-1308211839-project-member] [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 806.028129] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-159b7919-c2df-475c-991e-b30edaf1e964 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.035833] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e857884-a266-471c-ab99-5a5622e55ca7 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.068341] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6725d631-462b-4d6b-b3a2-76189ba4e45c {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.072187] env[61957]: DEBUG nova.compute.manager [req-99992813-2719-42dc-b268-af7a3e14607e req-0615b3c3-9c0b-48e9-a848-7d076054e45f service nova] [instance: 6e7bd89c-2c2a-450a-9858-3526d96c28ab] Received event network-vif-plugged-54e92ab2-af65-4141-994f-2ff5ddb01f84 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 806.072478] env[61957]: DEBUG oslo_concurrency.lockutils [req-99992813-2719-42dc-b268-af7a3e14607e req-0615b3c3-9c0b-48e9-a848-7d076054e45f service nova] Acquiring lock "6e7bd89c-2c2a-450a-9858-3526d96c28ab-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 806.072600] env[61957]: DEBUG oslo_concurrency.lockutils [req-99992813-2719-42dc-b268-af7a3e14607e req-0615b3c3-9c0b-48e9-a848-7d076054e45f service nova] Lock "6e7bd89c-2c2a-450a-9858-3526d96c28ab-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 806.072767] env[61957]: DEBUG oslo_concurrency.lockutils [req-99992813-2719-42dc-b268-af7a3e14607e req-0615b3c3-9c0b-48e9-a848-7d076054e45f service nova] Lock "6e7bd89c-2c2a-450a-9858-3526d96c28ab-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 806.073016] env[61957]: DEBUG nova.compute.manager [req-99992813-2719-42dc-b268-af7a3e14607e req-0615b3c3-9c0b-48e9-a848-7d076054e45f service nova] [instance: 6e7bd89c-2c2a-450a-9858-3526d96c28ab] No waiting events found dispatching network-vif-plugged-54e92ab2-af65-4141-994f-2ff5ddb01f84 {{(pid=61957) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 806.073373] env[61957]: WARNING nova.compute.manager [req-99992813-2719-42dc-b268-af7a3e14607e req-0615b3c3-9c0b-48e9-a848-7d076054e45f service nova] [instance: 6e7bd89c-2c2a-450a-9858-3526d96c28ab] Received unexpected event network-vif-plugged-54e92ab2-af65-4141-994f-2ff5ddb01f84 for instance with vm_state building and task_state spawning. [ 806.079424] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3ef874c-e709-4254-a598-d3f8bbb39377 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.096184] env[61957]: DEBUG nova.compute.provider_tree [None req-bb8cb37b-74c5-4607-9ba8-e5615134a1da tempest-InstanceActionsTestJSON-565478614 tempest-InstanceActionsTestJSON-565478614-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 806.171984] env[61957]: DEBUG nova.network.neutron [None req-da68b092-4602-4484-b1b0-adcac4aae9f8 tempest-InstanceActionsV221TestJSON-1768084888 tempest-InstanceActionsV221TestJSON-1768084888-project-member] [instance: 6e7bd89c-2c2a-450a-9858-3526d96c28ab] Successfully updated port: 54e92ab2-af65-4141-994f-2ff5ddb01f84 {{(pid=61957) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 806.452874] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d6c9937c-2526-4768-bade-4ca2ced8b569 tempest-ServersTestFqdnHostnames-1308211839 tempest-ServersTestFqdnHostnames-1308211839-project-member] Releasing lock "refresh_cache-5cb09b60-70b6-4d04-850e-049612ec4a89" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 806.452874] env[61957]: DEBUG nova.compute.manager [None req-d6c9937c-2526-4768-bade-4ca2ced8b569 tempest-ServersTestFqdnHostnames-1308211839 tempest-ServersTestFqdnHostnames-1308211839-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61957) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 806.453287] env[61957]: DEBUG nova.compute.manager [None req-d6c9937c-2526-4768-bade-4ca2ced8b569 tempest-ServersTestFqdnHostnames-1308211839 tempest-ServersTestFqdnHostnames-1308211839-project-member] [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 806.453287] env[61957]: DEBUG nova.network.neutron [None req-d6c9937c-2526-4768-bade-4ca2ced8b569 tempest-ServersTestFqdnHostnames-1308211839 tempest-ServersTestFqdnHostnames-1308211839-project-member] [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 806.487474] env[61957]: DEBUG nova.network.neutron [None req-d6c9937c-2526-4768-bade-4ca2ced8b569 tempest-ServersTestFqdnHostnames-1308211839 tempest-ServersTestFqdnHostnames-1308211839-project-member] [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 806.599424] env[61957]: DEBUG nova.scheduler.client.report [None req-bb8cb37b-74c5-4607-9ba8-e5615134a1da tempest-InstanceActionsTestJSON-565478614 tempest-InstanceActionsTestJSON-565478614-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 806.676702] env[61957]: DEBUG oslo_concurrency.lockutils [None req-da68b092-4602-4484-b1b0-adcac4aae9f8 tempest-InstanceActionsV221TestJSON-1768084888 tempest-InstanceActionsV221TestJSON-1768084888-project-member] Acquiring lock "refresh_cache-6e7bd89c-2c2a-450a-9858-3526d96c28ab" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 806.677673] env[61957]: DEBUG oslo_concurrency.lockutils [None req-da68b092-4602-4484-b1b0-adcac4aae9f8 tempest-InstanceActionsV221TestJSON-1768084888 tempest-InstanceActionsV221TestJSON-1768084888-project-member] Acquired lock "refresh_cache-6e7bd89c-2c2a-450a-9858-3526d96c28ab" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 806.677673] env[61957]: DEBUG nova.network.neutron [None req-da68b092-4602-4484-b1b0-adcac4aae9f8 tempest-InstanceActionsV221TestJSON-1768084888 tempest-InstanceActionsV221TestJSON-1768084888-project-member] [instance: 6e7bd89c-2c2a-450a-9858-3526d96c28ab] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 806.990328] env[61957]: DEBUG nova.network.neutron [None req-d6c9937c-2526-4768-bade-4ca2ced8b569 tempest-ServersTestFqdnHostnames-1308211839 tempest-ServersTestFqdnHostnames-1308211839-project-member] [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 807.106231] env[61957]: DEBUG oslo_concurrency.lockutils [None req-bb8cb37b-74c5-4607-9ba8-e5615134a1da tempest-InstanceActionsTestJSON-565478614 tempest-InstanceActionsTestJSON-565478614-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.907s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 807.106866] env[61957]: ERROR nova.compute.manager [None req-bb8cb37b-74c5-4607-9ba8-e5615134a1da tempest-InstanceActionsTestJSON-565478614 tempest-InstanceActionsTestJSON-565478614-project-member] [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 5ec2282a-722a-45ab-9124-906367749fe7, please check neutron logs for more information. [ 807.106866] env[61957]: ERROR nova.compute.manager [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] Traceback (most recent call last): [ 807.106866] env[61957]: ERROR nova.compute.manager [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 807.106866] env[61957]: ERROR nova.compute.manager [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] self.driver.spawn(context, instance, image_meta, [ 807.106866] env[61957]: ERROR nova.compute.manager [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 807.106866] env[61957]: ERROR nova.compute.manager [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 807.106866] env[61957]: ERROR nova.compute.manager [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 807.106866] env[61957]: ERROR nova.compute.manager [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] vm_ref = self.build_virtual_machine(instance, [ 807.106866] env[61957]: ERROR nova.compute.manager [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 807.106866] env[61957]: ERROR nova.compute.manager [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] vif_infos = vmwarevif.get_vif_info(self._session, [ 807.106866] env[61957]: ERROR nova.compute.manager [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 807.107290] env[61957]: ERROR nova.compute.manager [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] for vif in network_info: [ 807.107290] env[61957]: ERROR nova.compute.manager [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 807.107290] env[61957]: ERROR nova.compute.manager [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] return self._sync_wrapper(fn, *args, **kwargs) [ 807.107290] env[61957]: ERROR nova.compute.manager [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 807.107290] env[61957]: ERROR nova.compute.manager [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] self.wait() [ 807.107290] env[61957]: ERROR nova.compute.manager [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 807.107290] env[61957]: ERROR nova.compute.manager [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] self[:] = self._gt.wait() [ 807.107290] env[61957]: ERROR nova.compute.manager [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 807.107290] env[61957]: ERROR nova.compute.manager [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] return self._exit_event.wait() [ 807.107290] env[61957]: ERROR nova.compute.manager [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 807.107290] env[61957]: ERROR nova.compute.manager [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] current.throw(*self._exc) [ 807.107290] env[61957]: ERROR nova.compute.manager [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 807.107290] env[61957]: ERROR nova.compute.manager [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] result = function(*args, **kwargs) [ 807.107770] env[61957]: ERROR nova.compute.manager [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 807.107770] env[61957]: ERROR nova.compute.manager [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] return func(*args, **kwargs) [ 807.107770] env[61957]: ERROR nova.compute.manager [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 807.107770] env[61957]: ERROR nova.compute.manager [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] raise e [ 807.107770] env[61957]: ERROR nova.compute.manager [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 807.107770] env[61957]: ERROR nova.compute.manager [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] nwinfo = self.network_api.allocate_for_instance( [ 807.107770] env[61957]: ERROR nova.compute.manager [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 807.107770] env[61957]: ERROR nova.compute.manager [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] created_port_ids = self._update_ports_for_instance( [ 807.107770] env[61957]: ERROR nova.compute.manager [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 807.107770] env[61957]: ERROR nova.compute.manager [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] with excutils.save_and_reraise_exception(): [ 807.107770] env[61957]: ERROR nova.compute.manager [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 807.107770] env[61957]: ERROR nova.compute.manager [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] self.force_reraise() [ 807.107770] env[61957]: ERROR nova.compute.manager [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 807.108152] env[61957]: ERROR nova.compute.manager [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] raise self.value [ 807.108152] env[61957]: ERROR nova.compute.manager [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 807.108152] env[61957]: ERROR nova.compute.manager [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] updated_port = self._update_port( [ 807.108152] env[61957]: ERROR nova.compute.manager [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 807.108152] env[61957]: ERROR nova.compute.manager [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] _ensure_no_port_binding_failure(port) [ 807.108152] env[61957]: ERROR nova.compute.manager [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 807.108152] env[61957]: ERROR nova.compute.manager [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] raise exception.PortBindingFailed(port_id=port['id']) [ 807.108152] env[61957]: ERROR nova.compute.manager [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] nova.exception.PortBindingFailed: Binding failed for port 5ec2282a-722a-45ab-9124-906367749fe7, please check neutron logs for more information. [ 807.108152] env[61957]: ERROR nova.compute.manager [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] [ 807.108152] env[61957]: DEBUG nova.compute.utils [None req-bb8cb37b-74c5-4607-9ba8-e5615134a1da tempest-InstanceActionsTestJSON-565478614 tempest-InstanceActionsTestJSON-565478614-project-member] [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] Binding failed for port 5ec2282a-722a-45ab-9124-906367749fe7, please check neutron logs for more information. {{(pid=61957) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 807.110024] env[61957]: DEBUG oslo_concurrency.lockutils [None req-7860e034-eff8-489c-addf-33cf2776246c tempest-ServerActionsTestOtherA-1610268163 tempest-ServerActionsTestOtherA-1610268163-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 25.826s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 807.118301] env[61957]: DEBUG nova.compute.manager [None req-bb8cb37b-74c5-4607-9ba8-e5615134a1da tempest-InstanceActionsTestJSON-565478614 tempest-InstanceActionsTestJSON-565478614-project-member] [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] Build of instance 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e was re-scheduled: Binding failed for port 5ec2282a-722a-45ab-9124-906367749fe7, please check neutron logs for more information. {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 807.118301] env[61957]: DEBUG nova.compute.manager [None req-bb8cb37b-74c5-4607-9ba8-e5615134a1da tempest-InstanceActionsTestJSON-565478614 tempest-InstanceActionsTestJSON-565478614-project-member] [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] Unplugging VIFs for instance {{(pid=61957) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 807.118301] env[61957]: DEBUG oslo_concurrency.lockutils [None req-bb8cb37b-74c5-4607-9ba8-e5615134a1da tempest-InstanceActionsTestJSON-565478614 tempest-InstanceActionsTestJSON-565478614-project-member] Acquiring lock "refresh_cache-2f84d2df-00f0-4860-acb4-4e10ecf8ce4e" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 807.118301] env[61957]: DEBUG oslo_concurrency.lockutils [None req-bb8cb37b-74c5-4607-9ba8-e5615134a1da tempest-InstanceActionsTestJSON-565478614 tempest-InstanceActionsTestJSON-565478614-project-member] Acquired lock "refresh_cache-2f84d2df-00f0-4860-acb4-4e10ecf8ce4e" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 807.118624] env[61957]: DEBUG nova.network.neutron [None req-bb8cb37b-74c5-4607-9ba8-e5615134a1da tempest-InstanceActionsTestJSON-565478614 tempest-InstanceActionsTestJSON-565478614-project-member] [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 807.226055] env[61957]: DEBUG nova.network.neutron [None req-da68b092-4602-4484-b1b0-adcac4aae9f8 tempest-InstanceActionsV221TestJSON-1768084888 tempest-InstanceActionsV221TestJSON-1768084888-project-member] [instance: 6e7bd89c-2c2a-450a-9858-3526d96c28ab] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 807.456049] env[61957]: DEBUG nova.network.neutron [None req-da68b092-4602-4484-b1b0-adcac4aae9f8 tempest-InstanceActionsV221TestJSON-1768084888 tempest-InstanceActionsV221TestJSON-1768084888-project-member] [instance: 6e7bd89c-2c2a-450a-9858-3526d96c28ab] Updating instance_info_cache with network_info: [{"id": "54e92ab2-af65-4141-994f-2ff5ddb01f84", "address": "fa:16:3e:7e:36:5c", "network": {"id": "8e520db8-a454-455c-b3ed-3e2f6b951c28", "bridge": "br-int", "label": "tempest-InstanceActionsV221TestJSON-1491061537-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "645818a1252d49fcbea063f2215923e5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f87a752-ebb0-49a4-a67b-e356fa45b89b", "external-id": "nsx-vlan-transportzone-889", "segmentation_id": 889, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap54e92ab2-af", "ovs_interfaceid": "54e92ab2-af65-4141-994f-2ff5ddb01f84", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 807.496384] env[61957]: INFO nova.compute.manager [None req-d6c9937c-2526-4768-bade-4ca2ced8b569 tempest-ServersTestFqdnHostnames-1308211839 tempest-ServersTestFqdnHostnames-1308211839-project-member] [instance: 5cb09b60-70b6-4d04-850e-049612ec4a89] Took 1.04 seconds to deallocate network for instance. [ 807.646458] env[61957]: DEBUG nova.network.neutron [None req-bb8cb37b-74c5-4607-9ba8-e5615134a1da tempest-InstanceActionsTestJSON-565478614 tempest-InstanceActionsTestJSON-565478614-project-member] [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 807.774679] env[61957]: DEBUG oslo_concurrency.lockutils [None req-c13ea078-3782-4688-93ee-408e1ac828c0 tempest-ServerTagsTestJSON-969152625 tempest-ServerTagsTestJSON-969152625-project-member] Acquiring lock "8e04a098-25de-4d57-9e3c-ea44d234d57e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 807.775227] env[61957]: DEBUG oslo_concurrency.lockutils [None req-c13ea078-3782-4688-93ee-408e1ac828c0 tempest-ServerTagsTestJSON-969152625 tempest-ServerTagsTestJSON-969152625-project-member] Lock "8e04a098-25de-4d57-9e3c-ea44d234d57e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 807.812129] env[61957]: DEBUG nova.network.neutron [None req-bb8cb37b-74c5-4607-9ba8-e5615134a1da tempest-InstanceActionsTestJSON-565478614 tempest-InstanceActionsTestJSON-565478614-project-member] [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 807.946521] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3156066f-e38e-4f66-9309-bb9d1e12904b {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.955813] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-caf7d630-b805-4901-b170-3d5cf7770642 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.960170] env[61957]: DEBUG oslo_concurrency.lockutils [None req-da68b092-4602-4484-b1b0-adcac4aae9f8 tempest-InstanceActionsV221TestJSON-1768084888 tempest-InstanceActionsV221TestJSON-1768084888-project-member] Releasing lock "refresh_cache-6e7bd89c-2c2a-450a-9858-3526d96c28ab" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 807.960478] env[61957]: DEBUG nova.compute.manager [None req-da68b092-4602-4484-b1b0-adcac4aae9f8 tempest-InstanceActionsV221TestJSON-1768084888 tempest-InstanceActionsV221TestJSON-1768084888-project-member] [instance: 6e7bd89c-2c2a-450a-9858-3526d96c28ab] Instance network_info: |[{"id": "54e92ab2-af65-4141-994f-2ff5ddb01f84", "address": "fa:16:3e:7e:36:5c", "network": {"id": "8e520db8-a454-455c-b3ed-3e2f6b951c28", "bridge": "br-int", "label": "tempest-InstanceActionsV221TestJSON-1491061537-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "645818a1252d49fcbea063f2215923e5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f87a752-ebb0-49a4-a67b-e356fa45b89b", "external-id": "nsx-vlan-transportzone-889", "segmentation_id": 889, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap54e92ab2-af", "ovs_interfaceid": "54e92ab2-af65-4141-994f-2ff5ddb01f84", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 807.961272] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-da68b092-4602-4484-b1b0-adcac4aae9f8 tempest-InstanceActionsV221TestJSON-1768084888 tempest-InstanceActionsV221TestJSON-1768084888-project-member] [instance: 6e7bd89c-2c2a-450a-9858-3526d96c28ab] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:7e:36:5c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9f87a752-ebb0-49a4-a67b-e356fa45b89b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '54e92ab2-af65-4141-994f-2ff5ddb01f84', 'vif_model': 'vmxnet3'}] {{(pid=61957) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 807.968801] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-da68b092-4602-4484-b1b0-adcac4aae9f8 tempest-InstanceActionsV221TestJSON-1768084888 tempest-InstanceActionsV221TestJSON-1768084888-project-member] Creating folder: Project (645818a1252d49fcbea063f2215923e5). Parent ref: group-v274445. {{(pid=61957) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 807.992513] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-dbb2f21e-ed70-4a03-8cab-5cf9039fa0f4 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.994916] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-016716c3-3609-4d66-8341-79b4ebccd40d {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.002676] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-504691e5-6bb1-4c71-90dc-275bf1b69bfc {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.010844] env[61957]: INFO nova.virt.vmwareapi.vm_util [None req-da68b092-4602-4484-b1b0-adcac4aae9f8 tempest-InstanceActionsV221TestJSON-1768084888 tempest-InstanceActionsV221TestJSON-1768084888-project-member] Created folder: Project (645818a1252d49fcbea063f2215923e5) in parent group-v274445. [ 808.010969] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-da68b092-4602-4484-b1b0-adcac4aae9f8 tempest-InstanceActionsV221TestJSON-1768084888 tempest-InstanceActionsV221TestJSON-1768084888-project-member] Creating folder: Instances. Parent ref: group-v274469. {{(pid=61957) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 808.011691] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-94bcc804-fafc-4aed-a16a-b60172ae9b3f {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.020632] env[61957]: DEBUG nova.compute.provider_tree [None req-7860e034-eff8-489c-addf-33cf2776246c tempest-ServerActionsTestOtherA-1610268163 tempest-ServerActionsTestOtherA-1610268163-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 808.030526] env[61957]: INFO nova.virt.vmwareapi.vm_util [None req-da68b092-4602-4484-b1b0-adcac4aae9f8 tempest-InstanceActionsV221TestJSON-1768084888 tempest-InstanceActionsV221TestJSON-1768084888-project-member] Created folder: Instances in parent group-v274469. [ 808.030773] env[61957]: DEBUG oslo.service.loopingcall [None req-da68b092-4602-4484-b1b0-adcac4aae9f8 tempest-InstanceActionsV221TestJSON-1768084888 tempest-InstanceActionsV221TestJSON-1768084888-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 808.031446] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6e7bd89c-2c2a-450a-9858-3526d96c28ab] Creating VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 808.031446] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a779a2e9-65b5-4b63-8725-3df4615ecd4f {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.051486] env[61957]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 808.051486] env[61957]: value = "task-1277390" [ 808.051486] env[61957]: _type = "Task" [ 808.051486] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.060600] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277390, 'name': CreateVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.086200] env[61957]: DEBUG nova.compute.manager [req-c0a8523b-b203-47ee-be3a-88fc907dad19 req-c3ee1161-2b84-4c09-92fe-9671a604896b service nova] [instance: 6e7bd89c-2c2a-450a-9858-3526d96c28ab] Received event network-changed-54e92ab2-af65-4141-994f-2ff5ddb01f84 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 808.086512] env[61957]: DEBUG nova.compute.manager [req-c0a8523b-b203-47ee-be3a-88fc907dad19 req-c3ee1161-2b84-4c09-92fe-9671a604896b service nova] [instance: 6e7bd89c-2c2a-450a-9858-3526d96c28ab] Refreshing instance network info cache due to event network-changed-54e92ab2-af65-4141-994f-2ff5ddb01f84. {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 808.086784] env[61957]: DEBUG oslo_concurrency.lockutils [req-c0a8523b-b203-47ee-be3a-88fc907dad19 req-c3ee1161-2b84-4c09-92fe-9671a604896b service nova] Acquiring lock "refresh_cache-6e7bd89c-2c2a-450a-9858-3526d96c28ab" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 808.087058] env[61957]: DEBUG oslo_concurrency.lockutils [req-c0a8523b-b203-47ee-be3a-88fc907dad19 req-c3ee1161-2b84-4c09-92fe-9671a604896b service nova] Acquired lock "refresh_cache-6e7bd89c-2c2a-450a-9858-3526d96c28ab" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 808.087353] env[61957]: DEBUG nova.network.neutron [req-c0a8523b-b203-47ee-be3a-88fc907dad19 req-c3ee1161-2b84-4c09-92fe-9671a604896b service nova] [instance: 6e7bd89c-2c2a-450a-9858-3526d96c28ab] Refreshing network info cache for port 54e92ab2-af65-4141-994f-2ff5ddb01f84 {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 808.313778] env[61957]: DEBUG oslo_concurrency.lockutils [None req-bb8cb37b-74c5-4607-9ba8-e5615134a1da tempest-InstanceActionsTestJSON-565478614 tempest-InstanceActionsTestJSON-565478614-project-member] Releasing lock "refresh_cache-2f84d2df-00f0-4860-acb4-4e10ecf8ce4e" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 808.314088] env[61957]: DEBUG nova.compute.manager [None req-bb8cb37b-74c5-4607-9ba8-e5615134a1da tempest-InstanceActionsTestJSON-565478614 tempest-InstanceActionsTestJSON-565478614-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61957) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 808.314322] env[61957]: DEBUG nova.compute.manager [None req-bb8cb37b-74c5-4607-9ba8-e5615134a1da tempest-InstanceActionsTestJSON-565478614 tempest-InstanceActionsTestJSON-565478614-project-member] [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 808.315230] env[61957]: DEBUG nova.network.neutron [None req-bb8cb37b-74c5-4607-9ba8-e5615134a1da tempest-InstanceActionsTestJSON-565478614 tempest-InstanceActionsTestJSON-565478614-project-member] [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 808.340641] env[61957]: DEBUG nova.network.neutron [None req-bb8cb37b-74c5-4607-9ba8-e5615134a1da tempest-InstanceActionsTestJSON-565478614 tempest-InstanceActionsTestJSON-565478614-project-member] [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 808.523554] env[61957]: DEBUG nova.scheduler.client.report [None req-7860e034-eff8-489c-addf-33cf2776246c tempest-ServerActionsTestOtherA-1610268163 tempest-ServerActionsTestOtherA-1610268163-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 808.538779] env[61957]: INFO nova.scheduler.client.report [None req-d6c9937c-2526-4768-bade-4ca2ced8b569 tempest-ServersTestFqdnHostnames-1308211839 tempest-ServersTestFqdnHostnames-1308211839-project-member] Deleted allocations for instance 5cb09b60-70b6-4d04-850e-049612ec4a89 [ 808.568314] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277390, 'name': CreateVM_Task, 'duration_secs': 0.297166} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.568513] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6e7bd89c-2c2a-450a-9858-3526d96c28ab] Created VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 808.579094] env[61957]: DEBUG oslo_concurrency.lockutils [None req-da68b092-4602-4484-b1b0-adcac4aae9f8 tempest-InstanceActionsV221TestJSON-1768084888 tempest-InstanceActionsV221TestJSON-1768084888-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 808.579370] env[61957]: DEBUG oslo_concurrency.lockutils [None req-da68b092-4602-4484-b1b0-adcac4aae9f8 tempest-InstanceActionsV221TestJSON-1768084888 tempest-InstanceActionsV221TestJSON-1768084888-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 808.579697] env[61957]: DEBUG oslo_concurrency.lockutils [None req-da68b092-4602-4484-b1b0-adcac4aae9f8 tempest-InstanceActionsV221TestJSON-1768084888 tempest-InstanceActionsV221TestJSON-1768084888-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 808.580162] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-32532dea-eb58-4355-baa4-804ecba009cd {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.585392] env[61957]: DEBUG oslo_vmware.api [None req-da68b092-4602-4484-b1b0-adcac4aae9f8 tempest-InstanceActionsV221TestJSON-1768084888 tempest-InstanceActionsV221TestJSON-1768084888-project-member] Waiting for the task: (returnval){ [ 808.585392] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]524e27f5-4dcc-8a86-291c-c8c2a9ec56eb" [ 808.585392] env[61957]: _type = "Task" [ 808.585392] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.595625] env[61957]: DEBUG oslo_vmware.api [None req-da68b092-4602-4484-b1b0-adcac4aae9f8 tempest-InstanceActionsV221TestJSON-1768084888 tempest-InstanceActionsV221TestJSON-1768084888-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]524e27f5-4dcc-8a86-291c-c8c2a9ec56eb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.815333] env[61957]: DEBUG nova.network.neutron [req-c0a8523b-b203-47ee-be3a-88fc907dad19 req-c3ee1161-2b84-4c09-92fe-9671a604896b service nova] [instance: 6e7bd89c-2c2a-450a-9858-3526d96c28ab] Updated VIF entry in instance network info cache for port 54e92ab2-af65-4141-994f-2ff5ddb01f84. {{(pid=61957) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 808.815333] env[61957]: DEBUG nova.network.neutron [req-c0a8523b-b203-47ee-be3a-88fc907dad19 req-c3ee1161-2b84-4c09-92fe-9671a604896b service nova] [instance: 6e7bd89c-2c2a-450a-9858-3526d96c28ab] Updating instance_info_cache with network_info: [{"id": "54e92ab2-af65-4141-994f-2ff5ddb01f84", "address": "fa:16:3e:7e:36:5c", "network": {"id": "8e520db8-a454-455c-b3ed-3e2f6b951c28", "bridge": "br-int", "label": "tempest-InstanceActionsV221TestJSON-1491061537-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "645818a1252d49fcbea063f2215923e5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f87a752-ebb0-49a4-a67b-e356fa45b89b", "external-id": "nsx-vlan-transportzone-889", "segmentation_id": 889, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap54e92ab2-af", "ovs_interfaceid": "54e92ab2-af65-4141-994f-2ff5ddb01f84", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 808.843742] env[61957]: DEBUG nova.network.neutron [None req-bb8cb37b-74c5-4607-9ba8-e5615134a1da tempest-InstanceActionsTestJSON-565478614 tempest-InstanceActionsTestJSON-565478614-project-member] [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 809.034018] env[61957]: DEBUG oslo_concurrency.lockutils [None req-7860e034-eff8-489c-addf-33cf2776246c tempest-ServerActionsTestOtherA-1610268163 tempest-ServerActionsTestOtherA-1610268163-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.922s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 809.034018] env[61957]: ERROR nova.compute.manager [None req-7860e034-eff8-489c-addf-33cf2776246c tempest-ServerActionsTestOtherA-1610268163 tempest-ServerActionsTestOtherA-1610268163-project-member] [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 76ccb95a-3bfc-4be2-a607-679432fcf74d, please check neutron logs for more information. [ 809.034018] env[61957]: ERROR nova.compute.manager [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] Traceback (most recent call last): [ 809.034018] env[61957]: ERROR nova.compute.manager [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 809.034018] env[61957]: ERROR nova.compute.manager [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] self.driver.spawn(context, instance, image_meta, [ 809.034018] env[61957]: ERROR nova.compute.manager [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 809.034018] env[61957]: ERROR nova.compute.manager [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] self._vmops.spawn(context, instance, image_meta, injected_files, [ 809.034018] env[61957]: ERROR nova.compute.manager [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 809.034018] env[61957]: ERROR nova.compute.manager [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] vm_ref = self.build_virtual_machine(instance, [ 809.034378] env[61957]: ERROR nova.compute.manager [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 809.034378] env[61957]: ERROR nova.compute.manager [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] vif_infos = vmwarevif.get_vif_info(self._session, [ 809.034378] env[61957]: ERROR nova.compute.manager [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 809.034378] env[61957]: ERROR nova.compute.manager [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] for vif in network_info: [ 809.034378] env[61957]: ERROR nova.compute.manager [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 809.034378] env[61957]: ERROR nova.compute.manager [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] return self._sync_wrapper(fn, *args, **kwargs) [ 809.034378] env[61957]: ERROR nova.compute.manager [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 809.034378] env[61957]: ERROR nova.compute.manager [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] self.wait() [ 809.034378] env[61957]: ERROR nova.compute.manager [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 809.034378] env[61957]: ERROR nova.compute.manager [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] self[:] = self._gt.wait() [ 809.034378] env[61957]: ERROR nova.compute.manager [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 809.034378] env[61957]: ERROR nova.compute.manager [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] return self._exit_event.wait() [ 809.034378] env[61957]: ERROR nova.compute.manager [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 809.034802] env[61957]: ERROR nova.compute.manager [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] current.throw(*self._exc) [ 809.034802] env[61957]: ERROR nova.compute.manager [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 809.034802] env[61957]: ERROR nova.compute.manager [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] result = function(*args, **kwargs) [ 809.034802] env[61957]: ERROR nova.compute.manager [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 809.034802] env[61957]: ERROR nova.compute.manager [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] return func(*args, **kwargs) [ 809.034802] env[61957]: ERROR nova.compute.manager [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 809.034802] env[61957]: ERROR nova.compute.manager [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] raise e [ 809.034802] env[61957]: ERROR nova.compute.manager [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 809.034802] env[61957]: ERROR nova.compute.manager [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] nwinfo = self.network_api.allocate_for_instance( [ 809.034802] env[61957]: ERROR nova.compute.manager [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 809.034802] env[61957]: ERROR nova.compute.manager [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] created_port_ids = self._update_ports_for_instance( [ 809.034802] env[61957]: ERROR nova.compute.manager [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 809.034802] env[61957]: ERROR nova.compute.manager [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] with excutils.save_and_reraise_exception(): [ 809.035279] env[61957]: ERROR nova.compute.manager [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 809.035279] env[61957]: ERROR nova.compute.manager [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] self.force_reraise() [ 809.035279] env[61957]: ERROR nova.compute.manager [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 809.035279] env[61957]: ERROR nova.compute.manager [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] raise self.value [ 809.035279] env[61957]: ERROR nova.compute.manager [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 809.035279] env[61957]: ERROR nova.compute.manager [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] updated_port = self._update_port( [ 809.035279] env[61957]: ERROR nova.compute.manager [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 809.035279] env[61957]: ERROR nova.compute.manager [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] _ensure_no_port_binding_failure(port) [ 809.035279] env[61957]: ERROR nova.compute.manager [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 809.035279] env[61957]: ERROR nova.compute.manager [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] raise exception.PortBindingFailed(port_id=port['id']) [ 809.035279] env[61957]: ERROR nova.compute.manager [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] nova.exception.PortBindingFailed: Binding failed for port 76ccb95a-3bfc-4be2-a607-679432fcf74d, please check neutron logs for more information. [ 809.035279] env[61957]: ERROR nova.compute.manager [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] [ 809.035682] env[61957]: DEBUG nova.compute.utils [None req-7860e034-eff8-489c-addf-33cf2776246c tempest-ServerActionsTestOtherA-1610268163 tempest-ServerActionsTestOtherA-1610268163-project-member] [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] Binding failed for port 76ccb95a-3bfc-4be2-a607-679432fcf74d, please check neutron logs for more information. {{(pid=61957) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 809.035682] env[61957]: DEBUG oslo_concurrency.lockutils [None req-182a8e01-a1e1-4ecf-8692-2abcfc9944ba tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.308s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 809.036874] env[61957]: INFO nova.compute.claims [None req-182a8e01-a1e1-4ecf-8692-2abcfc9944ba tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] [instance: e0249936-d616-4ffb-8f77-d8107633c42a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 809.040056] env[61957]: DEBUG nova.compute.manager [None req-7860e034-eff8-489c-addf-33cf2776246c tempest-ServerActionsTestOtherA-1610268163 tempest-ServerActionsTestOtherA-1610268163-project-member] [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] Build of instance 6372a384-61a8-43e8-8f6c-376e2a735045 was re-scheduled: Binding failed for port 76ccb95a-3bfc-4be2-a607-679432fcf74d, please check neutron logs for more information. {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 809.040630] env[61957]: DEBUG nova.compute.manager [None req-7860e034-eff8-489c-addf-33cf2776246c tempest-ServerActionsTestOtherA-1610268163 tempest-ServerActionsTestOtherA-1610268163-project-member] [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] Unplugging VIFs for instance {{(pid=61957) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 809.040874] env[61957]: DEBUG oslo_concurrency.lockutils [None req-7860e034-eff8-489c-addf-33cf2776246c tempest-ServerActionsTestOtherA-1610268163 tempest-ServerActionsTestOtherA-1610268163-project-member] Acquiring lock "refresh_cache-6372a384-61a8-43e8-8f6c-376e2a735045" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 809.041024] env[61957]: DEBUG oslo_concurrency.lockutils [None req-7860e034-eff8-489c-addf-33cf2776246c tempest-ServerActionsTestOtherA-1610268163 tempest-ServerActionsTestOtherA-1610268163-project-member] Acquired lock "refresh_cache-6372a384-61a8-43e8-8f6c-376e2a735045" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 809.041187] env[61957]: DEBUG nova.network.neutron [None req-7860e034-eff8-489c-addf-33cf2776246c tempest-ServerActionsTestOtherA-1610268163 tempest-ServerActionsTestOtherA-1610268163-project-member] [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 809.045664] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d6c9937c-2526-4768-bade-4ca2ced8b569 tempest-ServersTestFqdnHostnames-1308211839 tempest-ServersTestFqdnHostnames-1308211839-project-member] Lock "5cb09b60-70b6-4d04-850e-049612ec4a89" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 155.847s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 809.099301] env[61957]: DEBUG oslo_vmware.api [None req-da68b092-4602-4484-b1b0-adcac4aae9f8 tempest-InstanceActionsV221TestJSON-1768084888 tempest-InstanceActionsV221TestJSON-1768084888-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]524e27f5-4dcc-8a86-291c-c8c2a9ec56eb, 'name': SearchDatastore_Task, 'duration_secs': 0.009202} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.100291] env[61957]: DEBUG oslo_concurrency.lockutils [None req-da68b092-4602-4484-b1b0-adcac4aae9f8 tempest-InstanceActionsV221TestJSON-1768084888 tempest-InstanceActionsV221TestJSON-1768084888-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 809.100626] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-da68b092-4602-4484-b1b0-adcac4aae9f8 tempest-InstanceActionsV221TestJSON-1768084888 tempest-InstanceActionsV221TestJSON-1768084888-project-member] [instance: 6e7bd89c-2c2a-450a-9858-3526d96c28ab] Processing image 11c76a2c-f705-470a-ba9d-4657858bab38 {{(pid=61957) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 809.100900] env[61957]: DEBUG oslo_concurrency.lockutils [None req-da68b092-4602-4484-b1b0-adcac4aae9f8 tempest-InstanceActionsV221TestJSON-1768084888 tempest-InstanceActionsV221TestJSON-1768084888-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 809.101088] env[61957]: DEBUG oslo_concurrency.lockutils [None req-da68b092-4602-4484-b1b0-adcac4aae9f8 tempest-InstanceActionsV221TestJSON-1768084888 tempest-InstanceActionsV221TestJSON-1768084888-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 809.101293] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-da68b092-4602-4484-b1b0-adcac4aae9f8 tempest-InstanceActionsV221TestJSON-1768084888 tempest-InstanceActionsV221TestJSON-1768084888-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 809.101842] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-bc66b995-34b0-425d-8df9-709e11c1a3be {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.110950] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-da68b092-4602-4484-b1b0-adcac4aae9f8 tempest-InstanceActionsV221TestJSON-1768084888 tempest-InstanceActionsV221TestJSON-1768084888-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 809.111267] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-da68b092-4602-4484-b1b0-adcac4aae9f8 tempest-InstanceActionsV221TestJSON-1768084888 tempest-InstanceActionsV221TestJSON-1768084888-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61957) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 809.112338] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-72e59e28-5d3d-497b-baaf-c1fbb9e77a88 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.120761] env[61957]: DEBUG oslo_vmware.api [None req-da68b092-4602-4484-b1b0-adcac4aae9f8 tempest-InstanceActionsV221TestJSON-1768084888 tempest-InstanceActionsV221TestJSON-1768084888-project-member] Waiting for the task: (returnval){ [ 809.120761] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52c2ed4b-086e-b519-f4a8-23aac8a44b04" [ 809.120761] env[61957]: _type = "Task" [ 809.120761] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.128774] env[61957]: DEBUG oslo_vmware.api [None req-da68b092-4602-4484-b1b0-adcac4aae9f8 tempest-InstanceActionsV221TestJSON-1768084888 tempest-InstanceActionsV221TestJSON-1768084888-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52c2ed4b-086e-b519-f4a8-23aac8a44b04, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.316975] env[61957]: DEBUG oslo_concurrency.lockutils [req-c0a8523b-b203-47ee-be3a-88fc907dad19 req-c3ee1161-2b84-4c09-92fe-9671a604896b service nova] Releasing lock "refresh_cache-6e7bd89c-2c2a-450a-9858-3526d96c28ab" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 809.345914] env[61957]: INFO nova.compute.manager [None req-bb8cb37b-74c5-4607-9ba8-e5615134a1da tempest-InstanceActionsTestJSON-565478614 tempest-InstanceActionsTestJSON-565478614-project-member] [instance: 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e] Took 1.03 seconds to deallocate network for instance. [ 809.551028] env[61957]: DEBUG nova.compute.manager [None req-7bbc9489-a498-401d-a2af-2f28d461499e tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 4951132e-7247-4772-8f88-3664c6a7e61e] Starting instance... {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 809.593969] env[61957]: DEBUG nova.network.neutron [None req-7860e034-eff8-489c-addf-33cf2776246c tempest-ServerActionsTestOtherA-1610268163 tempest-ServerActionsTestOtherA-1610268163-project-member] [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 809.629830] env[61957]: DEBUG oslo_vmware.api [None req-da68b092-4602-4484-b1b0-adcac4aae9f8 tempest-InstanceActionsV221TestJSON-1768084888 tempest-InstanceActionsV221TestJSON-1768084888-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52c2ed4b-086e-b519-f4a8-23aac8a44b04, 'name': SearchDatastore_Task, 'duration_secs': 0.008553} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.630665] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-382e4cc7-48ae-4092-9426-0ae2be00b354 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.636030] env[61957]: DEBUG oslo_vmware.api [None req-da68b092-4602-4484-b1b0-adcac4aae9f8 tempest-InstanceActionsV221TestJSON-1768084888 tempest-InstanceActionsV221TestJSON-1768084888-project-member] Waiting for the task: (returnval){ [ 809.636030] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52cfa47e-208c-71c1-7c67-ae27ce946506" [ 809.636030] env[61957]: _type = "Task" [ 809.636030] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.645806] env[61957]: DEBUG oslo_vmware.api [None req-da68b092-4602-4484-b1b0-adcac4aae9f8 tempest-InstanceActionsV221TestJSON-1768084888 tempest-InstanceActionsV221TestJSON-1768084888-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52cfa47e-208c-71c1-7c67-ae27ce946506, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.725998] env[61957]: DEBUG nova.network.neutron [None req-7860e034-eff8-489c-addf-33cf2776246c tempest-ServerActionsTestOtherA-1610268163 tempest-ServerActionsTestOtherA-1610268163-project-member] [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 810.080104] env[61957]: DEBUG oslo_concurrency.lockutils [None req-7bbc9489-a498-401d-a2af-2f28d461499e tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 810.148110] env[61957]: DEBUG oslo_vmware.api [None req-da68b092-4602-4484-b1b0-adcac4aae9f8 tempest-InstanceActionsV221TestJSON-1768084888 tempest-InstanceActionsV221TestJSON-1768084888-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52cfa47e-208c-71c1-7c67-ae27ce946506, 'name': SearchDatastore_Task, 'duration_secs': 0.008976} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.148110] env[61957]: DEBUG oslo_concurrency.lockutils [None req-da68b092-4602-4484-b1b0-adcac4aae9f8 tempest-InstanceActionsV221TestJSON-1768084888 tempest-InstanceActionsV221TestJSON-1768084888-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 810.148110] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-da68b092-4602-4484-b1b0-adcac4aae9f8 tempest-InstanceActionsV221TestJSON-1768084888 tempest-InstanceActionsV221TestJSON-1768084888-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore2] 6e7bd89c-2c2a-450a-9858-3526d96c28ab/6e7bd89c-2c2a-450a-9858-3526d96c28ab.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 810.148700] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-59d1b6b0-b64d-4ced-94c9-12b21c213416 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.160340] env[61957]: DEBUG oslo_vmware.api [None req-da68b092-4602-4484-b1b0-adcac4aae9f8 tempest-InstanceActionsV221TestJSON-1768084888 tempest-InstanceActionsV221TestJSON-1768084888-project-member] Waiting for the task: (returnval){ [ 810.160340] env[61957]: value = "task-1277391" [ 810.160340] env[61957]: _type = "Task" [ 810.160340] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.173797] env[61957]: DEBUG oslo_vmware.api [None req-da68b092-4602-4484-b1b0-adcac4aae9f8 tempest-InstanceActionsV221TestJSON-1768084888 tempest-InstanceActionsV221TestJSON-1768084888-project-member] Task: {'id': task-1277391, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.228253] env[61957]: DEBUG oslo_concurrency.lockutils [None req-7860e034-eff8-489c-addf-33cf2776246c tempest-ServerActionsTestOtherA-1610268163 tempest-ServerActionsTestOtherA-1610268163-project-member] Releasing lock "refresh_cache-6372a384-61a8-43e8-8f6c-376e2a735045" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 810.228485] env[61957]: DEBUG nova.compute.manager [None req-7860e034-eff8-489c-addf-33cf2776246c tempest-ServerActionsTestOtherA-1610268163 tempest-ServerActionsTestOtherA-1610268163-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61957) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 810.228716] env[61957]: DEBUG nova.compute.manager [None req-7860e034-eff8-489c-addf-33cf2776246c tempest-ServerActionsTestOtherA-1610268163 tempest-ServerActionsTestOtherA-1610268163-project-member] [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 810.228839] env[61957]: DEBUG nova.network.neutron [None req-7860e034-eff8-489c-addf-33cf2776246c tempest-ServerActionsTestOtherA-1610268163 tempest-ServerActionsTestOtherA-1610268163-project-member] [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 810.267163] env[61957]: DEBUG nova.network.neutron [None req-7860e034-eff8-489c-addf-33cf2776246c tempest-ServerActionsTestOtherA-1610268163 tempest-ServerActionsTestOtherA-1610268163-project-member] [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 810.378338] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3018f400-4973-4a83-a295-6bebb5228d1f {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.387596] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-446712ff-1dd5-4da4-bb19-7429b262ac2b {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.395914] env[61957]: INFO nova.scheduler.client.report [None req-bb8cb37b-74c5-4607-9ba8-e5615134a1da tempest-InstanceActionsTestJSON-565478614 tempest-InstanceActionsTestJSON-565478614-project-member] Deleted allocations for instance 2f84d2df-00f0-4860-acb4-4e10ecf8ce4e [ 810.432990] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b24e7ce-101e-44f6-9474-d27c9af116fc {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.441825] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0179b195-63cc-4e8f-b690-8c2abfbd6070 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.461463] env[61957]: DEBUG nova.compute.provider_tree [None req-182a8e01-a1e1-4ecf-8692-2abcfc9944ba tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 810.671581] env[61957]: DEBUG oslo_vmware.api [None req-da68b092-4602-4484-b1b0-adcac4aae9f8 tempest-InstanceActionsV221TestJSON-1768084888 tempest-InstanceActionsV221TestJSON-1768084888-project-member] Task: {'id': task-1277391, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.476313} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.671891] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-da68b092-4602-4484-b1b0-adcac4aae9f8 tempest-InstanceActionsV221TestJSON-1768084888 tempest-InstanceActionsV221TestJSON-1768084888-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore2] 6e7bd89c-2c2a-450a-9858-3526d96c28ab/6e7bd89c-2c2a-450a-9858-3526d96c28ab.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 810.672125] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-da68b092-4602-4484-b1b0-adcac4aae9f8 tempest-InstanceActionsV221TestJSON-1768084888 tempest-InstanceActionsV221TestJSON-1768084888-project-member] [instance: 6e7bd89c-2c2a-450a-9858-3526d96c28ab] Extending root virtual disk to 1048576 {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 810.672373] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8c4150b8-8f9f-4873-83ed-b249f5eb72b9 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.679495] env[61957]: DEBUG oslo_vmware.api [None req-da68b092-4602-4484-b1b0-adcac4aae9f8 tempest-InstanceActionsV221TestJSON-1768084888 tempest-InstanceActionsV221TestJSON-1768084888-project-member] Waiting for the task: (returnval){ [ 810.679495] env[61957]: value = "task-1277392" [ 810.679495] env[61957]: _type = "Task" [ 810.679495] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.687248] env[61957]: DEBUG oslo_vmware.api [None req-da68b092-4602-4484-b1b0-adcac4aae9f8 tempest-InstanceActionsV221TestJSON-1768084888 tempest-InstanceActionsV221TestJSON-1768084888-project-member] Task: {'id': task-1277392, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.771642] env[61957]: DEBUG nova.network.neutron [None req-7860e034-eff8-489c-addf-33cf2776246c tempest-ServerActionsTestOtherA-1610268163 tempest-ServerActionsTestOtherA-1610268163-project-member] [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 810.937991] env[61957]: DEBUG oslo_concurrency.lockutils [None req-bb8cb37b-74c5-4607-9ba8-e5615134a1da tempest-InstanceActionsTestJSON-565478614 tempest-InstanceActionsTestJSON-565478614-project-member] Lock "2f84d2df-00f0-4860-acb4-4e10ecf8ce4e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 156.310s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 810.964949] env[61957]: DEBUG nova.scheduler.client.report [None req-182a8e01-a1e1-4ecf-8692-2abcfc9944ba tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 811.194568] env[61957]: DEBUG oslo_vmware.api [None req-da68b092-4602-4484-b1b0-adcac4aae9f8 tempest-InstanceActionsV221TestJSON-1768084888 tempest-InstanceActionsV221TestJSON-1768084888-project-member] Task: {'id': task-1277392, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071626} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.194568] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-da68b092-4602-4484-b1b0-adcac4aae9f8 tempest-InstanceActionsV221TestJSON-1768084888 tempest-InstanceActionsV221TestJSON-1768084888-project-member] [instance: 6e7bd89c-2c2a-450a-9858-3526d96c28ab] Extended root virtual disk {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 811.194568] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-263a6589-94d7-4dbd-8fc9-0f426165a5ad {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.215482] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-da68b092-4602-4484-b1b0-adcac4aae9f8 tempest-InstanceActionsV221TestJSON-1768084888 tempest-InstanceActionsV221TestJSON-1768084888-project-member] [instance: 6e7bd89c-2c2a-450a-9858-3526d96c28ab] Reconfiguring VM instance instance-00000032 to attach disk [datastore2] 6e7bd89c-2c2a-450a-9858-3526d96c28ab/6e7bd89c-2c2a-450a-9858-3526d96c28ab.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 811.215797] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-857586e4-4214-487d-b0cc-ebd2b7de02dd {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.235742] env[61957]: DEBUG oslo_vmware.api [None req-da68b092-4602-4484-b1b0-adcac4aae9f8 tempest-InstanceActionsV221TestJSON-1768084888 tempest-InstanceActionsV221TestJSON-1768084888-project-member] Waiting for the task: (returnval){ [ 811.235742] env[61957]: value = "task-1277393" [ 811.235742] env[61957]: _type = "Task" [ 811.235742] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.246374] env[61957]: DEBUG oslo_vmware.api [None req-da68b092-4602-4484-b1b0-adcac4aae9f8 tempest-InstanceActionsV221TestJSON-1768084888 tempest-InstanceActionsV221TestJSON-1768084888-project-member] Task: {'id': task-1277393, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.274470] env[61957]: INFO nova.compute.manager [None req-7860e034-eff8-489c-addf-33cf2776246c tempest-ServerActionsTestOtherA-1610268163 tempest-ServerActionsTestOtherA-1610268163-project-member] [instance: 6372a384-61a8-43e8-8f6c-376e2a735045] Took 1.05 seconds to deallocate network for instance. [ 811.439214] env[61957]: DEBUG nova.compute.manager [None req-3c3df299-fd1f-460a-8cc2-979af1db30d1 tempest-ServersTestJSON-1093633677 tempest-ServersTestJSON-1093633677-project-member] [instance: ead6aae1-36b5-4f57-9129-3bb02cf103ce] Starting instance... {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 811.474522] env[61957]: DEBUG oslo_concurrency.lockutils [None req-182a8e01-a1e1-4ecf-8692-2abcfc9944ba tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.438s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 811.474522] env[61957]: DEBUG nova.compute.manager [None req-182a8e01-a1e1-4ecf-8692-2abcfc9944ba tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] [instance: e0249936-d616-4ffb-8f77-d8107633c42a] Start building networks asynchronously for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 811.480156] env[61957]: DEBUG oslo_concurrency.lockutils [None req-bc6cc00e-6b7d-4787-8382-d5cf71a4537c tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.793s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 811.480632] env[61957]: INFO nova.compute.claims [None req-bc6cc00e-6b7d-4787-8382-d5cf71a4537c tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] [instance: 615dbbed-2b02-4351-9e03-8c13f424a133] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 811.746146] env[61957]: DEBUG oslo_vmware.api [None req-da68b092-4602-4484-b1b0-adcac4aae9f8 tempest-InstanceActionsV221TestJSON-1768084888 tempest-InstanceActionsV221TestJSON-1768084888-project-member] Task: {'id': task-1277393, 'name': ReconfigVM_Task, 'duration_secs': 0.301626} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.746441] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-da68b092-4602-4484-b1b0-adcac4aae9f8 tempest-InstanceActionsV221TestJSON-1768084888 tempest-InstanceActionsV221TestJSON-1768084888-project-member] [instance: 6e7bd89c-2c2a-450a-9858-3526d96c28ab] Reconfigured VM instance instance-00000032 to attach disk [datastore2] 6e7bd89c-2c2a-450a-9858-3526d96c28ab/6e7bd89c-2c2a-450a-9858-3526d96c28ab.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 811.747086] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-bd47fdc0-1f75-43bd-a68c-2e1aafe943eb {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.753707] env[61957]: DEBUG oslo_vmware.api [None req-da68b092-4602-4484-b1b0-adcac4aae9f8 tempest-InstanceActionsV221TestJSON-1768084888 tempest-InstanceActionsV221TestJSON-1768084888-project-member] Waiting for the task: (returnval){ [ 811.753707] env[61957]: value = "task-1277394" [ 811.753707] env[61957]: _type = "Task" [ 811.753707] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.764311] env[61957]: DEBUG oslo_vmware.api [None req-da68b092-4602-4484-b1b0-adcac4aae9f8 tempest-InstanceActionsV221TestJSON-1768084888 tempest-InstanceActionsV221TestJSON-1768084888-project-member] Task: {'id': task-1277394, 'name': Rename_Task} progress is 5%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.965816] env[61957]: DEBUG oslo_concurrency.lockutils [None req-3c3df299-fd1f-460a-8cc2-979af1db30d1 tempest-ServersTestJSON-1093633677 tempest-ServersTestJSON-1093633677-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 811.984454] env[61957]: DEBUG nova.compute.utils [None req-182a8e01-a1e1-4ecf-8692-2abcfc9944ba tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Using /dev/sd instead of None {{(pid=61957) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 811.984454] env[61957]: DEBUG nova.compute.manager [None req-182a8e01-a1e1-4ecf-8692-2abcfc9944ba tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] [instance: e0249936-d616-4ffb-8f77-d8107633c42a] Allocating IP information in the background. {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 811.984454] env[61957]: DEBUG nova.network.neutron [None req-182a8e01-a1e1-4ecf-8692-2abcfc9944ba tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] [instance: e0249936-d616-4ffb-8f77-d8107633c42a] allocate_for_instance() {{(pid=61957) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 812.062308] env[61957]: DEBUG nova.policy [None req-182a8e01-a1e1-4ecf-8692-2abcfc9944ba tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9056f2767d99453f9f4d7b8e5eeb31be', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4ecb1795cf894c7eab36d8fc20024e98', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61957) authorize /opt/stack/nova/nova/policy.py:203}} [ 812.264359] env[61957]: DEBUG oslo_vmware.api [None req-da68b092-4602-4484-b1b0-adcac4aae9f8 tempest-InstanceActionsV221TestJSON-1768084888 tempest-InstanceActionsV221TestJSON-1768084888-project-member] Task: {'id': task-1277394, 'name': Rename_Task, 'duration_secs': 0.137917} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.264793] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-da68b092-4602-4484-b1b0-adcac4aae9f8 tempest-InstanceActionsV221TestJSON-1768084888 tempest-InstanceActionsV221TestJSON-1768084888-project-member] [instance: 6e7bd89c-2c2a-450a-9858-3526d96c28ab] Powering on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 812.265146] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-17cf0988-7b16-4524-ad0d-f07bd418de1e {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.272223] env[61957]: DEBUG oslo_vmware.api [None req-da68b092-4602-4484-b1b0-adcac4aae9f8 tempest-InstanceActionsV221TestJSON-1768084888 tempest-InstanceActionsV221TestJSON-1768084888-project-member] Waiting for the task: (returnval){ [ 812.272223] env[61957]: value = "task-1277395" [ 812.272223] env[61957]: _type = "Task" [ 812.272223] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.287116] env[61957]: DEBUG oslo_vmware.api [None req-da68b092-4602-4484-b1b0-adcac4aae9f8 tempest-InstanceActionsV221TestJSON-1768084888 tempest-InstanceActionsV221TestJSON-1768084888-project-member] Task: {'id': task-1277395, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.325214] env[61957]: INFO nova.scheduler.client.report [None req-7860e034-eff8-489c-addf-33cf2776246c tempest-ServerActionsTestOtherA-1610268163 tempest-ServerActionsTestOtherA-1610268163-project-member] Deleted allocations for instance 6372a384-61a8-43e8-8f6c-376e2a735045 [ 812.481392] env[61957]: DEBUG nova.network.neutron [None req-182a8e01-a1e1-4ecf-8692-2abcfc9944ba tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] [instance: e0249936-d616-4ffb-8f77-d8107633c42a] Successfully created port: 3852bc01-a0a4-4b76-9461-9eaf58378500 {{(pid=61957) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 812.489813] env[61957]: DEBUG nova.compute.manager [None req-182a8e01-a1e1-4ecf-8692-2abcfc9944ba tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] [instance: e0249936-d616-4ffb-8f77-d8107633c42a] Start building block device mappings for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 812.786149] env[61957]: DEBUG oslo_vmware.api [None req-da68b092-4602-4484-b1b0-adcac4aae9f8 tempest-InstanceActionsV221TestJSON-1768084888 tempest-InstanceActionsV221TestJSON-1768084888-project-member] Task: {'id': task-1277395, 'name': PowerOnVM_Task} progress is 98%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.855199] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1fb38bf-90be-461f-a8bd-a438ba2712ef {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.855199] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-765aa790-22aa-40fa-a9f0-6ed0fac8a883 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.869744] env[61957]: DEBUG oslo_concurrency.lockutils [None req-7860e034-eff8-489c-addf-33cf2776246c tempest-ServerActionsTestOtherA-1610268163 tempest-ServerActionsTestOtherA-1610268163-project-member] Lock "6372a384-61a8-43e8-8f6c-376e2a735045" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 156.625s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 812.871808] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77448d20-4188-4c6a-9932-bb49c4ffae1e {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.884656] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea44de4b-f2aa-428b-a5fd-8f775912a8c3 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.899171] env[61957]: DEBUG nova.compute.provider_tree [None req-bc6cc00e-6b7d-4787-8382-d5cf71a4537c tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 813.286403] env[61957]: DEBUG oslo_vmware.api [None req-da68b092-4602-4484-b1b0-adcac4aae9f8 tempest-InstanceActionsV221TestJSON-1768084888 tempest-InstanceActionsV221TestJSON-1768084888-project-member] Task: {'id': task-1277395, 'name': PowerOnVM_Task, 'duration_secs': 0.788608} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.287093] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-da68b092-4602-4484-b1b0-adcac4aae9f8 tempest-InstanceActionsV221TestJSON-1768084888 tempest-InstanceActionsV221TestJSON-1768084888-project-member] [instance: 6e7bd89c-2c2a-450a-9858-3526d96c28ab] Powered on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 813.287093] env[61957]: INFO nova.compute.manager [None req-da68b092-4602-4484-b1b0-adcac4aae9f8 tempest-InstanceActionsV221TestJSON-1768084888 tempest-InstanceActionsV221TestJSON-1768084888-project-member] [instance: 6e7bd89c-2c2a-450a-9858-3526d96c28ab] Took 7.91 seconds to spawn the instance on the hypervisor. [ 813.287297] env[61957]: DEBUG nova.compute.manager [None req-da68b092-4602-4484-b1b0-adcac4aae9f8 tempest-InstanceActionsV221TestJSON-1768084888 tempest-InstanceActionsV221TestJSON-1768084888-project-member] [instance: 6e7bd89c-2c2a-450a-9858-3526d96c28ab] Checking state {{(pid=61957) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 813.288244] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42d625c1-a745-4db4-9f63-c0dd9c868a6a {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.378444] env[61957]: DEBUG nova.compute.manager [None req-be584a65-2286-47ec-9ff6-29a7835a7d50 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: 37ce46a1-8f9a-4d15-bd81-e40845a0e48a] Starting instance... {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 813.402859] env[61957]: DEBUG nova.scheduler.client.report [None req-bc6cc00e-6b7d-4787-8382-d5cf71a4537c tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 813.506076] env[61957]: DEBUG nova.compute.manager [None req-182a8e01-a1e1-4ecf-8692-2abcfc9944ba tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] [instance: e0249936-d616-4ffb-8f77-d8107633c42a] Start spawning the instance on the hypervisor. {{(pid=61957) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 813.546116] env[61957]: DEBUG nova.virt.hardware [None req-182a8e01-a1e1-4ecf-8692-2abcfc9944ba tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T17:22:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T17:21:50Z,direct_url=,disk_format='vmdk',id=11c76a2c-f705-470a-ba9d-4657858bab38,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='81d3096d0f094373913fc3dc8f5c3c6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T17:21:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 813.546116] env[61957]: DEBUG nova.virt.hardware [None req-182a8e01-a1e1-4ecf-8692-2abcfc9944ba tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Flavor limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 813.546116] env[61957]: DEBUG nova.virt.hardware [None req-182a8e01-a1e1-4ecf-8692-2abcfc9944ba tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Image limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 813.546253] env[61957]: DEBUG nova.virt.hardware [None req-182a8e01-a1e1-4ecf-8692-2abcfc9944ba tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Flavor pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 813.546253] env[61957]: DEBUG nova.virt.hardware [None req-182a8e01-a1e1-4ecf-8692-2abcfc9944ba tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Image pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 813.546253] env[61957]: DEBUG nova.virt.hardware [None req-182a8e01-a1e1-4ecf-8692-2abcfc9944ba tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 813.546253] env[61957]: DEBUG nova.virt.hardware [None req-182a8e01-a1e1-4ecf-8692-2abcfc9944ba tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 813.546253] env[61957]: DEBUG nova.virt.hardware [None req-182a8e01-a1e1-4ecf-8692-2abcfc9944ba tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 813.546952] env[61957]: DEBUG nova.virt.hardware [None req-182a8e01-a1e1-4ecf-8692-2abcfc9944ba tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Got 1 possible topologies {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 813.547451] env[61957]: DEBUG nova.virt.hardware [None req-182a8e01-a1e1-4ecf-8692-2abcfc9944ba tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 813.549164] env[61957]: DEBUG nova.virt.hardware [None req-182a8e01-a1e1-4ecf-8692-2abcfc9944ba tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 813.550333] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47fb71ec-c2f4-47a9-afde-22e2ae3dc93f {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.561871] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4aceb9eb-c439-40b7-b379-0a096ea6374e {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.813647] env[61957]: INFO nova.compute.manager [None req-da68b092-4602-4484-b1b0-adcac4aae9f8 tempest-InstanceActionsV221TestJSON-1768084888 tempest-InstanceActionsV221TestJSON-1768084888-project-member] [instance: 6e7bd89c-2c2a-450a-9858-3526d96c28ab] Took 39.45 seconds to build instance. [ 813.905667] env[61957]: DEBUG oslo_concurrency.lockutils [None req-be584a65-2286-47ec-9ff6-29a7835a7d50 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 813.907772] env[61957]: DEBUG oslo_concurrency.lockutils [None req-bc6cc00e-6b7d-4787-8382-d5cf71a4537c tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.429s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 813.908374] env[61957]: DEBUG nova.compute.manager [None req-bc6cc00e-6b7d-4787-8382-d5cf71a4537c tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] [instance: 615dbbed-2b02-4351-9e03-8c13f424a133] Start building networks asynchronously for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 813.911519] env[61957]: DEBUG oslo_concurrency.lockutils [None req-dce171c8-b166-439d-a4ea-36b63ebfc3eb tempest-ServerGroupTestJSON-950728618 tempest-ServerGroupTestJSON-950728618-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.293s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 813.914886] env[61957]: INFO nova.compute.claims [None req-dce171c8-b166-439d-a4ea-36b63ebfc3eb tempest-ServerGroupTestJSON-950728618 tempest-ServerGroupTestJSON-950728618-project-member] [instance: e9ac118f-08b3-430b-848c-461c2b2e3e02] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 814.209561] env[61957]: DEBUG oslo_concurrency.lockutils [None req-85c79741-463e-4bc0-9ed2-4682b88fcc00 tempest-ServersV294TestFqdnHostnames-1426486100 tempest-ServersV294TestFqdnHostnames-1426486100-project-member] Acquiring lock "2e4a5344-600f-4b61-826e-c15f96b50af2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 814.210393] env[61957]: DEBUG oslo_concurrency.lockutils [None req-85c79741-463e-4bc0-9ed2-4682b88fcc00 tempest-ServersV294TestFqdnHostnames-1426486100 tempest-ServersV294TestFqdnHostnames-1426486100-project-member] Lock "2e4a5344-600f-4b61-826e-c15f96b50af2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 814.244159] env[61957]: DEBUG nova.compute.manager [req-debd4df9-f735-407f-b273-1e6a80ba55dc req-17c4b305-72c8-4c74-ab33-0e0cc524d18b service nova] [instance: e0249936-d616-4ffb-8f77-d8107633c42a] Received event network-vif-plugged-3852bc01-a0a4-4b76-9461-9eaf58378500 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 814.244384] env[61957]: DEBUG oslo_concurrency.lockutils [req-debd4df9-f735-407f-b273-1e6a80ba55dc req-17c4b305-72c8-4c74-ab33-0e0cc524d18b service nova] Acquiring lock "e0249936-d616-4ffb-8f77-d8107633c42a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 814.244583] env[61957]: DEBUG oslo_concurrency.lockutils [req-debd4df9-f735-407f-b273-1e6a80ba55dc req-17c4b305-72c8-4c74-ab33-0e0cc524d18b service nova] Lock "e0249936-d616-4ffb-8f77-d8107633c42a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 814.244761] env[61957]: DEBUG oslo_concurrency.lockutils [req-debd4df9-f735-407f-b273-1e6a80ba55dc req-17c4b305-72c8-4c74-ab33-0e0cc524d18b service nova] Lock "e0249936-d616-4ffb-8f77-d8107633c42a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 814.244902] env[61957]: DEBUG nova.compute.manager [req-debd4df9-f735-407f-b273-1e6a80ba55dc req-17c4b305-72c8-4c74-ab33-0e0cc524d18b service nova] [instance: e0249936-d616-4ffb-8f77-d8107633c42a] No waiting events found dispatching network-vif-plugged-3852bc01-a0a4-4b76-9461-9eaf58378500 {{(pid=61957) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 814.245081] env[61957]: WARNING nova.compute.manager [req-debd4df9-f735-407f-b273-1e6a80ba55dc req-17c4b305-72c8-4c74-ab33-0e0cc524d18b service nova] [instance: e0249936-d616-4ffb-8f77-d8107633c42a] Received unexpected event network-vif-plugged-3852bc01-a0a4-4b76-9461-9eaf58378500 for instance with vm_state building and task_state spawning. [ 814.316491] env[61957]: DEBUG oslo_concurrency.lockutils [None req-da68b092-4602-4484-b1b0-adcac4aae9f8 tempest-InstanceActionsV221TestJSON-1768084888 tempest-InstanceActionsV221TestJSON-1768084888-project-member] Lock "6e7bd89c-2c2a-450a-9858-3526d96c28ab" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 149.176s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 814.424062] env[61957]: DEBUG nova.compute.utils [None req-bc6cc00e-6b7d-4787-8382-d5cf71a4537c tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] Using /dev/sd instead of None {{(pid=61957) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 814.424062] env[61957]: DEBUG nova.compute.manager [None req-bc6cc00e-6b7d-4787-8382-d5cf71a4537c tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] [instance: 615dbbed-2b02-4351-9e03-8c13f424a133] Allocating IP information in the background. {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 814.424062] env[61957]: DEBUG nova.network.neutron [None req-bc6cc00e-6b7d-4787-8382-d5cf71a4537c tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] [instance: 615dbbed-2b02-4351-9e03-8c13f424a133] allocate_for_instance() {{(pid=61957) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 814.554628] env[61957]: DEBUG nova.policy [None req-bc6cc00e-6b7d-4787-8382-d5cf71a4537c tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2414df50921145e98bb4e5cade9c724b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8363e21fa47b46b8b2538b741564cb60', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61957) authorize /opt/stack/nova/nova/policy.py:203}} [ 814.579497] env[61957]: DEBUG nova.network.neutron [None req-182a8e01-a1e1-4ecf-8692-2abcfc9944ba tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] [instance: e0249936-d616-4ffb-8f77-d8107633c42a] Successfully updated port: 3852bc01-a0a4-4b76-9461-9eaf58378500 {{(pid=61957) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 814.707901] env[61957]: DEBUG oslo_concurrency.lockutils [None req-a20e46ee-0368-4600-9ced-5e54f65fe0cc tempest-InstanceActionsV221TestJSON-1768084888 tempest-InstanceActionsV221TestJSON-1768084888-project-member] Acquiring lock "6e7bd89c-2c2a-450a-9858-3526d96c28ab" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 814.708643] env[61957]: DEBUG oslo_concurrency.lockutils [None req-a20e46ee-0368-4600-9ced-5e54f65fe0cc tempest-InstanceActionsV221TestJSON-1768084888 tempest-InstanceActionsV221TestJSON-1768084888-project-member] Lock "6e7bd89c-2c2a-450a-9858-3526d96c28ab" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 814.708643] env[61957]: DEBUG oslo_concurrency.lockutils [None req-a20e46ee-0368-4600-9ced-5e54f65fe0cc tempest-InstanceActionsV221TestJSON-1768084888 tempest-InstanceActionsV221TestJSON-1768084888-project-member] Acquiring lock "6e7bd89c-2c2a-450a-9858-3526d96c28ab-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 814.708643] env[61957]: DEBUG oslo_concurrency.lockutils [None req-a20e46ee-0368-4600-9ced-5e54f65fe0cc tempest-InstanceActionsV221TestJSON-1768084888 tempest-InstanceActionsV221TestJSON-1768084888-project-member] Lock "6e7bd89c-2c2a-450a-9858-3526d96c28ab-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 814.708829] env[61957]: DEBUG oslo_concurrency.lockutils [None req-a20e46ee-0368-4600-9ced-5e54f65fe0cc tempest-InstanceActionsV221TestJSON-1768084888 tempest-InstanceActionsV221TestJSON-1768084888-project-member] Lock "6e7bd89c-2c2a-450a-9858-3526d96c28ab-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 814.711680] env[61957]: INFO nova.compute.manager [None req-a20e46ee-0368-4600-9ced-5e54f65fe0cc tempest-InstanceActionsV221TestJSON-1768084888 tempest-InstanceActionsV221TestJSON-1768084888-project-member] [instance: 6e7bd89c-2c2a-450a-9858-3526d96c28ab] Terminating instance [ 814.713671] env[61957]: DEBUG nova.compute.manager [None req-a20e46ee-0368-4600-9ced-5e54f65fe0cc tempest-InstanceActionsV221TestJSON-1768084888 tempest-InstanceActionsV221TestJSON-1768084888-project-member] [instance: 6e7bd89c-2c2a-450a-9858-3526d96c28ab] Start destroying the instance on the hypervisor. {{(pid=61957) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 814.715406] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-a20e46ee-0368-4600-9ced-5e54f65fe0cc tempest-InstanceActionsV221TestJSON-1768084888 tempest-InstanceActionsV221TestJSON-1768084888-project-member] [instance: 6e7bd89c-2c2a-450a-9858-3526d96c28ab] Destroying instance {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 814.715406] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64abafc1-aae4-41a7-96a9-66ff988eea58 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.724880] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-a20e46ee-0368-4600-9ced-5e54f65fe0cc tempest-InstanceActionsV221TestJSON-1768084888 tempest-InstanceActionsV221TestJSON-1768084888-project-member] [instance: 6e7bd89c-2c2a-450a-9858-3526d96c28ab] Powering off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 814.725154] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3233f567-4d2d-4f11-bf79-1de59d93b414 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.731483] env[61957]: DEBUG oslo_vmware.api [None req-a20e46ee-0368-4600-9ced-5e54f65fe0cc tempest-InstanceActionsV221TestJSON-1768084888 tempest-InstanceActionsV221TestJSON-1768084888-project-member] Waiting for the task: (returnval){ [ 814.731483] env[61957]: value = "task-1277396" [ 814.731483] env[61957]: _type = "Task" [ 814.731483] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 814.739918] env[61957]: DEBUG oslo_vmware.api [None req-a20e46ee-0368-4600-9ced-5e54f65fe0cc tempest-InstanceActionsV221TestJSON-1768084888 tempest-InstanceActionsV221TestJSON-1768084888-project-member] Task: {'id': task-1277396, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.823183] env[61957]: DEBUG nova.compute.manager [None req-227f29c0-ef6f-447c-b621-3f65bc420b4e tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] [instance: 37a9e1c1-9f6c-4047-9ff5-e141d0ca383e] Starting instance... {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 814.929095] env[61957]: DEBUG nova.compute.manager [None req-bc6cc00e-6b7d-4787-8382-d5cf71a4537c tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] [instance: 615dbbed-2b02-4351-9e03-8c13f424a133] Start building block device mappings for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 815.088853] env[61957]: DEBUG oslo_concurrency.lockutils [None req-182a8e01-a1e1-4ecf-8692-2abcfc9944ba tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Acquiring lock "refresh_cache-e0249936-d616-4ffb-8f77-d8107633c42a" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 815.088973] env[61957]: DEBUG oslo_concurrency.lockutils [None req-182a8e01-a1e1-4ecf-8692-2abcfc9944ba tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Acquired lock "refresh_cache-e0249936-d616-4ffb-8f77-d8107633c42a" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 815.089138] env[61957]: DEBUG nova.network.neutron [None req-182a8e01-a1e1-4ecf-8692-2abcfc9944ba tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] [instance: e0249936-d616-4ffb-8f77-d8107633c42a] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 815.242919] env[61957]: DEBUG oslo_vmware.api [None req-a20e46ee-0368-4600-9ced-5e54f65fe0cc tempest-InstanceActionsV221TestJSON-1768084888 tempest-InstanceActionsV221TestJSON-1768084888-project-member] Task: {'id': task-1277396, 'name': PowerOffVM_Task, 'duration_secs': 0.272} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.246667] env[61957]: DEBUG nova.network.neutron [None req-bc6cc00e-6b7d-4787-8382-d5cf71a4537c tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] [instance: 615dbbed-2b02-4351-9e03-8c13f424a133] Successfully created port: 31850256-5668-4fd1-9758-bf86c83c33e4 {{(pid=61957) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 815.248862] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-a20e46ee-0368-4600-9ced-5e54f65fe0cc tempest-InstanceActionsV221TestJSON-1768084888 tempest-InstanceActionsV221TestJSON-1768084888-project-member] [instance: 6e7bd89c-2c2a-450a-9858-3526d96c28ab] Powered off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 815.250957] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-a20e46ee-0368-4600-9ced-5e54f65fe0cc tempest-InstanceActionsV221TestJSON-1768084888 tempest-InstanceActionsV221TestJSON-1768084888-project-member] [instance: 6e7bd89c-2c2a-450a-9858-3526d96c28ab] Unregistering the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 815.250957] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-59071d5c-bd77-4388-bcfa-98d06e9fa14b {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.282184] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d48385c-8012-4894-87f8-de9cc36351b2 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.289942] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-978c1ffd-36c9-4359-801c-bdf594d1ae6b {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.325044] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b54c1130-f923-4b32-a25a-ca0f2fe201ab {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.327402] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-a20e46ee-0368-4600-9ced-5e54f65fe0cc tempest-InstanceActionsV221TestJSON-1768084888 tempest-InstanceActionsV221TestJSON-1768084888-project-member] [instance: 6e7bd89c-2c2a-450a-9858-3526d96c28ab] Unregistered the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 815.327656] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-a20e46ee-0368-4600-9ced-5e54f65fe0cc tempest-InstanceActionsV221TestJSON-1768084888 tempest-InstanceActionsV221TestJSON-1768084888-project-member] [instance: 6e7bd89c-2c2a-450a-9858-3526d96c28ab] Deleting contents of the VM from datastore datastore2 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 815.327870] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-a20e46ee-0368-4600-9ced-5e54f65fe0cc tempest-InstanceActionsV221TestJSON-1768084888 tempest-InstanceActionsV221TestJSON-1768084888-project-member] Deleting the datastore file [datastore2] 6e7bd89c-2c2a-450a-9858-3526d96c28ab {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 815.328153] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0981e6f4-f351-437e-9307-1b77bbf5fa80 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.336174] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7dac71cd-6c00-4f2a-8b1a-cfed9641cee5 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.343408] env[61957]: DEBUG oslo_vmware.api [None req-a20e46ee-0368-4600-9ced-5e54f65fe0cc tempest-InstanceActionsV221TestJSON-1768084888 tempest-InstanceActionsV221TestJSON-1768084888-project-member] Waiting for the task: (returnval){ [ 815.343408] env[61957]: value = "task-1277398" [ 815.343408] env[61957]: _type = "Task" [ 815.343408] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.355025] env[61957]: DEBUG nova.compute.provider_tree [None req-dce171c8-b166-439d-a4ea-36b63ebfc3eb tempest-ServerGroupTestJSON-950728618 tempest-ServerGroupTestJSON-950728618-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 815.357591] env[61957]: DEBUG oslo_concurrency.lockutils [None req-227f29c0-ef6f-447c-b621-3f65bc420b4e tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 815.362536] env[61957]: DEBUG oslo_vmware.api [None req-a20e46ee-0368-4600-9ced-5e54f65fe0cc tempest-InstanceActionsV221TestJSON-1768084888 tempest-InstanceActionsV221TestJSON-1768084888-project-member] Task: {'id': task-1277398, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.542092] env[61957]: DEBUG nova.network.neutron [None req-bc6cc00e-6b7d-4787-8382-d5cf71a4537c tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] [instance: 615dbbed-2b02-4351-9e03-8c13f424a133] Successfully created port: bbe103a8-3b31-4432-a8d8-34647bec0527 {{(pid=61957) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 815.631062] env[61957]: DEBUG nova.network.neutron [None req-182a8e01-a1e1-4ecf-8692-2abcfc9944ba tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] [instance: e0249936-d616-4ffb-8f77-d8107633c42a] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 815.854858] env[61957]: DEBUG oslo_vmware.api [None req-a20e46ee-0368-4600-9ced-5e54f65fe0cc tempest-InstanceActionsV221TestJSON-1768084888 tempest-InstanceActionsV221TestJSON-1768084888-project-member] Task: {'id': task-1277398, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.14397} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.855657] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-a20e46ee-0368-4600-9ced-5e54f65fe0cc tempest-InstanceActionsV221TestJSON-1768084888 tempest-InstanceActionsV221TestJSON-1768084888-project-member] Deleted the datastore file {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 815.855657] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-a20e46ee-0368-4600-9ced-5e54f65fe0cc tempest-InstanceActionsV221TestJSON-1768084888 tempest-InstanceActionsV221TestJSON-1768084888-project-member] [instance: 6e7bd89c-2c2a-450a-9858-3526d96c28ab] Deleted contents of the VM from datastore datastore2 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 815.855960] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-a20e46ee-0368-4600-9ced-5e54f65fe0cc tempest-InstanceActionsV221TestJSON-1768084888 tempest-InstanceActionsV221TestJSON-1768084888-project-member] [instance: 6e7bd89c-2c2a-450a-9858-3526d96c28ab] Instance destroyed {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 815.856226] env[61957]: INFO nova.compute.manager [None req-a20e46ee-0368-4600-9ced-5e54f65fe0cc tempest-InstanceActionsV221TestJSON-1768084888 tempest-InstanceActionsV221TestJSON-1768084888-project-member] [instance: 6e7bd89c-2c2a-450a-9858-3526d96c28ab] Took 1.14 seconds to destroy the instance on the hypervisor. [ 815.856999] env[61957]: DEBUG oslo.service.loopingcall [None req-a20e46ee-0368-4600-9ced-5e54f65fe0cc tempest-InstanceActionsV221TestJSON-1768084888 tempest-InstanceActionsV221TestJSON-1768084888-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 815.856999] env[61957]: DEBUG nova.compute.manager [-] [instance: 6e7bd89c-2c2a-450a-9858-3526d96c28ab] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 815.856999] env[61957]: DEBUG nova.network.neutron [-] [instance: 6e7bd89c-2c2a-450a-9858-3526d96c28ab] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 815.859507] env[61957]: DEBUG nova.scheduler.client.report [None req-dce171c8-b166-439d-a4ea-36b63ebfc3eb tempest-ServerGroupTestJSON-950728618 tempest-ServerGroupTestJSON-950728618-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 815.891917] env[61957]: DEBUG nova.network.neutron [None req-182a8e01-a1e1-4ecf-8692-2abcfc9944ba tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] [instance: e0249936-d616-4ffb-8f77-d8107633c42a] Updating instance_info_cache with network_info: [{"id": "3852bc01-a0a4-4b76-9461-9eaf58378500", "address": "fa:16:3e:c7:5b:78", "network": {"id": "2d30f2d8-cbca-4889-ab21-e4670539418c", "bridge": "br-int", "label": "tempest-ImagesOneServerTestJSON-13606442-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4ecb1795cf894c7eab36d8fc20024e98", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dd72ef32-a57c-43b0-93df-e8a030987d44", "external-id": "nsx-vlan-transportzone-340", "segmentation_id": 340, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3852bc01-a0", "ovs_interfaceid": "3852bc01-a0a4-4b76-9461-9eaf58378500", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 815.940791] env[61957]: DEBUG nova.compute.manager [None req-bc6cc00e-6b7d-4787-8382-d5cf71a4537c tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] [instance: 615dbbed-2b02-4351-9e03-8c13f424a133] Start spawning the instance on the hypervisor. {{(pid=61957) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 815.970984] env[61957]: DEBUG nova.virt.hardware [None req-bc6cc00e-6b7d-4787-8382-d5cf71a4537c tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T17:22:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T17:21:50Z,direct_url=,disk_format='vmdk',id=11c76a2c-f705-470a-ba9d-4657858bab38,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='81d3096d0f094373913fc3dc8f5c3c6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T17:21:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 815.971260] env[61957]: DEBUG nova.virt.hardware [None req-bc6cc00e-6b7d-4787-8382-d5cf71a4537c tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] Flavor limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 815.971416] env[61957]: DEBUG nova.virt.hardware [None req-bc6cc00e-6b7d-4787-8382-d5cf71a4537c tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] Image limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 815.971593] env[61957]: DEBUG nova.virt.hardware [None req-bc6cc00e-6b7d-4787-8382-d5cf71a4537c tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] Flavor pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 815.971740] env[61957]: DEBUG nova.virt.hardware [None req-bc6cc00e-6b7d-4787-8382-d5cf71a4537c tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] Image pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 815.971884] env[61957]: DEBUG nova.virt.hardware [None req-bc6cc00e-6b7d-4787-8382-d5cf71a4537c tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 815.972295] env[61957]: DEBUG nova.virt.hardware [None req-bc6cc00e-6b7d-4787-8382-d5cf71a4537c tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 815.972608] env[61957]: DEBUG nova.virt.hardware [None req-bc6cc00e-6b7d-4787-8382-d5cf71a4537c tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 815.972890] env[61957]: DEBUG nova.virt.hardware [None req-bc6cc00e-6b7d-4787-8382-d5cf71a4537c tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] Got 1 possible topologies {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 815.973106] env[61957]: DEBUG nova.virt.hardware [None req-bc6cc00e-6b7d-4787-8382-d5cf71a4537c tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 815.973290] env[61957]: DEBUG nova.virt.hardware [None req-bc6cc00e-6b7d-4787-8382-d5cf71a4537c tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 815.974512] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94a88894-b988-4353-a930-8469b6dd9e92 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.985126] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d19c00e6-6c69-4874-a8c1-ff7f7cc7f8fb {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.365266] env[61957]: DEBUG oslo_concurrency.lockutils [None req-dce171c8-b166-439d-a4ea-36b63ebfc3eb tempest-ServerGroupTestJSON-950728618 tempest-ServerGroupTestJSON-950728618-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.454s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 816.366798] env[61957]: DEBUG oslo_concurrency.lockutils [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 17.698s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 816.366989] env[61957]: DEBUG oslo_concurrency.lockutils [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 816.367171] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61957) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 816.367466] env[61957]: DEBUG oslo_concurrency.lockutils [None req-36e797a2-06a5-481b-9dd9-66f4ec9dadf3 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.969s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 816.369411] env[61957]: INFO nova.compute.claims [None req-36e797a2-06a5-481b-9dd9-66f4ec9dadf3 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 14ba830d-4e2c-4e9a-a059-3c86209f0127] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 816.373723] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c413207-6138-40e1-867c-a312ac428181 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.386512] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85eb3c59-4f0a-4311-b0a3-bfb8514ca3c3 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.395934] env[61957]: DEBUG oslo_concurrency.lockutils [None req-182a8e01-a1e1-4ecf-8692-2abcfc9944ba tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Releasing lock "refresh_cache-e0249936-d616-4ffb-8f77-d8107633c42a" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 816.395934] env[61957]: DEBUG nova.compute.manager [None req-182a8e01-a1e1-4ecf-8692-2abcfc9944ba tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] [instance: e0249936-d616-4ffb-8f77-d8107633c42a] Instance network_info: |[{"id": "3852bc01-a0a4-4b76-9461-9eaf58378500", "address": "fa:16:3e:c7:5b:78", "network": {"id": "2d30f2d8-cbca-4889-ab21-e4670539418c", "bridge": "br-int", "label": "tempest-ImagesOneServerTestJSON-13606442-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4ecb1795cf894c7eab36d8fc20024e98", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dd72ef32-a57c-43b0-93df-e8a030987d44", "external-id": "nsx-vlan-transportzone-340", "segmentation_id": 340, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3852bc01-a0", "ovs_interfaceid": "3852bc01-a0a4-4b76-9461-9eaf58378500", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 816.408210] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-182a8e01-a1e1-4ecf-8692-2abcfc9944ba tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] [instance: e0249936-d616-4ffb-8f77-d8107633c42a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c7:5b:78', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'dd72ef32-a57c-43b0-93df-e8a030987d44', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3852bc01-a0a4-4b76-9461-9eaf58378500', 'vif_model': 'vmxnet3'}] {{(pid=61957) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 816.417162] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-182a8e01-a1e1-4ecf-8692-2abcfc9944ba tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Creating folder: Project (4ecb1795cf894c7eab36d8fc20024e98). Parent ref: group-v274445. {{(pid=61957) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 816.420083] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8b7afdf4-839b-4745-af64-b0cc93e5bcdb {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.423317] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80678dab-6492-4ce6-ba2a-9a1a33f521c2 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.435597] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a22dc504-e34e-4303-8755-8d53e4a9e980 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.441243] env[61957]: DEBUG nova.compute.manager [req-208f42fc-4cbf-4ffb-9f83-f953ca179778 req-d1e58031-875f-4768-b7ff-9cadb09ead4a service nova] [instance: e0249936-d616-4ffb-8f77-d8107633c42a] Received event network-changed-3852bc01-a0a4-4b76-9461-9eaf58378500 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 816.441944] env[61957]: DEBUG nova.compute.manager [req-208f42fc-4cbf-4ffb-9f83-f953ca179778 req-d1e58031-875f-4768-b7ff-9cadb09ead4a service nova] [instance: e0249936-d616-4ffb-8f77-d8107633c42a] Refreshing instance network info cache due to event network-changed-3852bc01-a0a4-4b76-9461-9eaf58378500. {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 816.441944] env[61957]: DEBUG oslo_concurrency.lockutils [req-208f42fc-4cbf-4ffb-9f83-f953ca179778 req-d1e58031-875f-4768-b7ff-9cadb09ead4a service nova] Acquiring lock "refresh_cache-e0249936-d616-4ffb-8f77-d8107633c42a" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 816.441944] env[61957]: DEBUG oslo_concurrency.lockutils [req-208f42fc-4cbf-4ffb-9f83-f953ca179778 req-d1e58031-875f-4768-b7ff-9cadb09ead4a service nova] Acquired lock "refresh_cache-e0249936-d616-4ffb-8f77-d8107633c42a" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 816.441944] env[61957]: DEBUG nova.network.neutron [req-208f42fc-4cbf-4ffb-9f83-f953ca179778 req-d1e58031-875f-4768-b7ff-9cadb09ead4a service nova] [instance: e0249936-d616-4ffb-8f77-d8107633c42a] Refreshing network info cache for port 3852bc01-a0a4-4b76-9461-9eaf58378500 {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 816.445170] env[61957]: INFO nova.virt.vmwareapi.vm_util [None req-182a8e01-a1e1-4ecf-8692-2abcfc9944ba tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Created folder: Project (4ecb1795cf894c7eab36d8fc20024e98) in parent group-v274445. [ 816.445383] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-182a8e01-a1e1-4ecf-8692-2abcfc9944ba tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Creating folder: Instances. Parent ref: group-v274472. {{(pid=61957) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 816.445812] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2bfca045-de55-45c6-a17c-9e940f6c99f5 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.475912] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181487MB free_disk=142GB free_vcpus=48 pci_devices=None {{(pid=61957) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 816.475912] env[61957]: DEBUG oslo_concurrency.lockutils [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 816.485065] env[61957]: INFO nova.virt.vmwareapi.vm_util [None req-182a8e01-a1e1-4ecf-8692-2abcfc9944ba tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Created folder: Instances in parent group-v274472. [ 816.485339] env[61957]: DEBUG oslo.service.loopingcall [None req-182a8e01-a1e1-4ecf-8692-2abcfc9944ba tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 816.485955] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e0249936-d616-4ffb-8f77-d8107633c42a] Creating VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 816.486234] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7ada0d4e-3d11-45ac-bf35-b12f1388899d {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.515468] env[61957]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 816.515468] env[61957]: value = "task-1277401" [ 816.515468] env[61957]: _type = "Task" [ 816.515468] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.524234] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277401, 'name': CreateVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.772396] env[61957]: DEBUG nova.network.neutron [-] [instance: 6e7bd89c-2c2a-450a-9858-3526d96c28ab] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 816.874383] env[61957]: DEBUG oslo_concurrency.lockutils [None req-dce171c8-b166-439d-a4ea-36b63ebfc3eb tempest-ServerGroupTestJSON-950728618 tempest-ServerGroupTestJSON-950728618-project-member] Acquiring lock "b6ba77be-bdee-486f-8e90-838deb77903c" by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 816.874623] env[61957]: DEBUG oslo_concurrency.lockutils [None req-dce171c8-b166-439d-a4ea-36b63ebfc3eb tempest-ServerGroupTestJSON-950728618 tempest-ServerGroupTestJSON-950728618-project-member] Lock "b6ba77be-bdee-486f-8e90-838deb77903c" acquired by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 817.025666] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277401, 'name': CreateVM_Task, 'duration_secs': 0.290228} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.028187] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e0249936-d616-4ffb-8f77-d8107633c42a] Created VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 817.028898] env[61957]: DEBUG oslo_concurrency.lockutils [None req-182a8e01-a1e1-4ecf-8692-2abcfc9944ba tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 817.029134] env[61957]: DEBUG oslo_concurrency.lockutils [None req-182a8e01-a1e1-4ecf-8692-2abcfc9944ba tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 817.029720] env[61957]: DEBUG oslo_concurrency.lockutils [None req-182a8e01-a1e1-4ecf-8692-2abcfc9944ba tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 817.030147] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0138d3bf-00ec-4214-91cf-ab8d2f0de462 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.035035] env[61957]: DEBUG oslo_vmware.api [None req-182a8e01-a1e1-4ecf-8692-2abcfc9944ba tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Waiting for the task: (returnval){ [ 817.035035] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]524573c8-1fd3-bd0e-f442-23fbda9b0350" [ 817.035035] env[61957]: _type = "Task" [ 817.035035] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.044631] env[61957]: DEBUG oslo_vmware.api [None req-182a8e01-a1e1-4ecf-8692-2abcfc9944ba tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]524573c8-1fd3-bd0e-f442-23fbda9b0350, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.256169] env[61957]: DEBUG nova.network.neutron [req-208f42fc-4cbf-4ffb-9f83-f953ca179778 req-d1e58031-875f-4768-b7ff-9cadb09ead4a service nova] [instance: e0249936-d616-4ffb-8f77-d8107633c42a] Updated VIF entry in instance network info cache for port 3852bc01-a0a4-4b76-9461-9eaf58378500. {{(pid=61957) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 817.256568] env[61957]: DEBUG nova.network.neutron [req-208f42fc-4cbf-4ffb-9f83-f953ca179778 req-d1e58031-875f-4768-b7ff-9cadb09ead4a service nova] [instance: e0249936-d616-4ffb-8f77-d8107633c42a] Updating instance_info_cache with network_info: [{"id": "3852bc01-a0a4-4b76-9461-9eaf58378500", "address": "fa:16:3e:c7:5b:78", "network": {"id": "2d30f2d8-cbca-4889-ab21-e4670539418c", "bridge": "br-int", "label": "tempest-ImagesOneServerTestJSON-13606442-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4ecb1795cf894c7eab36d8fc20024e98", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dd72ef32-a57c-43b0-93df-e8a030987d44", "external-id": "nsx-vlan-transportzone-340", "segmentation_id": 340, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3852bc01-a0", "ovs_interfaceid": "3852bc01-a0a4-4b76-9461-9eaf58378500", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 817.277796] env[61957]: INFO nova.compute.manager [-] [instance: 6e7bd89c-2c2a-450a-9858-3526d96c28ab] Took 1.42 seconds to deallocate network for instance. [ 817.377294] env[61957]: DEBUG oslo_concurrency.lockutils [None req-dce171c8-b166-439d-a4ea-36b63ebfc3eb tempest-ServerGroupTestJSON-950728618 tempest-ServerGroupTestJSON-950728618-project-member] Lock "b6ba77be-bdee-486f-8e90-838deb77903c" "released" by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" :: held 0.503s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 817.377962] env[61957]: DEBUG nova.compute.manager [None req-dce171c8-b166-439d-a4ea-36b63ebfc3eb tempest-ServerGroupTestJSON-950728618 tempest-ServerGroupTestJSON-950728618-project-member] [instance: e9ac118f-08b3-430b-848c-461c2b2e3e02] Start building networks asynchronously for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 817.442445] env[61957]: DEBUG nova.network.neutron [None req-bc6cc00e-6b7d-4787-8382-d5cf71a4537c tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] [instance: 615dbbed-2b02-4351-9e03-8c13f424a133] Successfully updated port: 31850256-5668-4fd1-9758-bf86c83c33e4 {{(pid=61957) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 817.548299] env[61957]: DEBUG oslo_vmware.api [None req-182a8e01-a1e1-4ecf-8692-2abcfc9944ba tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]524573c8-1fd3-bd0e-f442-23fbda9b0350, 'name': SearchDatastore_Task, 'duration_secs': 0.0084} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.551147] env[61957]: DEBUG oslo_concurrency.lockutils [None req-182a8e01-a1e1-4ecf-8692-2abcfc9944ba tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 817.553250] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-182a8e01-a1e1-4ecf-8692-2abcfc9944ba tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] [instance: e0249936-d616-4ffb-8f77-d8107633c42a] Processing image 11c76a2c-f705-470a-ba9d-4657858bab38 {{(pid=61957) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 817.553613] env[61957]: DEBUG oslo_concurrency.lockutils [None req-182a8e01-a1e1-4ecf-8692-2abcfc9944ba tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 817.553946] env[61957]: DEBUG oslo_concurrency.lockutils [None req-182a8e01-a1e1-4ecf-8692-2abcfc9944ba tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 817.554243] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-182a8e01-a1e1-4ecf-8692-2abcfc9944ba tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 817.554849] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c880ade9-418f-44c2-914c-6067a84e1584 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.563237] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-182a8e01-a1e1-4ecf-8692-2abcfc9944ba tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 817.563425] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-182a8e01-a1e1-4ecf-8692-2abcfc9944ba tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61957) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 817.564272] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-918d8adc-1013-4f9c-88e3-7c40a73f95ec {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.571990] env[61957]: DEBUG oslo_vmware.api [None req-182a8e01-a1e1-4ecf-8692-2abcfc9944ba tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Waiting for the task: (returnval){ [ 817.571990] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]5297c28e-9c54-2c33-389b-18d5646f16de" [ 817.571990] env[61957]: _type = "Task" [ 817.571990] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.580748] env[61957]: DEBUG oslo_vmware.api [None req-182a8e01-a1e1-4ecf-8692-2abcfc9944ba tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]5297c28e-9c54-2c33-389b-18d5646f16de, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.686109] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2fc0971-be1d-46b8-9e11-0b6e15ee47d0 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.693824] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1863bd5-3b1e-4890-9ddc-1979fbe08da4 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.726579] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8095a95-325e-4a01-a191-a137b7b9c650 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.734635] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db773918-747e-46a4-9f1b-aa8fa8f497b5 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.747992] env[61957]: DEBUG nova.compute.provider_tree [None req-36e797a2-06a5-481b-9dd9-66f4ec9dadf3 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 817.759749] env[61957]: DEBUG oslo_concurrency.lockutils [req-208f42fc-4cbf-4ffb-9f83-f953ca179778 req-d1e58031-875f-4768-b7ff-9cadb09ead4a service nova] Releasing lock "refresh_cache-e0249936-d616-4ffb-8f77-d8107633c42a" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 817.790011] env[61957]: DEBUG oslo_concurrency.lockutils [None req-a20e46ee-0368-4600-9ced-5e54f65fe0cc tempest-InstanceActionsV221TestJSON-1768084888 tempest-InstanceActionsV221TestJSON-1768084888-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 817.885038] env[61957]: DEBUG nova.compute.utils [None req-dce171c8-b166-439d-a4ea-36b63ebfc3eb tempest-ServerGroupTestJSON-950728618 tempest-ServerGroupTestJSON-950728618-project-member] Using /dev/sd instead of None {{(pid=61957) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 817.886417] env[61957]: DEBUG nova.compute.manager [None req-dce171c8-b166-439d-a4ea-36b63ebfc3eb tempest-ServerGroupTestJSON-950728618 tempest-ServerGroupTestJSON-950728618-project-member] [instance: e9ac118f-08b3-430b-848c-461c2b2e3e02] Allocating IP information in the background. {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 817.889942] env[61957]: DEBUG nova.network.neutron [None req-dce171c8-b166-439d-a4ea-36b63ebfc3eb tempest-ServerGroupTestJSON-950728618 tempest-ServerGroupTestJSON-950728618-project-member] [instance: e9ac118f-08b3-430b-848c-461c2b2e3e02] allocate_for_instance() {{(pid=61957) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 817.932747] env[61957]: DEBUG nova.policy [None req-dce171c8-b166-439d-a4ea-36b63ebfc3eb tempest-ServerGroupTestJSON-950728618 tempest-ServerGroupTestJSON-950728618-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6c5d5d3e07b74f2cbdbc52749450da06', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd5cc9b51da4d4c07b155a481af0aaaf0', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61957) authorize /opt/stack/nova/nova/policy.py:203}} [ 818.090187] env[61957]: DEBUG oslo_concurrency.lockutils [None req-582b6bd7-3488-48cf-a2f8-f44952b10345 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Acquiring lock "e40fa112-4648-428f-a403-b3e3b8319ea8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 818.090711] env[61957]: DEBUG oslo_concurrency.lockutils [None req-582b6bd7-3488-48cf-a2f8-f44952b10345 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Lock "e40fa112-4648-428f-a403-b3e3b8319ea8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 818.091107] env[61957]: DEBUG oslo_vmware.api [None req-182a8e01-a1e1-4ecf-8692-2abcfc9944ba tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]5297c28e-9c54-2c33-389b-18d5646f16de, 'name': SearchDatastore_Task, 'duration_secs': 0.011837} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.092237] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-114c6166-3e80-423c-be3a-e45b851c198a {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.104129] env[61957]: DEBUG oslo_vmware.api [None req-182a8e01-a1e1-4ecf-8692-2abcfc9944ba tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Waiting for the task: (returnval){ [ 818.104129] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]526b6619-261b-761b-527e-2992e8c79a08" [ 818.104129] env[61957]: _type = "Task" [ 818.104129] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.113481] env[61957]: DEBUG oslo_vmware.api [None req-182a8e01-a1e1-4ecf-8692-2abcfc9944ba tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]526b6619-261b-761b-527e-2992e8c79a08, 'name': SearchDatastore_Task, 'duration_secs': 0.00927} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.113851] env[61957]: DEBUG oslo_concurrency.lockutils [None req-182a8e01-a1e1-4ecf-8692-2abcfc9944ba tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 818.114289] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-182a8e01-a1e1-4ecf-8692-2abcfc9944ba tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore2] e0249936-d616-4ffb-8f77-d8107633c42a/e0249936-d616-4ffb-8f77-d8107633c42a.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 818.114670] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8919e99b-6786-4533-8e76-423ff9b3135b {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.123271] env[61957]: DEBUG oslo_vmware.api [None req-182a8e01-a1e1-4ecf-8692-2abcfc9944ba tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Waiting for the task: (returnval){ [ 818.123271] env[61957]: value = "task-1277402" [ 818.123271] env[61957]: _type = "Task" [ 818.123271] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.132599] env[61957]: DEBUG oslo_vmware.api [None req-182a8e01-a1e1-4ecf-8692-2abcfc9944ba tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Task: {'id': task-1277402, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.251723] env[61957]: DEBUG nova.scheduler.client.report [None req-36e797a2-06a5-481b-9dd9-66f4ec9dadf3 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 818.391385] env[61957]: DEBUG nova.compute.manager [None req-dce171c8-b166-439d-a4ea-36b63ebfc3eb tempest-ServerGroupTestJSON-950728618 tempest-ServerGroupTestJSON-950728618-project-member] [instance: e9ac118f-08b3-430b-848c-461c2b2e3e02] Start building block device mappings for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 818.487780] env[61957]: DEBUG nova.network.neutron [None req-dce171c8-b166-439d-a4ea-36b63ebfc3eb tempest-ServerGroupTestJSON-950728618 tempest-ServerGroupTestJSON-950728618-project-member] [instance: e9ac118f-08b3-430b-848c-461c2b2e3e02] Successfully created port: d6407909-c811-4b71-9ac5-f6128992d3f0 {{(pid=61957) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 818.625135] env[61957]: DEBUG nova.compute.manager [req-1849b8ce-0f99-4470-8eb1-8db2f0550df7 req-58f98b69-e226-45ce-8d91-b233f682f381 service nova] [instance: 6e7bd89c-2c2a-450a-9858-3526d96c28ab] Received event network-vif-deleted-54e92ab2-af65-4141-994f-2ff5ddb01f84 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 818.625485] env[61957]: DEBUG nova.compute.manager [req-1849b8ce-0f99-4470-8eb1-8db2f0550df7 req-58f98b69-e226-45ce-8d91-b233f682f381 service nova] [instance: 615dbbed-2b02-4351-9e03-8c13f424a133] Received event network-vif-plugged-31850256-5668-4fd1-9758-bf86c83c33e4 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 818.625554] env[61957]: DEBUG oslo_concurrency.lockutils [req-1849b8ce-0f99-4470-8eb1-8db2f0550df7 req-58f98b69-e226-45ce-8d91-b233f682f381 service nova] Acquiring lock "615dbbed-2b02-4351-9e03-8c13f424a133-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 818.625842] env[61957]: DEBUG oslo_concurrency.lockutils [req-1849b8ce-0f99-4470-8eb1-8db2f0550df7 req-58f98b69-e226-45ce-8d91-b233f682f381 service nova] Lock "615dbbed-2b02-4351-9e03-8c13f424a133-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 818.625934] env[61957]: DEBUG oslo_concurrency.lockutils [req-1849b8ce-0f99-4470-8eb1-8db2f0550df7 req-58f98b69-e226-45ce-8d91-b233f682f381 service nova] Lock "615dbbed-2b02-4351-9e03-8c13f424a133-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 818.626160] env[61957]: DEBUG nova.compute.manager [req-1849b8ce-0f99-4470-8eb1-8db2f0550df7 req-58f98b69-e226-45ce-8d91-b233f682f381 service nova] [instance: 615dbbed-2b02-4351-9e03-8c13f424a133] No waiting events found dispatching network-vif-plugged-31850256-5668-4fd1-9758-bf86c83c33e4 {{(pid=61957) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 818.626258] env[61957]: WARNING nova.compute.manager [req-1849b8ce-0f99-4470-8eb1-8db2f0550df7 req-58f98b69-e226-45ce-8d91-b233f682f381 service nova] [instance: 615dbbed-2b02-4351-9e03-8c13f424a133] Received unexpected event network-vif-plugged-31850256-5668-4fd1-9758-bf86c83c33e4 for instance with vm_state building and task_state spawning. [ 818.626894] env[61957]: DEBUG nova.compute.manager [req-1849b8ce-0f99-4470-8eb1-8db2f0550df7 req-58f98b69-e226-45ce-8d91-b233f682f381 service nova] [instance: 615dbbed-2b02-4351-9e03-8c13f424a133] Received event network-changed-31850256-5668-4fd1-9758-bf86c83c33e4 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 818.626894] env[61957]: DEBUG nova.compute.manager [req-1849b8ce-0f99-4470-8eb1-8db2f0550df7 req-58f98b69-e226-45ce-8d91-b233f682f381 service nova] [instance: 615dbbed-2b02-4351-9e03-8c13f424a133] Refreshing instance network info cache due to event network-changed-31850256-5668-4fd1-9758-bf86c83c33e4. {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 818.626894] env[61957]: DEBUG oslo_concurrency.lockutils [req-1849b8ce-0f99-4470-8eb1-8db2f0550df7 req-58f98b69-e226-45ce-8d91-b233f682f381 service nova] Acquiring lock "refresh_cache-615dbbed-2b02-4351-9e03-8c13f424a133" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 818.626894] env[61957]: DEBUG oslo_concurrency.lockutils [req-1849b8ce-0f99-4470-8eb1-8db2f0550df7 req-58f98b69-e226-45ce-8d91-b233f682f381 service nova] Acquired lock "refresh_cache-615dbbed-2b02-4351-9e03-8c13f424a133" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 818.627049] env[61957]: DEBUG nova.network.neutron [req-1849b8ce-0f99-4470-8eb1-8db2f0550df7 req-58f98b69-e226-45ce-8d91-b233f682f381 service nova] [instance: 615dbbed-2b02-4351-9e03-8c13f424a133] Refreshing network info cache for port 31850256-5668-4fd1-9758-bf86c83c33e4 {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 818.637587] env[61957]: DEBUG oslo_concurrency.lockutils [None req-06bfc4e2-99e2-4510-98f7-f5b198d6231a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Acquiring lock "1a0327c2-5671-4970-9db7-c7cc912d8678" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 818.637825] env[61957]: DEBUG oslo_concurrency.lockutils [None req-06bfc4e2-99e2-4510-98f7-f5b198d6231a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Lock "1a0327c2-5671-4970-9db7-c7cc912d8678" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 818.642997] env[61957]: DEBUG oslo_vmware.api [None req-182a8e01-a1e1-4ecf-8692-2abcfc9944ba tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Task: {'id': task-1277402, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.459325} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.643295] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-182a8e01-a1e1-4ecf-8692-2abcfc9944ba tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore2] e0249936-d616-4ffb-8f77-d8107633c42a/e0249936-d616-4ffb-8f77-d8107633c42a.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 818.643503] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-182a8e01-a1e1-4ecf-8692-2abcfc9944ba tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] [instance: e0249936-d616-4ffb-8f77-d8107633c42a] Extending root virtual disk to 1048576 {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 818.643740] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-47859e0a-5329-46a0-8a74-e97a0c5823ab {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.651787] env[61957]: DEBUG oslo_vmware.api [None req-182a8e01-a1e1-4ecf-8692-2abcfc9944ba tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Waiting for the task: (returnval){ [ 818.651787] env[61957]: value = "task-1277403" [ 818.651787] env[61957]: _type = "Task" [ 818.651787] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.661901] env[61957]: DEBUG oslo_vmware.api [None req-182a8e01-a1e1-4ecf-8692-2abcfc9944ba tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Task: {'id': task-1277403, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.760583] env[61957]: DEBUG oslo_concurrency.lockutils [None req-36e797a2-06a5-481b-9dd9-66f4ec9dadf3 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.393s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 818.761355] env[61957]: DEBUG nova.compute.manager [None req-36e797a2-06a5-481b-9dd9-66f4ec9dadf3 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 14ba830d-4e2c-4e9a-a059-3c86209f0127] Start building networks asynchronously for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 818.764267] env[61957]: DEBUG oslo_concurrency.lockutils [None req-79f56a2c-6af4-4a0f-b13f-0dc131681bf7 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.689s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 818.896042] env[61957]: DEBUG oslo_concurrency.lockutils [None req-bafb915b-a249-4473-bf3a-4c41863ba98f tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Acquiring lock "28ab7d23-487f-4ae1-8fe4-58db55b59918" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 818.896042] env[61957]: DEBUG oslo_concurrency.lockutils [None req-bafb915b-a249-4473-bf3a-4c41863ba98f tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Lock "28ab7d23-487f-4ae1-8fe4-58db55b59918" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 819.167021] env[61957]: DEBUG oslo_vmware.api [None req-182a8e01-a1e1-4ecf-8692-2abcfc9944ba tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Task: {'id': task-1277403, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06743} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.167021] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-182a8e01-a1e1-4ecf-8692-2abcfc9944ba tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] [instance: e0249936-d616-4ffb-8f77-d8107633c42a] Extended root virtual disk {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 819.167021] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be4c9d77-a908-44ab-b563-4c0d239b5e89 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.189352] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-182a8e01-a1e1-4ecf-8692-2abcfc9944ba tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] [instance: e0249936-d616-4ffb-8f77-d8107633c42a] Reconfiguring VM instance instance-00000033 to attach disk [datastore2] e0249936-d616-4ffb-8f77-d8107633c42a/e0249936-d616-4ffb-8f77-d8107633c42a.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 819.190375] env[61957]: DEBUG nova.network.neutron [req-1849b8ce-0f99-4470-8eb1-8db2f0550df7 req-58f98b69-e226-45ce-8d91-b233f682f381 service nova] [instance: 615dbbed-2b02-4351-9e03-8c13f424a133] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 819.193829] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-86de5fb4-0a7b-4d9b-9c5f-8e327f2812ba {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.213771] env[61957]: DEBUG oslo_vmware.api [None req-182a8e01-a1e1-4ecf-8692-2abcfc9944ba tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Waiting for the task: (returnval){ [ 819.213771] env[61957]: value = "task-1277404" [ 819.213771] env[61957]: _type = "Task" [ 819.213771] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.222337] env[61957]: DEBUG oslo_vmware.api [None req-182a8e01-a1e1-4ecf-8692-2abcfc9944ba tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Task: {'id': task-1277404, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.270817] env[61957]: DEBUG nova.compute.utils [None req-36e797a2-06a5-481b-9dd9-66f4ec9dadf3 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Using /dev/sd instead of None {{(pid=61957) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 819.276904] env[61957]: DEBUG nova.compute.manager [None req-36e797a2-06a5-481b-9dd9-66f4ec9dadf3 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 14ba830d-4e2c-4e9a-a059-3c86209f0127] Allocating IP information in the background. {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 819.277243] env[61957]: DEBUG nova.network.neutron [None req-36e797a2-06a5-481b-9dd9-66f4ec9dadf3 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 14ba830d-4e2c-4e9a-a059-3c86209f0127] allocate_for_instance() {{(pid=61957) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 819.354697] env[61957]: DEBUG nova.policy [None req-36e797a2-06a5-481b-9dd9-66f4ec9dadf3 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '85d4f8bb3aa64e7285aac42d14087ff1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e16f6dce3f0e44fb96de516f17d4c6f5', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61957) authorize /opt/stack/nova/nova/policy.py:203}} [ 819.394129] env[61957]: DEBUG nova.network.neutron [req-1849b8ce-0f99-4470-8eb1-8db2f0550df7 req-58f98b69-e226-45ce-8d91-b233f682f381 service nova] [instance: 615dbbed-2b02-4351-9e03-8c13f424a133] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 819.407147] env[61957]: DEBUG nova.compute.manager [None req-dce171c8-b166-439d-a4ea-36b63ebfc3eb tempest-ServerGroupTestJSON-950728618 tempest-ServerGroupTestJSON-950728618-project-member] [instance: e9ac118f-08b3-430b-848c-461c2b2e3e02] Start spawning the instance on the hypervisor. {{(pid=61957) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 819.449451] env[61957]: DEBUG nova.virt.hardware [None req-dce171c8-b166-439d-a4ea-36b63ebfc3eb tempest-ServerGroupTestJSON-950728618 tempest-ServerGroupTestJSON-950728618-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T17:22:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T17:21:50Z,direct_url=,disk_format='vmdk',id=11c76a2c-f705-470a-ba9d-4657858bab38,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='81d3096d0f094373913fc3dc8f5c3c6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T17:21:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 819.449738] env[61957]: DEBUG nova.virt.hardware [None req-dce171c8-b166-439d-a4ea-36b63ebfc3eb tempest-ServerGroupTestJSON-950728618 tempest-ServerGroupTestJSON-950728618-project-member] Flavor limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 819.449901] env[61957]: DEBUG nova.virt.hardware [None req-dce171c8-b166-439d-a4ea-36b63ebfc3eb tempest-ServerGroupTestJSON-950728618 tempest-ServerGroupTestJSON-950728618-project-member] Image limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 819.450097] env[61957]: DEBUG nova.virt.hardware [None req-dce171c8-b166-439d-a4ea-36b63ebfc3eb tempest-ServerGroupTestJSON-950728618 tempest-ServerGroupTestJSON-950728618-project-member] Flavor pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 819.450359] env[61957]: DEBUG nova.virt.hardware [None req-dce171c8-b166-439d-a4ea-36b63ebfc3eb tempest-ServerGroupTestJSON-950728618 tempest-ServerGroupTestJSON-950728618-project-member] Image pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 819.450428] env[61957]: DEBUG nova.virt.hardware [None req-dce171c8-b166-439d-a4ea-36b63ebfc3eb tempest-ServerGroupTestJSON-950728618 tempest-ServerGroupTestJSON-950728618-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 819.450585] env[61957]: DEBUG nova.virt.hardware [None req-dce171c8-b166-439d-a4ea-36b63ebfc3eb tempest-ServerGroupTestJSON-950728618 tempest-ServerGroupTestJSON-950728618-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 819.450838] env[61957]: DEBUG nova.virt.hardware [None req-dce171c8-b166-439d-a4ea-36b63ebfc3eb tempest-ServerGroupTestJSON-950728618 tempest-ServerGroupTestJSON-950728618-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 819.451104] env[61957]: DEBUG nova.virt.hardware [None req-dce171c8-b166-439d-a4ea-36b63ebfc3eb tempest-ServerGroupTestJSON-950728618 tempest-ServerGroupTestJSON-950728618-project-member] Got 1 possible topologies {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 819.451351] env[61957]: DEBUG nova.virt.hardware [None req-dce171c8-b166-439d-a4ea-36b63ebfc3eb tempest-ServerGroupTestJSON-950728618 tempest-ServerGroupTestJSON-950728618-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 819.451537] env[61957]: DEBUG nova.virt.hardware [None req-dce171c8-b166-439d-a4ea-36b63ebfc3eb tempest-ServerGroupTestJSON-950728618 tempest-ServerGroupTestJSON-950728618-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 819.452504] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d3cea64-a891-4c3d-965e-c6eef699418c {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.463273] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ca55c2b-13a8-4a92-abff-9b95aa6f95ef {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.555299] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ca59f931-a2d6-4c9e-954c-a383ffdeea61 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Acquiring lock "d5d10a31-0e78-4ed7-b944-9208138a4861" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 819.555299] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ca59f931-a2d6-4c9e-954c-a383ffdeea61 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Lock "d5d10a31-0e78-4ed7-b944-9208138a4861" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 819.662328] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7317b902-210c-42a1-8785-00f57b4ac359 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.670475] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f54bdc39-3bd7-4cdb-9478-19a0289514d0 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.705216] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-793a8bd6-30e4-4892-a37b-2995c2b1e96c {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.712765] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccf2bb81-401e-48ec-b6ec-17c82c3d89ac {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.728650] env[61957]: DEBUG nova.compute.provider_tree [None req-79f56a2c-6af4-4a0f-b13f-0dc131681bf7 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 819.732784] env[61957]: DEBUG oslo_vmware.api [None req-182a8e01-a1e1-4ecf-8692-2abcfc9944ba tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Task: {'id': task-1277404, 'name': ReconfigVM_Task, 'duration_secs': 0.283709} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.733253] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-182a8e01-a1e1-4ecf-8692-2abcfc9944ba tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] [instance: e0249936-d616-4ffb-8f77-d8107633c42a] Reconfigured VM instance instance-00000033 to attach disk [datastore2] e0249936-d616-4ffb-8f77-d8107633c42a/e0249936-d616-4ffb-8f77-d8107633c42a.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 819.733895] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-715fa02e-0a26-4f5c-9c47-a61f4a5d9167 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.739748] env[61957]: DEBUG oslo_vmware.api [None req-182a8e01-a1e1-4ecf-8692-2abcfc9944ba tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Waiting for the task: (returnval){ [ 819.739748] env[61957]: value = "task-1277405" [ 819.739748] env[61957]: _type = "Task" [ 819.739748] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.748190] env[61957]: DEBUG oslo_vmware.api [None req-182a8e01-a1e1-4ecf-8692-2abcfc9944ba tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Task: {'id': task-1277405, 'name': Rename_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.772821] env[61957]: DEBUG nova.compute.manager [None req-36e797a2-06a5-481b-9dd9-66f4ec9dadf3 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 14ba830d-4e2c-4e9a-a059-3c86209f0127] Start building block device mappings for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 819.780910] env[61957]: DEBUG oslo_concurrency.lockutils [None req-7916de58-8954-4a2d-b3f9-5ec9b9a16797 tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Acquiring lock "0d0f2d34-de35-4e80-8d9f-12693add0786" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 819.782484] env[61957]: DEBUG oslo_concurrency.lockutils [None req-7916de58-8954-4a2d-b3f9-5ec9b9a16797 tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Lock "0d0f2d34-de35-4e80-8d9f-12693add0786" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 819.867824] env[61957]: DEBUG nova.network.neutron [None req-36e797a2-06a5-481b-9dd9-66f4ec9dadf3 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 14ba830d-4e2c-4e9a-a059-3c86209f0127] Successfully created port: 6b5aa7fe-7f6f-4305-b6b9-eef2a2c90d69 {{(pid=61957) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 819.899043] env[61957]: DEBUG oslo_concurrency.lockutils [req-1849b8ce-0f99-4470-8eb1-8db2f0550df7 req-58f98b69-e226-45ce-8d91-b233f682f381 service nova] Releasing lock "refresh_cache-615dbbed-2b02-4351-9e03-8c13f424a133" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 819.999440] env[61957]: DEBUG nova.network.neutron [None req-bc6cc00e-6b7d-4787-8382-d5cf71a4537c tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] [instance: 615dbbed-2b02-4351-9e03-8c13f424a133] Successfully updated port: bbe103a8-3b31-4432-a8d8-34647bec0527 {{(pid=61957) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 820.016284] env[61957]: DEBUG oslo_concurrency.lockutils [None req-accdd3a1-4d57-48f0-a206-1602b908905b tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Acquiring lock "a8ca1ad6-d636-4fa2-b0ac-53b020e392ce" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 820.016515] env[61957]: DEBUG oslo_concurrency.lockutils [None req-accdd3a1-4d57-48f0-a206-1602b908905b tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Lock "a8ca1ad6-d636-4fa2-b0ac-53b020e392ce" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 820.235499] env[61957]: DEBUG nova.scheduler.client.report [None req-79f56a2c-6af4-4a0f-b13f-0dc131681bf7 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 820.252767] env[61957]: DEBUG oslo_vmware.api [None req-182a8e01-a1e1-4ecf-8692-2abcfc9944ba tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Task: {'id': task-1277405, 'name': Rename_Task, 'duration_secs': 0.13652} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.253136] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-182a8e01-a1e1-4ecf-8692-2abcfc9944ba tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] [instance: e0249936-d616-4ffb-8f77-d8107633c42a] Powering on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 820.258408] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-dc435f5b-fc9c-4737-b8bf-778fa8f4be8c {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.267107] env[61957]: DEBUG oslo_vmware.api [None req-182a8e01-a1e1-4ecf-8692-2abcfc9944ba tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Waiting for the task: (returnval){ [ 820.267107] env[61957]: value = "task-1277406" [ 820.267107] env[61957]: _type = "Task" [ 820.267107] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.280552] env[61957]: DEBUG oslo_vmware.api [None req-182a8e01-a1e1-4ecf-8692-2abcfc9944ba tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Task: {'id': task-1277406, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.314989] env[61957]: DEBUG nova.network.neutron [None req-dce171c8-b166-439d-a4ea-36b63ebfc3eb tempest-ServerGroupTestJSON-950728618 tempest-ServerGroupTestJSON-950728618-project-member] [instance: e9ac118f-08b3-430b-848c-461c2b2e3e02] Successfully updated port: d6407909-c811-4b71-9ac5-f6128992d3f0 {{(pid=61957) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 820.503205] env[61957]: DEBUG oslo_concurrency.lockutils [None req-bc6cc00e-6b7d-4787-8382-d5cf71a4537c tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] Acquiring lock "refresh_cache-615dbbed-2b02-4351-9e03-8c13f424a133" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 820.503205] env[61957]: DEBUG oslo_concurrency.lockutils [None req-bc6cc00e-6b7d-4787-8382-d5cf71a4537c tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] Acquired lock "refresh_cache-615dbbed-2b02-4351-9e03-8c13f424a133" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 820.503205] env[61957]: DEBUG nova.network.neutron [None req-bc6cc00e-6b7d-4787-8382-d5cf71a4537c tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] [instance: 615dbbed-2b02-4351-9e03-8c13f424a133] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 820.742689] env[61957]: DEBUG oslo_concurrency.lockutils [None req-79f56a2c-6af4-4a0f-b13f-0dc131681bf7 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.979s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 820.745958] env[61957]: ERROR nova.compute.manager [None req-79f56a2c-6af4-4a0f-b13f-0dc131681bf7 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 0584a64b-4127-4f93-80bc-48deadb5339c, please check neutron logs for more information. [ 820.745958] env[61957]: ERROR nova.compute.manager [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] Traceback (most recent call last): [ 820.745958] env[61957]: ERROR nova.compute.manager [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 820.745958] env[61957]: ERROR nova.compute.manager [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] self.driver.spawn(context, instance, image_meta, [ 820.745958] env[61957]: ERROR nova.compute.manager [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 545, in spawn [ 820.745958] env[61957]: ERROR nova.compute.manager [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] self._vmops.spawn(context, instance, image_meta, injected_files, [ 820.745958] env[61957]: ERROR nova.compute.manager [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 820.745958] env[61957]: ERROR nova.compute.manager [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] vm_ref = self.build_virtual_machine(instance, [ 820.745958] env[61957]: ERROR nova.compute.manager [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 820.745958] env[61957]: ERROR nova.compute.manager [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] vif_infos = vmwarevif.get_vif_info(self._session, [ 820.745958] env[61957]: ERROR nova.compute.manager [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 820.746352] env[61957]: ERROR nova.compute.manager [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] for vif in network_info: [ 820.746352] env[61957]: ERROR nova.compute.manager [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 820.746352] env[61957]: ERROR nova.compute.manager [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] return self._sync_wrapper(fn, *args, **kwargs) [ 820.746352] env[61957]: ERROR nova.compute.manager [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 820.746352] env[61957]: ERROR nova.compute.manager [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] self.wait() [ 820.746352] env[61957]: ERROR nova.compute.manager [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 820.746352] env[61957]: ERROR nova.compute.manager [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] self[:] = self._gt.wait() [ 820.746352] env[61957]: ERROR nova.compute.manager [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 820.746352] env[61957]: ERROR nova.compute.manager [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] return self._exit_event.wait() [ 820.746352] env[61957]: ERROR nova.compute.manager [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 820.746352] env[61957]: ERROR nova.compute.manager [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] current.throw(*self._exc) [ 820.746352] env[61957]: ERROR nova.compute.manager [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 820.746352] env[61957]: ERROR nova.compute.manager [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] result = function(*args, **kwargs) [ 820.746726] env[61957]: ERROR nova.compute.manager [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 820.746726] env[61957]: ERROR nova.compute.manager [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] return func(*args, **kwargs) [ 820.746726] env[61957]: ERROR nova.compute.manager [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 820.746726] env[61957]: ERROR nova.compute.manager [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] raise e [ 820.746726] env[61957]: ERROR nova.compute.manager [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 820.746726] env[61957]: ERROR nova.compute.manager [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] nwinfo = self.network_api.allocate_for_instance( [ 820.746726] env[61957]: ERROR nova.compute.manager [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 820.746726] env[61957]: ERROR nova.compute.manager [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] created_port_ids = self._update_ports_for_instance( [ 820.746726] env[61957]: ERROR nova.compute.manager [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 820.746726] env[61957]: ERROR nova.compute.manager [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] with excutils.save_and_reraise_exception(): [ 820.746726] env[61957]: ERROR nova.compute.manager [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 820.746726] env[61957]: ERROR nova.compute.manager [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] self.force_reraise() [ 820.746726] env[61957]: ERROR nova.compute.manager [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 820.747123] env[61957]: ERROR nova.compute.manager [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] raise self.value [ 820.747123] env[61957]: ERROR nova.compute.manager [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 820.747123] env[61957]: ERROR nova.compute.manager [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] updated_port = self._update_port( [ 820.747123] env[61957]: ERROR nova.compute.manager [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 820.747123] env[61957]: ERROR nova.compute.manager [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] _ensure_no_port_binding_failure(port) [ 820.747123] env[61957]: ERROR nova.compute.manager [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 820.747123] env[61957]: ERROR nova.compute.manager [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] raise exception.PortBindingFailed(port_id=port['id']) [ 820.747123] env[61957]: ERROR nova.compute.manager [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] nova.exception.PortBindingFailed: Binding failed for port 0584a64b-4127-4f93-80bc-48deadb5339c, please check neutron logs for more information. [ 820.747123] env[61957]: ERROR nova.compute.manager [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] [ 820.747123] env[61957]: DEBUG nova.compute.utils [None req-79f56a2c-6af4-4a0f-b13f-0dc131681bf7 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] Binding failed for port 0584a64b-4127-4f93-80bc-48deadb5339c, please check neutron logs for more information. {{(pid=61957) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 820.747941] env[61957]: DEBUG nova.compute.manager [None req-79f56a2c-6af4-4a0f-b13f-0dc131681bf7 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] Build of instance b0af6294-f30c-4266-ae46-6fb03dc0cbeb was re-scheduled: Binding failed for port 0584a64b-4127-4f93-80bc-48deadb5339c, please check neutron logs for more information. {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 820.748532] env[61957]: DEBUG nova.compute.manager [None req-79f56a2c-6af4-4a0f-b13f-0dc131681bf7 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] Unplugging VIFs for instance {{(pid=61957) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 820.752057] env[61957]: DEBUG oslo_concurrency.lockutils [None req-79f56a2c-6af4-4a0f-b13f-0dc131681bf7 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Acquiring lock "refresh_cache-b0af6294-f30c-4266-ae46-6fb03dc0cbeb" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 820.752057] env[61957]: DEBUG oslo_concurrency.lockutils [None req-79f56a2c-6af4-4a0f-b13f-0dc131681bf7 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Acquired lock "refresh_cache-b0af6294-f30c-4266-ae46-6fb03dc0cbeb" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 820.752057] env[61957]: DEBUG nova.network.neutron [None req-79f56a2c-6af4-4a0f-b13f-0dc131681bf7 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 820.752057] env[61957]: DEBUG oslo_concurrency.lockutils [None req-9387ee6a-989e-4086-be03-e20e994ba789 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.003s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 820.752891] env[61957]: INFO nova.compute.claims [None req-9387ee6a-989e-4086-be03-e20e994ba789 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] [instance: 17f00b20-9d3b-45e6-919d-6fab9999ec77] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 820.776815] env[61957]: DEBUG oslo_vmware.api [None req-182a8e01-a1e1-4ecf-8692-2abcfc9944ba tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Task: {'id': task-1277406, 'name': PowerOnVM_Task, 'duration_secs': 0.467913} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.777168] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-182a8e01-a1e1-4ecf-8692-2abcfc9944ba tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] [instance: e0249936-d616-4ffb-8f77-d8107633c42a] Powered on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 820.777389] env[61957]: INFO nova.compute.manager [None req-182a8e01-a1e1-4ecf-8692-2abcfc9944ba tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] [instance: e0249936-d616-4ffb-8f77-d8107633c42a] Took 7.27 seconds to spawn the instance on the hypervisor. [ 820.777593] env[61957]: DEBUG nova.compute.manager [None req-182a8e01-a1e1-4ecf-8692-2abcfc9944ba tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] [instance: e0249936-d616-4ffb-8f77-d8107633c42a] Checking state {{(pid=61957) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 820.778386] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-065d977d-c21e-4d75-8057-aa86f49c082e {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.781575] env[61957]: DEBUG nova.compute.manager [None req-36e797a2-06a5-481b-9dd9-66f4ec9dadf3 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 14ba830d-4e2c-4e9a-a059-3c86209f0127] Start spawning the instance on the hypervisor. {{(pid=61957) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 820.805527] env[61957]: DEBUG nova.compute.manager [req-0e0abacf-48f4-4b02-bc3b-a2a7a794dd62 req-5b6defff-e62f-44aa-a4e2-30a1e643367c service nova] [instance: 615dbbed-2b02-4351-9e03-8c13f424a133] Received event network-vif-plugged-bbe103a8-3b31-4432-a8d8-34647bec0527 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 820.805770] env[61957]: DEBUG oslo_concurrency.lockutils [req-0e0abacf-48f4-4b02-bc3b-a2a7a794dd62 req-5b6defff-e62f-44aa-a4e2-30a1e643367c service nova] Acquiring lock "615dbbed-2b02-4351-9e03-8c13f424a133-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 820.806394] env[61957]: DEBUG oslo_concurrency.lockutils [req-0e0abacf-48f4-4b02-bc3b-a2a7a794dd62 req-5b6defff-e62f-44aa-a4e2-30a1e643367c service nova] Lock "615dbbed-2b02-4351-9e03-8c13f424a133-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 820.806531] env[61957]: DEBUG oslo_concurrency.lockutils [req-0e0abacf-48f4-4b02-bc3b-a2a7a794dd62 req-5b6defff-e62f-44aa-a4e2-30a1e643367c service nova] Lock "615dbbed-2b02-4351-9e03-8c13f424a133-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 820.806816] env[61957]: DEBUG nova.compute.manager [req-0e0abacf-48f4-4b02-bc3b-a2a7a794dd62 req-5b6defff-e62f-44aa-a4e2-30a1e643367c service nova] [instance: 615dbbed-2b02-4351-9e03-8c13f424a133] No waiting events found dispatching network-vif-plugged-bbe103a8-3b31-4432-a8d8-34647bec0527 {{(pid=61957) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 820.807044] env[61957]: WARNING nova.compute.manager [req-0e0abacf-48f4-4b02-bc3b-a2a7a794dd62 req-5b6defff-e62f-44aa-a4e2-30a1e643367c service nova] [instance: 615dbbed-2b02-4351-9e03-8c13f424a133] Received unexpected event network-vif-plugged-bbe103a8-3b31-4432-a8d8-34647bec0527 for instance with vm_state building and task_state spawning. [ 820.807201] env[61957]: DEBUG nova.compute.manager [req-0e0abacf-48f4-4b02-bc3b-a2a7a794dd62 req-5b6defff-e62f-44aa-a4e2-30a1e643367c service nova] [instance: 615dbbed-2b02-4351-9e03-8c13f424a133] Received event network-changed-bbe103a8-3b31-4432-a8d8-34647bec0527 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 820.807355] env[61957]: DEBUG nova.compute.manager [req-0e0abacf-48f4-4b02-bc3b-a2a7a794dd62 req-5b6defff-e62f-44aa-a4e2-30a1e643367c service nova] [instance: 615dbbed-2b02-4351-9e03-8c13f424a133] Refreshing instance network info cache due to event network-changed-bbe103a8-3b31-4432-a8d8-34647bec0527. {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 820.807525] env[61957]: DEBUG oslo_concurrency.lockutils [req-0e0abacf-48f4-4b02-bc3b-a2a7a794dd62 req-5b6defff-e62f-44aa-a4e2-30a1e643367c service nova] Acquiring lock "refresh_cache-615dbbed-2b02-4351-9e03-8c13f424a133" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 820.812556] env[61957]: DEBUG nova.virt.hardware [None req-36e797a2-06a5-481b-9dd9-66f4ec9dadf3 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T17:22:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T17:21:50Z,direct_url=,disk_format='vmdk',id=11c76a2c-f705-470a-ba9d-4657858bab38,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='81d3096d0f094373913fc3dc8f5c3c6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T17:21:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 820.812777] env[61957]: DEBUG nova.virt.hardware [None req-36e797a2-06a5-481b-9dd9-66f4ec9dadf3 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Flavor limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 820.812930] env[61957]: DEBUG nova.virt.hardware [None req-36e797a2-06a5-481b-9dd9-66f4ec9dadf3 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Image limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 820.813124] env[61957]: DEBUG nova.virt.hardware [None req-36e797a2-06a5-481b-9dd9-66f4ec9dadf3 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Flavor pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 820.813268] env[61957]: DEBUG nova.virt.hardware [None req-36e797a2-06a5-481b-9dd9-66f4ec9dadf3 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Image pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 820.813410] env[61957]: DEBUG nova.virt.hardware [None req-36e797a2-06a5-481b-9dd9-66f4ec9dadf3 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 820.813614] env[61957]: DEBUG nova.virt.hardware [None req-36e797a2-06a5-481b-9dd9-66f4ec9dadf3 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 820.813769] env[61957]: DEBUG nova.virt.hardware [None req-36e797a2-06a5-481b-9dd9-66f4ec9dadf3 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 820.813933] env[61957]: DEBUG nova.virt.hardware [None req-36e797a2-06a5-481b-9dd9-66f4ec9dadf3 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Got 1 possible topologies {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 820.814299] env[61957]: DEBUG nova.virt.hardware [None req-36e797a2-06a5-481b-9dd9-66f4ec9dadf3 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 820.814500] env[61957]: DEBUG nova.virt.hardware [None req-36e797a2-06a5-481b-9dd9-66f4ec9dadf3 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 820.815776] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7de8a145-8a54-4c84-a2a4-cc6f4f617dec {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.818625] env[61957]: DEBUG oslo_concurrency.lockutils [None req-dce171c8-b166-439d-a4ea-36b63ebfc3eb tempest-ServerGroupTestJSON-950728618 tempest-ServerGroupTestJSON-950728618-project-member] Acquiring lock "refresh_cache-e9ac118f-08b3-430b-848c-461c2b2e3e02" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 820.818806] env[61957]: DEBUG oslo_concurrency.lockutils [None req-dce171c8-b166-439d-a4ea-36b63ebfc3eb tempest-ServerGroupTestJSON-950728618 tempest-ServerGroupTestJSON-950728618-project-member] Acquired lock "refresh_cache-e9ac118f-08b3-430b-848c-461c2b2e3e02" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 820.818892] env[61957]: DEBUG nova.network.neutron [None req-dce171c8-b166-439d-a4ea-36b63ebfc3eb tempest-ServerGroupTestJSON-950728618 tempest-ServerGroupTestJSON-950728618-project-member] [instance: e9ac118f-08b3-430b-848c-461c2b2e3e02] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 820.827291] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52b3f9d1-4ad4-43ae-b0a4-b8ba2a30611f {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.041196] env[61957]: DEBUG nova.network.neutron [None req-bc6cc00e-6b7d-4787-8382-d5cf71a4537c tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] [instance: 615dbbed-2b02-4351-9e03-8c13f424a133] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 821.277058] env[61957]: DEBUG nova.network.neutron [None req-79f56a2c-6af4-4a0f-b13f-0dc131681bf7 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 821.305193] env[61957]: INFO nova.compute.manager [None req-182a8e01-a1e1-4ecf-8692-2abcfc9944ba tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] [instance: e0249936-d616-4ffb-8f77-d8107633c42a] Took 37.59 seconds to build instance. [ 821.377915] env[61957]: DEBUG nova.network.neutron [None req-bc6cc00e-6b7d-4787-8382-d5cf71a4537c tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] [instance: 615dbbed-2b02-4351-9e03-8c13f424a133] Updating instance_info_cache with network_info: [{"id": "31850256-5668-4fd1-9758-bf86c83c33e4", "address": "fa:16:3e:e0:80:a6", "network": {"id": "20aa8adb-b6a9-42c4-94c0-17673dfad27a", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-865860169", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.116", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8363e21fa47b46b8b2538b741564cb60", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "40859343-2baa-45fd-88e3-ebf8aaed2b19", "external-id": "nsx-vlan-transportzone-10", "segmentation_id": 10, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap31850256-56", "ovs_interfaceid": "31850256-5668-4fd1-9758-bf86c83c33e4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "bbe103a8-3b31-4432-a8d8-34647bec0527", "address": "fa:16:3e:2d:a4:c3", "network": {"id": "48edddaa-5fbf-42a9-a7a9-53ead80056af", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1780374469", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.125", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "8363e21fa47b46b8b2538b741564cb60", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ae4e3171-21cd-4094-b6cf-81bf366c75bd", "external-id": "nsx-vlan-transportzone-193", "segmentation_id": 193, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbbe103a8-3b", "ovs_interfaceid": "bbe103a8-3b31-4432-a8d8-34647bec0527", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 821.398628] env[61957]: DEBUG nova.network.neutron [None req-dce171c8-b166-439d-a4ea-36b63ebfc3eb tempest-ServerGroupTestJSON-950728618 tempest-ServerGroupTestJSON-950728618-project-member] [instance: e9ac118f-08b3-430b-848c-461c2b2e3e02] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 821.467565] env[61957]: DEBUG nova.network.neutron [None req-79f56a2c-6af4-4a0f-b13f-0dc131681bf7 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 821.780193] env[61957]: DEBUG nova.network.neutron [None req-dce171c8-b166-439d-a4ea-36b63ebfc3eb tempest-ServerGroupTestJSON-950728618 tempest-ServerGroupTestJSON-950728618-project-member] [instance: e9ac118f-08b3-430b-848c-461c2b2e3e02] Updating instance_info_cache with network_info: [{"id": "d6407909-c811-4b71-9ac5-f6128992d3f0", "address": "fa:16:3e:e7:ab:79", "network": {"id": "801ead5c-62aa-4304-bfbb-0a7dfaea9be0", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-1943350964-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d5cc9b51da4d4c07b155a481af0aaaf0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "04ccbc7a-cf8d-4ea2-8411-291a1e27df7b", "external-id": "nsx-vlan-transportzone-998", "segmentation_id": 998, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd6407909-c8", "ovs_interfaceid": "d6407909-c811-4b71-9ac5-f6128992d3f0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 821.808140] env[61957]: DEBUG oslo_concurrency.lockutils [None req-182a8e01-a1e1-4ecf-8692-2abcfc9944ba tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Lock "e0249936-d616-4ffb-8f77-d8107633c42a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 151.765s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 821.881814] env[61957]: DEBUG oslo_concurrency.lockutils [None req-bc6cc00e-6b7d-4787-8382-d5cf71a4537c tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] Releasing lock "refresh_cache-615dbbed-2b02-4351-9e03-8c13f424a133" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 821.882184] env[61957]: DEBUG nova.compute.manager [None req-bc6cc00e-6b7d-4787-8382-d5cf71a4537c tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] [instance: 615dbbed-2b02-4351-9e03-8c13f424a133] Instance network_info: |[{"id": "31850256-5668-4fd1-9758-bf86c83c33e4", "address": "fa:16:3e:e0:80:a6", "network": {"id": "20aa8adb-b6a9-42c4-94c0-17673dfad27a", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-865860169", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.116", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8363e21fa47b46b8b2538b741564cb60", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "40859343-2baa-45fd-88e3-ebf8aaed2b19", "external-id": "nsx-vlan-transportzone-10", "segmentation_id": 10, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap31850256-56", "ovs_interfaceid": "31850256-5668-4fd1-9758-bf86c83c33e4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "bbe103a8-3b31-4432-a8d8-34647bec0527", "address": "fa:16:3e:2d:a4:c3", "network": {"id": "48edddaa-5fbf-42a9-a7a9-53ead80056af", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1780374469", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.125", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "8363e21fa47b46b8b2538b741564cb60", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ae4e3171-21cd-4094-b6cf-81bf366c75bd", "external-id": "nsx-vlan-transportzone-193", "segmentation_id": 193, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbbe103a8-3b", "ovs_interfaceid": "bbe103a8-3b31-4432-a8d8-34647bec0527", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 821.882467] env[61957]: DEBUG oslo_concurrency.lockutils [req-0e0abacf-48f4-4b02-bc3b-a2a7a794dd62 req-5b6defff-e62f-44aa-a4e2-30a1e643367c service nova] Acquired lock "refresh_cache-615dbbed-2b02-4351-9e03-8c13f424a133" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 821.882639] env[61957]: DEBUG nova.network.neutron [req-0e0abacf-48f4-4b02-bc3b-a2a7a794dd62 req-5b6defff-e62f-44aa-a4e2-30a1e643367c service nova] [instance: 615dbbed-2b02-4351-9e03-8c13f424a133] Refreshing network info cache for port bbe103a8-3b31-4432-a8d8-34647bec0527 {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 821.883727] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-bc6cc00e-6b7d-4787-8382-d5cf71a4537c tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] [instance: 615dbbed-2b02-4351-9e03-8c13f424a133] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e0:80:a6', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '40859343-2baa-45fd-88e3-ebf8aaed2b19', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '31850256-5668-4fd1-9758-bf86c83c33e4', 'vif_model': 'vmxnet3'}, {'network_name': 'br-int', 'mac_address': 'fa:16:3e:2d:a4:c3', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ae4e3171-21cd-4094-b6cf-81bf366c75bd', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'bbe103a8-3b31-4432-a8d8-34647bec0527', 'vif_model': 'vmxnet3'}] {{(pid=61957) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 821.897729] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-bc6cc00e-6b7d-4787-8382-d5cf71a4537c tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] Creating folder: Project (8363e21fa47b46b8b2538b741564cb60). Parent ref: group-v274445. {{(pid=61957) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 821.904213] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-84ae0fcb-3d35-4fda-a25e-50b50e2afcc5 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.908649] env[61957]: DEBUG nova.network.neutron [None req-36e797a2-06a5-481b-9dd9-66f4ec9dadf3 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 14ba830d-4e2c-4e9a-a059-3c86209f0127] Successfully updated port: 6b5aa7fe-7f6f-4305-b6b9-eef2a2c90d69 {{(pid=61957) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 821.918059] env[61957]: INFO nova.virt.vmwareapi.vm_util [None req-bc6cc00e-6b7d-4787-8382-d5cf71a4537c tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] Created folder: Project (8363e21fa47b46b8b2538b741564cb60) in parent group-v274445. [ 821.918274] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-bc6cc00e-6b7d-4787-8382-d5cf71a4537c tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] Creating folder: Instances. Parent ref: group-v274475. {{(pid=61957) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 821.921286] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1ba06f67-0947-495f-878e-77c5eeb68f45 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.934297] env[61957]: INFO nova.virt.vmwareapi.vm_util [None req-bc6cc00e-6b7d-4787-8382-d5cf71a4537c tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] Created folder: Instances in parent group-v274475. [ 821.934396] env[61957]: DEBUG oslo.service.loopingcall [None req-bc6cc00e-6b7d-4787-8382-d5cf71a4537c tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 821.934683] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 615dbbed-2b02-4351-9e03-8c13f424a133] Creating VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 821.934982] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-af703b02-63d2-4994-b297-f1b78c8f3bac {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.968929] env[61957]: DEBUG oslo_concurrency.lockutils [None req-79f56a2c-6af4-4a0f-b13f-0dc131681bf7 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Releasing lock "refresh_cache-b0af6294-f30c-4266-ae46-6fb03dc0cbeb" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 821.969140] env[61957]: DEBUG nova.compute.manager [None req-79f56a2c-6af4-4a0f-b13f-0dc131681bf7 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61957) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 821.969321] env[61957]: DEBUG nova.compute.manager [None req-79f56a2c-6af4-4a0f-b13f-0dc131681bf7 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 821.969487] env[61957]: DEBUG nova.network.neutron [None req-79f56a2c-6af4-4a0f-b13f-0dc131681bf7 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 821.972495] env[61957]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 821.972495] env[61957]: value = "task-1277409" [ 821.972495] env[61957]: _type = "Task" [ 821.972495] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.980229] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277409, 'name': CreateVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.987916] env[61957]: DEBUG nova.network.neutron [None req-79f56a2c-6af4-4a0f-b13f-0dc131681bf7 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 822.177840] env[61957]: DEBUG nova.network.neutron [req-0e0abacf-48f4-4b02-bc3b-a2a7a794dd62 req-5b6defff-e62f-44aa-a4e2-30a1e643367c service nova] [instance: 615dbbed-2b02-4351-9e03-8c13f424a133] Updated VIF entry in instance network info cache for port bbe103a8-3b31-4432-a8d8-34647bec0527. {{(pid=61957) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 822.178278] env[61957]: DEBUG nova.network.neutron [req-0e0abacf-48f4-4b02-bc3b-a2a7a794dd62 req-5b6defff-e62f-44aa-a4e2-30a1e643367c service nova] [instance: 615dbbed-2b02-4351-9e03-8c13f424a133] Updating instance_info_cache with network_info: [{"id": "31850256-5668-4fd1-9758-bf86c83c33e4", "address": "fa:16:3e:e0:80:a6", "network": {"id": "20aa8adb-b6a9-42c4-94c0-17673dfad27a", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-865860169", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.116", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8363e21fa47b46b8b2538b741564cb60", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "40859343-2baa-45fd-88e3-ebf8aaed2b19", "external-id": "nsx-vlan-transportzone-10", "segmentation_id": 10, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap31850256-56", "ovs_interfaceid": "31850256-5668-4fd1-9758-bf86c83c33e4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "bbe103a8-3b31-4432-a8d8-34647bec0527", "address": "fa:16:3e:2d:a4:c3", "network": {"id": "48edddaa-5fbf-42a9-a7a9-53ead80056af", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1780374469", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.125", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "8363e21fa47b46b8b2538b741564cb60", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ae4e3171-21cd-4094-b6cf-81bf366c75bd", "external-id": "nsx-vlan-transportzone-193", "segmentation_id": 193, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbbe103a8-3b", "ovs_interfaceid": "bbe103a8-3b31-4432-a8d8-34647bec0527", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 822.268875] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de91e792-6e75-4ca0-a9c5-98a5c0a9e2b4 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.276382] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-925051b0-c1b2-46de-b9fc-398406988029 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.307024] env[61957]: DEBUG oslo_concurrency.lockutils [None req-dce171c8-b166-439d-a4ea-36b63ebfc3eb tempest-ServerGroupTestJSON-950728618 tempest-ServerGroupTestJSON-950728618-project-member] Releasing lock "refresh_cache-e9ac118f-08b3-430b-848c-461c2b2e3e02" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 822.307372] env[61957]: DEBUG nova.compute.manager [None req-dce171c8-b166-439d-a4ea-36b63ebfc3eb tempest-ServerGroupTestJSON-950728618 tempest-ServerGroupTestJSON-950728618-project-member] [instance: e9ac118f-08b3-430b-848c-461c2b2e3e02] Instance network_info: |[{"id": "d6407909-c811-4b71-9ac5-f6128992d3f0", "address": "fa:16:3e:e7:ab:79", "network": {"id": "801ead5c-62aa-4304-bfbb-0a7dfaea9be0", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-1943350964-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d5cc9b51da4d4c07b155a481af0aaaf0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "04ccbc7a-cf8d-4ea2-8411-291a1e27df7b", "external-id": "nsx-vlan-transportzone-998", "segmentation_id": 998, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd6407909-c8", "ovs_interfaceid": "d6407909-c811-4b71-9ac5-f6128992d3f0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 822.307986] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-dce171c8-b166-439d-a4ea-36b63ebfc3eb tempest-ServerGroupTestJSON-950728618 tempest-ServerGroupTestJSON-950728618-project-member] [instance: e9ac118f-08b3-430b-848c-461c2b2e3e02] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e7:ab:79', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '04ccbc7a-cf8d-4ea2-8411-291a1e27df7b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd6407909-c811-4b71-9ac5-f6128992d3f0', 'vif_model': 'vmxnet3'}] {{(pid=61957) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 822.315525] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-dce171c8-b166-439d-a4ea-36b63ebfc3eb tempest-ServerGroupTestJSON-950728618 tempest-ServerGroupTestJSON-950728618-project-member] Creating folder: Project (d5cc9b51da4d4c07b155a481af0aaaf0). Parent ref: group-v274445. {{(pid=61957) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 822.316214] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99646004-3453-4cc0-aab5-8b97126aa630 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.318772] env[61957]: DEBUG nova.compute.manager [None req-e4109c08-00a8-4666-8e84-183191cbc2fb tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: 0adae8e1-8c2f-4110-805b-1f286debc833] Starting instance... {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 822.321489] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ecb059d9-300d-47b7-9a4e-e579e178857f {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.327832] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1582e685-ddc8-4370-91f4-e1743e8f05c9 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.332505] env[61957]: INFO nova.virt.vmwareapi.vm_util [None req-dce171c8-b166-439d-a4ea-36b63ebfc3eb tempest-ServerGroupTestJSON-950728618 tempest-ServerGroupTestJSON-950728618-project-member] Created folder: Project (d5cc9b51da4d4c07b155a481af0aaaf0) in parent group-v274445. [ 822.332772] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-dce171c8-b166-439d-a4ea-36b63ebfc3eb tempest-ServerGroupTestJSON-950728618 tempest-ServerGroupTestJSON-950728618-project-member] Creating folder: Instances. Parent ref: group-v274478. {{(pid=61957) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 822.333696] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-568eaf54-f5e5-4979-a932-aae207211f96 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.342587] env[61957]: DEBUG nova.compute.provider_tree [None req-9387ee6a-989e-4086-be03-e20e994ba789 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 822.351111] env[61957]: DEBUG nova.compute.manager [None req-83fd723b-e45f-4910-94da-c61980bb25f7 tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] [instance: e0249936-d616-4ffb-8f77-d8107633c42a] Checking state {{(pid=61957) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 822.351505] env[61957]: INFO nova.virt.vmwareapi.vm_util [None req-dce171c8-b166-439d-a4ea-36b63ebfc3eb tempest-ServerGroupTestJSON-950728618 tempest-ServerGroupTestJSON-950728618-project-member] Created folder: Instances in parent group-v274478. [ 822.351655] env[61957]: DEBUG oslo.service.loopingcall [None req-dce171c8-b166-439d-a4ea-36b63ebfc3eb tempest-ServerGroupTestJSON-950728618 tempest-ServerGroupTestJSON-950728618-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 822.353078] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8f4aa12-5535-45e6-9e24-8f7c196a16c4 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.355799] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e9ac118f-08b3-430b-848c-461c2b2e3e02] Creating VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 822.356203] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c3e9b673-02eb-4a0a-a748-cb826bd90c90 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.378689] env[61957]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 822.378689] env[61957]: value = "task-1277412" [ 822.378689] env[61957]: _type = "Task" [ 822.378689] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.386643] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277412, 'name': CreateVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.412388] env[61957]: DEBUG oslo_concurrency.lockutils [None req-36e797a2-06a5-481b-9dd9-66f4ec9dadf3 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Acquiring lock "refresh_cache-14ba830d-4e2c-4e9a-a059-3c86209f0127" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 822.412388] env[61957]: DEBUG oslo_concurrency.lockutils [None req-36e797a2-06a5-481b-9dd9-66f4ec9dadf3 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Acquired lock "refresh_cache-14ba830d-4e2c-4e9a-a059-3c86209f0127" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 822.412388] env[61957]: DEBUG nova.network.neutron [None req-36e797a2-06a5-481b-9dd9-66f4ec9dadf3 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 14ba830d-4e2c-4e9a-a059-3c86209f0127] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 822.482883] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277409, 'name': CreateVM_Task, 'duration_secs': 0.3447} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.482883] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 615dbbed-2b02-4351-9e03-8c13f424a133] Created VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 822.482883] env[61957]: DEBUG oslo_concurrency.lockutils [None req-bc6cc00e-6b7d-4787-8382-d5cf71a4537c tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 822.483181] env[61957]: DEBUG oslo_concurrency.lockutils [None req-bc6cc00e-6b7d-4787-8382-d5cf71a4537c tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 822.483321] env[61957]: DEBUG oslo_concurrency.lockutils [None req-bc6cc00e-6b7d-4787-8382-d5cf71a4537c tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 822.483569] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-507f5a8c-dab1-473e-a4fb-8017d6bd0e5e {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.488467] env[61957]: DEBUG oslo_vmware.api [None req-bc6cc00e-6b7d-4787-8382-d5cf71a4537c tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] Waiting for the task: (returnval){ [ 822.488467] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52aeb11b-bd23-5d8d-b903-f6f55ca6f243" [ 822.488467] env[61957]: _type = "Task" [ 822.488467] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.492028] env[61957]: DEBUG nova.network.neutron [None req-79f56a2c-6af4-4a0f-b13f-0dc131681bf7 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 822.497209] env[61957]: DEBUG oslo_vmware.api [None req-bc6cc00e-6b7d-4787-8382-d5cf71a4537c tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52aeb11b-bd23-5d8d-b903-f6f55ca6f243, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.682036] env[61957]: DEBUG oslo_concurrency.lockutils [req-0e0abacf-48f4-4b02-bc3b-a2a7a794dd62 req-5b6defff-e62f-44aa-a4e2-30a1e643367c service nova] Releasing lock "refresh_cache-615dbbed-2b02-4351-9e03-8c13f424a133" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 822.682036] env[61957]: DEBUG nova.compute.manager [req-0e0abacf-48f4-4b02-bc3b-a2a7a794dd62 req-5b6defff-e62f-44aa-a4e2-30a1e643367c service nova] [instance: e9ac118f-08b3-430b-848c-461c2b2e3e02] Received event network-vif-plugged-d6407909-c811-4b71-9ac5-f6128992d3f0 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 822.682036] env[61957]: DEBUG oslo_concurrency.lockutils [req-0e0abacf-48f4-4b02-bc3b-a2a7a794dd62 req-5b6defff-e62f-44aa-a4e2-30a1e643367c service nova] Acquiring lock "e9ac118f-08b3-430b-848c-461c2b2e3e02-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 822.682036] env[61957]: DEBUG oslo_concurrency.lockutils [req-0e0abacf-48f4-4b02-bc3b-a2a7a794dd62 req-5b6defff-e62f-44aa-a4e2-30a1e643367c service nova] Lock "e9ac118f-08b3-430b-848c-461c2b2e3e02-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 822.682256] env[61957]: DEBUG oslo_concurrency.lockutils [req-0e0abacf-48f4-4b02-bc3b-a2a7a794dd62 req-5b6defff-e62f-44aa-a4e2-30a1e643367c service nova] Lock "e9ac118f-08b3-430b-848c-461c2b2e3e02-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 822.682480] env[61957]: DEBUG nova.compute.manager [req-0e0abacf-48f4-4b02-bc3b-a2a7a794dd62 req-5b6defff-e62f-44aa-a4e2-30a1e643367c service nova] [instance: e9ac118f-08b3-430b-848c-461c2b2e3e02] No waiting events found dispatching network-vif-plugged-d6407909-c811-4b71-9ac5-f6128992d3f0 {{(pid=61957) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 822.682701] env[61957]: WARNING nova.compute.manager [req-0e0abacf-48f4-4b02-bc3b-a2a7a794dd62 req-5b6defff-e62f-44aa-a4e2-30a1e643367c service nova] [instance: e9ac118f-08b3-430b-848c-461c2b2e3e02] Received unexpected event network-vif-plugged-d6407909-c811-4b71-9ac5-f6128992d3f0 for instance with vm_state building and task_state spawning. [ 822.682878] env[61957]: DEBUG nova.compute.manager [req-0e0abacf-48f4-4b02-bc3b-a2a7a794dd62 req-5b6defff-e62f-44aa-a4e2-30a1e643367c service nova] [instance: e9ac118f-08b3-430b-848c-461c2b2e3e02] Received event network-changed-d6407909-c811-4b71-9ac5-f6128992d3f0 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 822.683045] env[61957]: DEBUG nova.compute.manager [req-0e0abacf-48f4-4b02-bc3b-a2a7a794dd62 req-5b6defff-e62f-44aa-a4e2-30a1e643367c service nova] [instance: e9ac118f-08b3-430b-848c-461c2b2e3e02] Refreshing instance network info cache due to event network-changed-d6407909-c811-4b71-9ac5-f6128992d3f0. {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 822.683235] env[61957]: DEBUG oslo_concurrency.lockutils [req-0e0abacf-48f4-4b02-bc3b-a2a7a794dd62 req-5b6defff-e62f-44aa-a4e2-30a1e643367c service nova] Acquiring lock "refresh_cache-e9ac118f-08b3-430b-848c-461c2b2e3e02" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 822.683370] env[61957]: DEBUG oslo_concurrency.lockutils [req-0e0abacf-48f4-4b02-bc3b-a2a7a794dd62 req-5b6defff-e62f-44aa-a4e2-30a1e643367c service nova] Acquired lock "refresh_cache-e9ac118f-08b3-430b-848c-461c2b2e3e02" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 822.683865] env[61957]: DEBUG nova.network.neutron [req-0e0abacf-48f4-4b02-bc3b-a2a7a794dd62 req-5b6defff-e62f-44aa-a4e2-30a1e643367c service nova] [instance: e9ac118f-08b3-430b-848c-461c2b2e3e02] Refreshing network info cache for port d6407909-c811-4b71-9ac5-f6128992d3f0 {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 822.830290] env[61957]: DEBUG nova.compute.manager [req-f549238e-e23b-41f0-a8f7-2b2351579b25 req-2c5fb744-df2c-43de-a0c7-70498a2d2b39 service nova] [instance: 14ba830d-4e2c-4e9a-a059-3c86209f0127] Received event network-vif-plugged-6b5aa7fe-7f6f-4305-b6b9-eef2a2c90d69 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 822.830508] env[61957]: DEBUG oslo_concurrency.lockutils [req-f549238e-e23b-41f0-a8f7-2b2351579b25 req-2c5fb744-df2c-43de-a0c7-70498a2d2b39 service nova] Acquiring lock "14ba830d-4e2c-4e9a-a059-3c86209f0127-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 822.830776] env[61957]: DEBUG oslo_concurrency.lockutils [req-f549238e-e23b-41f0-a8f7-2b2351579b25 req-2c5fb744-df2c-43de-a0c7-70498a2d2b39 service nova] Lock "14ba830d-4e2c-4e9a-a059-3c86209f0127-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 822.830949] env[61957]: DEBUG oslo_concurrency.lockutils [req-f549238e-e23b-41f0-a8f7-2b2351579b25 req-2c5fb744-df2c-43de-a0c7-70498a2d2b39 service nova] Lock "14ba830d-4e2c-4e9a-a059-3c86209f0127-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 822.831126] env[61957]: DEBUG nova.compute.manager [req-f549238e-e23b-41f0-a8f7-2b2351579b25 req-2c5fb744-df2c-43de-a0c7-70498a2d2b39 service nova] [instance: 14ba830d-4e2c-4e9a-a059-3c86209f0127] No waiting events found dispatching network-vif-plugged-6b5aa7fe-7f6f-4305-b6b9-eef2a2c90d69 {{(pid=61957) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 822.831290] env[61957]: WARNING nova.compute.manager [req-f549238e-e23b-41f0-a8f7-2b2351579b25 req-2c5fb744-df2c-43de-a0c7-70498a2d2b39 service nova] [instance: 14ba830d-4e2c-4e9a-a059-3c86209f0127] Received unexpected event network-vif-plugged-6b5aa7fe-7f6f-4305-b6b9-eef2a2c90d69 for instance with vm_state building and task_state spawning. [ 822.831446] env[61957]: DEBUG nova.compute.manager [req-f549238e-e23b-41f0-a8f7-2b2351579b25 req-2c5fb744-df2c-43de-a0c7-70498a2d2b39 service nova] [instance: 14ba830d-4e2c-4e9a-a059-3c86209f0127] Received event network-changed-6b5aa7fe-7f6f-4305-b6b9-eef2a2c90d69 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 822.831595] env[61957]: DEBUG nova.compute.manager [req-f549238e-e23b-41f0-a8f7-2b2351579b25 req-2c5fb744-df2c-43de-a0c7-70498a2d2b39 service nova] [instance: 14ba830d-4e2c-4e9a-a059-3c86209f0127] Refreshing instance network info cache due to event network-changed-6b5aa7fe-7f6f-4305-b6b9-eef2a2c90d69. {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 822.831764] env[61957]: DEBUG oslo_concurrency.lockutils [req-f549238e-e23b-41f0-a8f7-2b2351579b25 req-2c5fb744-df2c-43de-a0c7-70498a2d2b39 service nova] Acquiring lock "refresh_cache-14ba830d-4e2c-4e9a-a059-3c86209f0127" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 822.838410] env[61957]: DEBUG oslo_concurrency.lockutils [None req-e4109c08-00a8-4666-8e84-183191cbc2fb tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 822.846472] env[61957]: DEBUG nova.scheduler.client.report [None req-9387ee6a-989e-4086-be03-e20e994ba789 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 822.880522] env[61957]: INFO nova.compute.manager [None req-83fd723b-e45f-4910-94da-c61980bb25f7 tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] [instance: e0249936-d616-4ffb-8f77-d8107633c42a] instance snapshotting [ 822.885984] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-819d5de5-460c-46ff-a643-5a7a2a3e4339 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.894339] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277412, 'name': CreateVM_Task, 'duration_secs': 0.298634} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.906735] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e9ac118f-08b3-430b-848c-461c2b2e3e02] Created VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 822.907856] env[61957]: DEBUG oslo_concurrency.lockutils [None req-dce171c8-b166-439d-a4ea-36b63ebfc3eb tempest-ServerGroupTestJSON-950728618 tempest-ServerGroupTestJSON-950728618-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 822.908016] env[61957]: DEBUG oslo_concurrency.lockutils [None req-dce171c8-b166-439d-a4ea-36b63ebfc3eb tempest-ServerGroupTestJSON-950728618 tempest-ServerGroupTestJSON-950728618-project-member] Acquired lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 822.908315] env[61957]: DEBUG oslo_concurrency.lockutils [None req-dce171c8-b166-439d-a4ea-36b63ebfc3eb tempest-ServerGroupTestJSON-950728618 tempest-ServerGroupTestJSON-950728618-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 822.909014] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7743e38f-4dec-4e8e-9bbe-83c436a76d97 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.911302] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7553a524-b1d4-4f8a-a92a-a6ce85353dad {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.920638] env[61957]: DEBUG oslo_vmware.api [None req-dce171c8-b166-439d-a4ea-36b63ebfc3eb tempest-ServerGroupTestJSON-950728618 tempest-ServerGroupTestJSON-950728618-project-member] Waiting for the task: (returnval){ [ 822.920638] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]521f4674-0aa4-0a4b-d56b-f013f12d5129" [ 822.920638] env[61957]: _type = "Task" [ 822.920638] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.928757] env[61957]: DEBUG oslo_vmware.api [None req-dce171c8-b166-439d-a4ea-36b63ebfc3eb tempest-ServerGroupTestJSON-950728618 tempest-ServerGroupTestJSON-950728618-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]521f4674-0aa4-0a4b-d56b-f013f12d5129, 'name': SearchDatastore_Task, 'duration_secs': 0.008378} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.929012] env[61957]: DEBUG oslo_concurrency.lockutils [None req-dce171c8-b166-439d-a4ea-36b63ebfc3eb tempest-ServerGroupTestJSON-950728618 tempest-ServerGroupTestJSON-950728618-project-member] Releasing lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 822.929230] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-dce171c8-b166-439d-a4ea-36b63ebfc3eb tempest-ServerGroupTestJSON-950728618 tempest-ServerGroupTestJSON-950728618-project-member] [instance: e9ac118f-08b3-430b-848c-461c2b2e3e02] Processing image 11c76a2c-f705-470a-ba9d-4657858bab38 {{(pid=61957) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 822.929438] env[61957]: DEBUG oslo_concurrency.lockutils [None req-dce171c8-b166-439d-a4ea-36b63ebfc3eb tempest-ServerGroupTestJSON-950728618 tempest-ServerGroupTestJSON-950728618-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 822.929585] env[61957]: DEBUG oslo_concurrency.lockutils [None req-dce171c8-b166-439d-a4ea-36b63ebfc3eb tempest-ServerGroupTestJSON-950728618 tempest-ServerGroupTestJSON-950728618-project-member] Acquired lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 822.929766] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-dce171c8-b166-439d-a4ea-36b63ebfc3eb tempest-ServerGroupTestJSON-950728618 tempest-ServerGroupTestJSON-950728618-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 822.929980] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-54c095c7-8b52-43f7-86da-520ede63c814 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.937061] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-dce171c8-b166-439d-a4ea-36b63ebfc3eb tempest-ServerGroupTestJSON-950728618 tempest-ServerGroupTestJSON-950728618-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 822.937061] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-dce171c8-b166-439d-a4ea-36b63ebfc3eb tempest-ServerGroupTestJSON-950728618 tempest-ServerGroupTestJSON-950728618-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61957) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 822.937323] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f0c1d9a8-ca64-4971-a700-c432e5d27a67 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.942381] env[61957]: DEBUG oslo_vmware.api [None req-dce171c8-b166-439d-a4ea-36b63ebfc3eb tempest-ServerGroupTestJSON-950728618 tempest-ServerGroupTestJSON-950728618-project-member] Waiting for the task: (returnval){ [ 822.942381] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]525cfbe6-2e88-bf57-184a-988f3ad60a06" [ 822.942381] env[61957]: _type = "Task" [ 822.942381] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.950161] env[61957]: DEBUG oslo_vmware.api [None req-dce171c8-b166-439d-a4ea-36b63ebfc3eb tempest-ServerGroupTestJSON-950728618 tempest-ServerGroupTestJSON-950728618-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]525cfbe6-2e88-bf57-184a-988f3ad60a06, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.952248] env[61957]: DEBUG nova.network.neutron [None req-36e797a2-06a5-481b-9dd9-66f4ec9dadf3 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 14ba830d-4e2c-4e9a-a059-3c86209f0127] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 822.994965] env[61957]: INFO nova.compute.manager [None req-79f56a2c-6af4-4a0f-b13f-0dc131681bf7 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] [instance: b0af6294-f30c-4266-ae46-6fb03dc0cbeb] Took 1.03 seconds to deallocate network for instance. [ 823.002452] env[61957]: DEBUG oslo_vmware.api [None req-bc6cc00e-6b7d-4787-8382-d5cf71a4537c tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52aeb11b-bd23-5d8d-b903-f6f55ca6f243, 'name': SearchDatastore_Task, 'duration_secs': 0.009127} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.003547] env[61957]: DEBUG oslo_concurrency.lockutils [None req-bc6cc00e-6b7d-4787-8382-d5cf71a4537c tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 823.003547] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-bc6cc00e-6b7d-4787-8382-d5cf71a4537c tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] [instance: 615dbbed-2b02-4351-9e03-8c13f424a133] Processing image 11c76a2c-f705-470a-ba9d-4657858bab38 {{(pid=61957) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 823.003547] env[61957]: DEBUG oslo_concurrency.lockutils [None req-bc6cc00e-6b7d-4787-8382-d5cf71a4537c tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 823.003547] env[61957]: DEBUG oslo_concurrency.lockutils [None req-bc6cc00e-6b7d-4787-8382-d5cf71a4537c tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 823.003547] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-bc6cc00e-6b7d-4787-8382-d5cf71a4537c tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 823.003755] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-61274546-c569-412d-92c1-545ac332abb0 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.011462] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-bc6cc00e-6b7d-4787-8382-d5cf71a4537c tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 823.011462] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-bc6cc00e-6b7d-4787-8382-d5cf71a4537c tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61957) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 823.013269] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1b773dd0-9b88-4eb4-8ba5-5009be72ad3f {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.018282] env[61957]: DEBUG oslo_vmware.api [None req-bc6cc00e-6b7d-4787-8382-d5cf71a4537c tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] Waiting for the task: (returnval){ [ 823.018282] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52fce788-9995-f840-2659-c45748fda231" [ 823.018282] env[61957]: _type = "Task" [ 823.018282] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.025927] env[61957]: DEBUG oslo_vmware.api [None req-bc6cc00e-6b7d-4787-8382-d5cf71a4537c tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52fce788-9995-f840-2659-c45748fda231, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.105147] env[61957]: DEBUG nova.network.neutron [None req-36e797a2-06a5-481b-9dd9-66f4ec9dadf3 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 14ba830d-4e2c-4e9a-a059-3c86209f0127] Updating instance_info_cache with network_info: [{"id": "6b5aa7fe-7f6f-4305-b6b9-eef2a2c90d69", "address": "fa:16:3e:51:67:61", "network": {"id": "efe08823-51fe-4620-a849-5345e262cc71", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1035436819-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e16f6dce3f0e44fb96de516f17d4c6f5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "667a2e97-c1be-421d-9941-6b84c2629b43", "external-id": "nsx-vlan-transportzone-484", "segmentation_id": 484, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6b5aa7fe-7f", "ovs_interfaceid": "6b5aa7fe-7f6f-4305-b6b9-eef2a2c90d69", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 823.351268] env[61957]: DEBUG oslo_concurrency.lockutils [None req-9387ee6a-989e-4086-be03-e20e994ba789 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.600s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 823.351767] env[61957]: DEBUG nova.compute.manager [None req-9387ee6a-989e-4086-be03-e20e994ba789 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] [instance: 17f00b20-9d3b-45e6-919d-6fab9999ec77] Start building networks asynchronously for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 823.354294] env[61957]: DEBUG oslo_concurrency.lockutils [None req-7bbc9489-a498-401d-a2af-2f28d461499e tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.277s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 823.356075] env[61957]: INFO nova.compute.claims [None req-7bbc9489-a498-401d-a2af-2f28d461499e tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 4951132e-7247-4772-8f88-3664c6a7e61e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 823.385959] env[61957]: DEBUG nova.network.neutron [req-0e0abacf-48f4-4b02-bc3b-a2a7a794dd62 req-5b6defff-e62f-44aa-a4e2-30a1e643367c service nova] [instance: e9ac118f-08b3-430b-848c-461c2b2e3e02] Updated VIF entry in instance network info cache for port d6407909-c811-4b71-9ac5-f6128992d3f0. {{(pid=61957) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 823.386327] env[61957]: DEBUG nova.network.neutron [req-0e0abacf-48f4-4b02-bc3b-a2a7a794dd62 req-5b6defff-e62f-44aa-a4e2-30a1e643367c service nova] [instance: e9ac118f-08b3-430b-848c-461c2b2e3e02] Updating instance_info_cache with network_info: [{"id": "d6407909-c811-4b71-9ac5-f6128992d3f0", "address": "fa:16:3e:e7:ab:79", "network": {"id": "801ead5c-62aa-4304-bfbb-0a7dfaea9be0", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-1943350964-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d5cc9b51da4d4c07b155a481af0aaaf0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "04ccbc7a-cf8d-4ea2-8411-291a1e27df7b", "external-id": "nsx-vlan-transportzone-998", "segmentation_id": 998, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd6407909-c8", "ovs_interfaceid": "d6407909-c811-4b71-9ac5-f6128992d3f0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 823.422370] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-83fd723b-e45f-4910-94da-c61980bb25f7 tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] [instance: e0249936-d616-4ffb-8f77-d8107633c42a] Creating Snapshot of the VM instance {{(pid=61957) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 823.422683] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-23be5a1a-d532-42ee-9705-e9736f3443b4 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.430397] env[61957]: DEBUG oslo_vmware.api [None req-83fd723b-e45f-4910-94da-c61980bb25f7 tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Waiting for the task: (returnval){ [ 823.430397] env[61957]: value = "task-1277413" [ 823.430397] env[61957]: _type = "Task" [ 823.430397] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.440344] env[61957]: DEBUG oslo_vmware.api [None req-83fd723b-e45f-4910-94da-c61980bb25f7 tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Task: {'id': task-1277413, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.452064] env[61957]: DEBUG oslo_vmware.api [None req-dce171c8-b166-439d-a4ea-36b63ebfc3eb tempest-ServerGroupTestJSON-950728618 tempest-ServerGroupTestJSON-950728618-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]525cfbe6-2e88-bf57-184a-988f3ad60a06, 'name': SearchDatastore_Task, 'duration_secs': 0.007921} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.452858] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-be429c03-472c-4d38-9adc-d27439baf1e1 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.457863] env[61957]: DEBUG oslo_vmware.api [None req-dce171c8-b166-439d-a4ea-36b63ebfc3eb tempest-ServerGroupTestJSON-950728618 tempest-ServerGroupTestJSON-950728618-project-member] Waiting for the task: (returnval){ [ 823.457863] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52fb7f04-0ddb-dc9e-fb1a-45d5f67ca820" [ 823.457863] env[61957]: _type = "Task" [ 823.457863] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.465260] env[61957]: DEBUG oslo_vmware.api [None req-dce171c8-b166-439d-a4ea-36b63ebfc3eb tempest-ServerGroupTestJSON-950728618 tempest-ServerGroupTestJSON-950728618-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52fb7f04-0ddb-dc9e-fb1a-45d5f67ca820, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.529036] env[61957]: DEBUG oslo_vmware.api [None req-bc6cc00e-6b7d-4787-8382-d5cf71a4537c tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52fce788-9995-f840-2659-c45748fda231, 'name': SearchDatastore_Task, 'duration_secs': 0.0085} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.529036] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cd5dfa40-e665-4bdd-913f-6e52d66555e2 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.534231] env[61957]: DEBUG oslo_vmware.api [None req-bc6cc00e-6b7d-4787-8382-d5cf71a4537c tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] Waiting for the task: (returnval){ [ 823.534231] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52b7fdc8-a893-6dd7-1cc5-e8639091b843" [ 823.534231] env[61957]: _type = "Task" [ 823.534231] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.542585] env[61957]: DEBUG oslo_vmware.api [None req-bc6cc00e-6b7d-4787-8382-d5cf71a4537c tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52b7fdc8-a893-6dd7-1cc5-e8639091b843, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.611352] env[61957]: DEBUG oslo_concurrency.lockutils [None req-36e797a2-06a5-481b-9dd9-66f4ec9dadf3 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Releasing lock "refresh_cache-14ba830d-4e2c-4e9a-a059-3c86209f0127" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 823.611352] env[61957]: DEBUG nova.compute.manager [None req-36e797a2-06a5-481b-9dd9-66f4ec9dadf3 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 14ba830d-4e2c-4e9a-a059-3c86209f0127] Instance network_info: |[{"id": "6b5aa7fe-7f6f-4305-b6b9-eef2a2c90d69", "address": "fa:16:3e:51:67:61", "network": {"id": "efe08823-51fe-4620-a849-5345e262cc71", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1035436819-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e16f6dce3f0e44fb96de516f17d4c6f5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "667a2e97-c1be-421d-9941-6b84c2629b43", "external-id": "nsx-vlan-transportzone-484", "segmentation_id": 484, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6b5aa7fe-7f", "ovs_interfaceid": "6b5aa7fe-7f6f-4305-b6b9-eef2a2c90d69", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 823.611603] env[61957]: DEBUG oslo_concurrency.lockutils [req-f549238e-e23b-41f0-a8f7-2b2351579b25 req-2c5fb744-df2c-43de-a0c7-70498a2d2b39 service nova] Acquired lock "refresh_cache-14ba830d-4e2c-4e9a-a059-3c86209f0127" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 823.611897] env[61957]: DEBUG nova.network.neutron [req-f549238e-e23b-41f0-a8f7-2b2351579b25 req-2c5fb744-df2c-43de-a0c7-70498a2d2b39 service nova] [instance: 14ba830d-4e2c-4e9a-a059-3c86209f0127] Refreshing network info cache for port 6b5aa7fe-7f6f-4305-b6b9-eef2a2c90d69 {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 823.613463] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-36e797a2-06a5-481b-9dd9-66f4ec9dadf3 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 14ba830d-4e2c-4e9a-a059-3c86209f0127] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:51:67:61', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '667a2e97-c1be-421d-9941-6b84c2629b43', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6b5aa7fe-7f6f-4305-b6b9-eef2a2c90d69', 'vif_model': 'vmxnet3'}] {{(pid=61957) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 823.626473] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-36e797a2-06a5-481b-9dd9-66f4ec9dadf3 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Creating folder: Project (e16f6dce3f0e44fb96de516f17d4c6f5). Parent ref: group-v274445. {{(pid=61957) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 823.627346] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-11e21c83-4386-45a7-9e8a-cfa3cde0cae3 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.638875] env[61957]: INFO nova.virt.vmwareapi.vm_util [None req-36e797a2-06a5-481b-9dd9-66f4ec9dadf3 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Created folder: Project (e16f6dce3f0e44fb96de516f17d4c6f5) in parent group-v274445. [ 823.639224] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-36e797a2-06a5-481b-9dd9-66f4ec9dadf3 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Creating folder: Instances. Parent ref: group-v274481. {{(pid=61957) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 823.639529] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e473af5d-ec9d-49e2-b514-a750db99ae9e {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.648408] env[61957]: INFO nova.virt.vmwareapi.vm_util [None req-36e797a2-06a5-481b-9dd9-66f4ec9dadf3 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Created folder: Instances in parent group-v274481. [ 823.648642] env[61957]: DEBUG oslo.service.loopingcall [None req-36e797a2-06a5-481b-9dd9-66f4ec9dadf3 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 823.648836] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 14ba830d-4e2c-4e9a-a059-3c86209f0127] Creating VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 823.649039] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-de7e9402-0872-46c9-8a2c-9d7fcf30f14c {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.667828] env[61957]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 823.667828] env[61957]: value = "task-1277416" [ 823.667828] env[61957]: _type = "Task" [ 823.667828] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.675422] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277416, 'name': CreateVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.860883] env[61957]: DEBUG nova.compute.utils [None req-9387ee6a-989e-4086-be03-e20e994ba789 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Using /dev/sd instead of None {{(pid=61957) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 823.864521] env[61957]: DEBUG nova.compute.manager [None req-9387ee6a-989e-4086-be03-e20e994ba789 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] [instance: 17f00b20-9d3b-45e6-919d-6fab9999ec77] Not allocating networking since 'none' was specified. {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 823.888384] env[61957]: DEBUG oslo_concurrency.lockutils [req-0e0abacf-48f4-4b02-bc3b-a2a7a794dd62 req-5b6defff-e62f-44aa-a4e2-30a1e643367c service nova] Releasing lock "refresh_cache-e9ac118f-08b3-430b-848c-461c2b2e3e02" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 823.941302] env[61957]: DEBUG oslo_vmware.api [None req-83fd723b-e45f-4910-94da-c61980bb25f7 tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Task: {'id': task-1277413, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.967868] env[61957]: DEBUG oslo_vmware.api [None req-dce171c8-b166-439d-a4ea-36b63ebfc3eb tempest-ServerGroupTestJSON-950728618 tempest-ServerGroupTestJSON-950728618-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52fb7f04-0ddb-dc9e-fb1a-45d5f67ca820, 'name': SearchDatastore_Task, 'duration_secs': 0.007836} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.968104] env[61957]: DEBUG oslo_concurrency.lockutils [None req-dce171c8-b166-439d-a4ea-36b63ebfc3eb tempest-ServerGroupTestJSON-950728618 tempest-ServerGroupTestJSON-950728618-project-member] Releasing lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 823.968366] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-dce171c8-b166-439d-a4ea-36b63ebfc3eb tempest-ServerGroupTestJSON-950728618 tempest-ServerGroupTestJSON-950728618-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore1] e9ac118f-08b3-430b-848c-461c2b2e3e02/e9ac118f-08b3-430b-848c-461c2b2e3e02.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 823.968627] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c1f3d3a0-3de0-4c20-b75f-fbd87a6135c9 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.975134] env[61957]: DEBUG oslo_vmware.api [None req-dce171c8-b166-439d-a4ea-36b63ebfc3eb tempest-ServerGroupTestJSON-950728618 tempest-ServerGroupTestJSON-950728618-project-member] Waiting for the task: (returnval){ [ 823.975134] env[61957]: value = "task-1277417" [ 823.975134] env[61957]: _type = "Task" [ 823.975134] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.982505] env[61957]: DEBUG oslo_vmware.api [None req-dce171c8-b166-439d-a4ea-36b63ebfc3eb tempest-ServerGroupTestJSON-950728618 tempest-ServerGroupTestJSON-950728618-project-member] Task: {'id': task-1277417, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.030851] env[61957]: INFO nova.scheduler.client.report [None req-79f56a2c-6af4-4a0f-b13f-0dc131681bf7 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Deleted allocations for instance b0af6294-f30c-4266-ae46-6fb03dc0cbeb [ 824.047861] env[61957]: DEBUG oslo_vmware.api [None req-bc6cc00e-6b7d-4787-8382-d5cf71a4537c tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52b7fdc8-a893-6dd7-1cc5-e8639091b843, 'name': SearchDatastore_Task, 'duration_secs': 0.008453} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.048141] env[61957]: DEBUG oslo_concurrency.lockutils [None req-bc6cc00e-6b7d-4787-8382-d5cf71a4537c tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 824.048386] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-bc6cc00e-6b7d-4787-8382-d5cf71a4537c tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore2] 615dbbed-2b02-4351-9e03-8c13f424a133/615dbbed-2b02-4351-9e03-8c13f424a133.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 824.048639] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b902aa0e-03f7-41a7-a40f-3f02a4606afa {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.054891] env[61957]: DEBUG oslo_vmware.api [None req-bc6cc00e-6b7d-4787-8382-d5cf71a4537c tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] Waiting for the task: (returnval){ [ 824.054891] env[61957]: value = "task-1277418" [ 824.054891] env[61957]: _type = "Task" [ 824.054891] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.064312] env[61957]: DEBUG oslo_vmware.api [None req-bc6cc00e-6b7d-4787-8382-d5cf71a4537c tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] Task: {'id': task-1277418, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.177634] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277416, 'name': CreateVM_Task, 'duration_secs': 0.295472} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.177855] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 14ba830d-4e2c-4e9a-a059-3c86209f0127] Created VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 824.178583] env[61957]: DEBUG oslo_concurrency.lockutils [None req-36e797a2-06a5-481b-9dd9-66f4ec9dadf3 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 824.178762] env[61957]: DEBUG oslo_concurrency.lockutils [None req-36e797a2-06a5-481b-9dd9-66f4ec9dadf3 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 824.179125] env[61957]: DEBUG oslo_concurrency.lockutils [None req-36e797a2-06a5-481b-9dd9-66f4ec9dadf3 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 824.179384] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-162c2ca4-430f-40b2-bc17-78f417619094 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.184640] env[61957]: DEBUG oslo_vmware.api [None req-36e797a2-06a5-481b-9dd9-66f4ec9dadf3 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Waiting for the task: (returnval){ [ 824.184640] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]5275a64f-be69-0f16-10e3-a9455d3e7440" [ 824.184640] env[61957]: _type = "Task" [ 824.184640] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.192816] env[61957]: DEBUG oslo_vmware.api [None req-36e797a2-06a5-481b-9dd9-66f4ec9dadf3 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]5275a64f-be69-0f16-10e3-a9455d3e7440, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.365449] env[61957]: DEBUG nova.compute.manager [None req-9387ee6a-989e-4086-be03-e20e994ba789 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] [instance: 17f00b20-9d3b-45e6-919d-6fab9999ec77] Start building block device mappings for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 824.403463] env[61957]: DEBUG nova.scheduler.client.report [None req-7bbc9489-a498-401d-a2af-2f28d461499e tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Refreshing inventories for resource provider 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 824.421335] env[61957]: DEBUG nova.scheduler.client.report [None req-7bbc9489-a498-401d-a2af-2f28d461499e tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Updating ProviderTree inventory for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 824.421758] env[61957]: DEBUG nova.compute.provider_tree [None req-7bbc9489-a498-401d-a2af-2f28d461499e tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Updating inventory in ProviderTree for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 824.430915] env[61957]: DEBUG nova.network.neutron [req-f549238e-e23b-41f0-a8f7-2b2351579b25 req-2c5fb744-df2c-43de-a0c7-70498a2d2b39 service nova] [instance: 14ba830d-4e2c-4e9a-a059-3c86209f0127] Updated VIF entry in instance network info cache for port 6b5aa7fe-7f6f-4305-b6b9-eef2a2c90d69. {{(pid=61957) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 824.431329] env[61957]: DEBUG nova.network.neutron [req-f549238e-e23b-41f0-a8f7-2b2351579b25 req-2c5fb744-df2c-43de-a0c7-70498a2d2b39 service nova] [instance: 14ba830d-4e2c-4e9a-a059-3c86209f0127] Updating instance_info_cache with network_info: [{"id": "6b5aa7fe-7f6f-4305-b6b9-eef2a2c90d69", "address": "fa:16:3e:51:67:61", "network": {"id": "efe08823-51fe-4620-a849-5345e262cc71", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1035436819-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e16f6dce3f0e44fb96de516f17d4c6f5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "667a2e97-c1be-421d-9941-6b84c2629b43", "external-id": "nsx-vlan-transportzone-484", "segmentation_id": 484, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6b5aa7fe-7f", "ovs_interfaceid": "6b5aa7fe-7f6f-4305-b6b9-eef2a2c90d69", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 824.444565] env[61957]: DEBUG oslo_vmware.api [None req-83fd723b-e45f-4910-94da-c61980bb25f7 tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Task: {'id': task-1277413, 'name': CreateSnapshot_Task, 'duration_secs': 0.614386} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.444565] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-83fd723b-e45f-4910-94da-c61980bb25f7 tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] [instance: e0249936-d616-4ffb-8f77-d8107633c42a] Created Snapshot of the VM instance {{(pid=61957) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 824.445323] env[61957]: DEBUG nova.scheduler.client.report [None req-7bbc9489-a498-401d-a2af-2f28d461499e tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Refreshing aggregate associations for resource provider 0ceb6c9e-61c6-496d-8579-9d32627e96da, aggregates: None {{(pid=61957) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 824.447943] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d21d5b69-9591-4cd2-a244-15e5d0885d64 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.471511] env[61957]: DEBUG nova.scheduler.client.report [None req-7bbc9489-a498-401d-a2af-2f28d461499e tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Refreshing trait associations for resource provider 0ceb6c9e-61c6-496d-8579-9d32627e96da, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO,HW_ARCH_X86_64,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=61957) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 824.488684] env[61957]: DEBUG oslo_vmware.api [None req-dce171c8-b166-439d-a4ea-36b63ebfc3eb tempest-ServerGroupTestJSON-950728618 tempest-ServerGroupTestJSON-950728618-project-member] Task: {'id': task-1277417, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.544540] env[61957]: DEBUG oslo_concurrency.lockutils [None req-79f56a2c-6af4-4a0f-b13f-0dc131681bf7 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Lock "b0af6294-f30c-4266-ae46-6fb03dc0cbeb" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 167.315s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 824.571628] env[61957]: DEBUG oslo_vmware.api [None req-bc6cc00e-6b7d-4787-8382-d5cf71a4537c tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] Task: {'id': task-1277418, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.696364] env[61957]: DEBUG oslo_vmware.api [None req-36e797a2-06a5-481b-9dd9-66f4ec9dadf3 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]5275a64f-be69-0f16-10e3-a9455d3e7440, 'name': SearchDatastore_Task, 'duration_secs': 0.017543} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.696675] env[61957]: DEBUG oslo_concurrency.lockutils [None req-36e797a2-06a5-481b-9dd9-66f4ec9dadf3 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 824.696901] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-36e797a2-06a5-481b-9dd9-66f4ec9dadf3 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 14ba830d-4e2c-4e9a-a059-3c86209f0127] Processing image 11c76a2c-f705-470a-ba9d-4657858bab38 {{(pid=61957) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 824.697149] env[61957]: DEBUG oslo_concurrency.lockutils [None req-36e797a2-06a5-481b-9dd9-66f4ec9dadf3 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 824.697292] env[61957]: DEBUG oslo_concurrency.lockutils [None req-36e797a2-06a5-481b-9dd9-66f4ec9dadf3 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 824.697466] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-36e797a2-06a5-481b-9dd9-66f4ec9dadf3 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 824.697716] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-045b6fd6-085b-4e24-a3b6-9f343e8e770a {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.713082] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-36e797a2-06a5-481b-9dd9-66f4ec9dadf3 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 824.713281] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-36e797a2-06a5-481b-9dd9-66f4ec9dadf3 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61957) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 824.713990] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cecfb5d5-1836-4978-abe4-90b73fba14bf {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.718762] env[61957]: DEBUG oslo_vmware.api [None req-36e797a2-06a5-481b-9dd9-66f4ec9dadf3 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Waiting for the task: (returnval){ [ 824.718762] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]5274178e-4ead-9150-c57d-bf8ad1b41440" [ 824.718762] env[61957]: _type = "Task" [ 824.718762] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.728388] env[61957]: DEBUG oslo_vmware.api [None req-36e797a2-06a5-481b-9dd9-66f4ec9dadf3 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]5274178e-4ead-9150-c57d-bf8ad1b41440, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.816808] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-398fbb9b-1253-4d4f-93c7-8a49223b37f5 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.824382] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a597456-2053-4723-a4dd-a6a54d82bfe5 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.855932] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ad078f3-290e-40b9-bf2c-1a4f17f62bcd {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.863485] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-effaf0d6-ec35-4b27-b58d-e9d56431ee6b {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.879858] env[61957]: DEBUG nova.compute.provider_tree [None req-7bbc9489-a498-401d-a2af-2f28d461499e tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 824.941476] env[61957]: DEBUG oslo_concurrency.lockutils [req-f549238e-e23b-41f0-a8f7-2b2351579b25 req-2c5fb744-df2c-43de-a0c7-70498a2d2b39 service nova] Releasing lock "refresh_cache-14ba830d-4e2c-4e9a-a059-3c86209f0127" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 824.974261] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-83fd723b-e45f-4910-94da-c61980bb25f7 tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] [instance: e0249936-d616-4ffb-8f77-d8107633c42a] Creating linked-clone VM from snapshot {{(pid=61957) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 824.974613] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-f1ea6497-b17b-45af-8025-46fe0e73c3cf {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.986594] env[61957]: DEBUG oslo_vmware.api [None req-dce171c8-b166-439d-a4ea-36b63ebfc3eb tempest-ServerGroupTestJSON-950728618 tempest-ServerGroupTestJSON-950728618-project-member] Task: {'id': task-1277417, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.651973} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.987792] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-dce171c8-b166-439d-a4ea-36b63ebfc3eb tempest-ServerGroupTestJSON-950728618 tempest-ServerGroupTestJSON-950728618-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore1] e9ac118f-08b3-430b-848c-461c2b2e3e02/e9ac118f-08b3-430b-848c-461c2b2e3e02.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 824.988012] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-dce171c8-b166-439d-a4ea-36b63ebfc3eb tempest-ServerGroupTestJSON-950728618 tempest-ServerGroupTestJSON-950728618-project-member] [instance: e9ac118f-08b3-430b-848c-461c2b2e3e02] Extending root virtual disk to 1048576 {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 824.988317] env[61957]: DEBUG oslo_vmware.api [None req-83fd723b-e45f-4910-94da-c61980bb25f7 tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Waiting for the task: (returnval){ [ 824.988317] env[61957]: value = "task-1277419" [ 824.988317] env[61957]: _type = "Task" [ 824.988317] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.988508] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-01fd3972-7ad5-4b33-978d-1cd70e6888c1 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.996850] env[61957]: DEBUG oslo_vmware.api [None req-dce171c8-b166-439d-a4ea-36b63ebfc3eb tempest-ServerGroupTestJSON-950728618 tempest-ServerGroupTestJSON-950728618-project-member] Waiting for the task: (returnval){ [ 824.996850] env[61957]: value = "task-1277420" [ 824.996850] env[61957]: _type = "Task" [ 824.996850] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.999512] env[61957]: DEBUG oslo_vmware.api [None req-83fd723b-e45f-4910-94da-c61980bb25f7 tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Task: {'id': task-1277419, 'name': CloneVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.006557] env[61957]: DEBUG oslo_vmware.api [None req-dce171c8-b166-439d-a4ea-36b63ebfc3eb tempest-ServerGroupTestJSON-950728618 tempest-ServerGroupTestJSON-950728618-project-member] Task: {'id': task-1277420, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.047939] env[61957]: DEBUG nova.compute.manager [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: 6438fe40-046c-45d5-9986-8f182ecde49f] Starting instance... {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 825.066740] env[61957]: DEBUG oslo_vmware.api [None req-bc6cc00e-6b7d-4787-8382-d5cf71a4537c tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] Task: {'id': task-1277418, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.690764} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.067235] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-bc6cc00e-6b7d-4787-8382-d5cf71a4537c tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore2] 615dbbed-2b02-4351-9e03-8c13f424a133/615dbbed-2b02-4351-9e03-8c13f424a133.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 825.067235] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-bc6cc00e-6b7d-4787-8382-d5cf71a4537c tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] [instance: 615dbbed-2b02-4351-9e03-8c13f424a133] Extending root virtual disk to 1048576 {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 825.067472] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0a6d6b13-6b2c-4466-be33-f02660817f01 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.074095] env[61957]: DEBUG oslo_vmware.api [None req-bc6cc00e-6b7d-4787-8382-d5cf71a4537c tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] Waiting for the task: (returnval){ [ 825.074095] env[61957]: value = "task-1277421" [ 825.074095] env[61957]: _type = "Task" [ 825.074095] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.082371] env[61957]: DEBUG oslo_vmware.api [None req-bc6cc00e-6b7d-4787-8382-d5cf71a4537c tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] Task: {'id': task-1277421, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.233210] env[61957]: DEBUG oslo_vmware.api [None req-36e797a2-06a5-481b-9dd9-66f4ec9dadf3 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]5274178e-4ead-9150-c57d-bf8ad1b41440, 'name': SearchDatastore_Task, 'duration_secs': 0.030807} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.234386] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bcc1f543-e3cf-4285-8182-bc3e594820f4 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.240969] env[61957]: DEBUG oslo_vmware.api [None req-36e797a2-06a5-481b-9dd9-66f4ec9dadf3 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Waiting for the task: (returnval){ [ 825.240969] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]5238c74d-82e7-6286-7aa4-8921d91593e7" [ 825.240969] env[61957]: _type = "Task" [ 825.240969] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.252990] env[61957]: DEBUG oslo_vmware.api [None req-36e797a2-06a5-481b-9dd9-66f4ec9dadf3 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]5238c74d-82e7-6286-7aa4-8921d91593e7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.383260] env[61957]: DEBUG nova.compute.manager [None req-9387ee6a-989e-4086-be03-e20e994ba789 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] [instance: 17f00b20-9d3b-45e6-919d-6fab9999ec77] Start spawning the instance on the hypervisor. {{(pid=61957) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 825.386091] env[61957]: DEBUG nova.scheduler.client.report [None req-7bbc9489-a498-401d-a2af-2f28d461499e tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 825.418665] env[61957]: DEBUG nova.virt.hardware [None req-9387ee6a-989e-4086-be03-e20e994ba789 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T17:22:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T17:21:50Z,direct_url=,disk_format='vmdk',id=11c76a2c-f705-470a-ba9d-4657858bab38,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='81d3096d0f094373913fc3dc8f5c3c6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T17:21:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 825.418949] env[61957]: DEBUG nova.virt.hardware [None req-9387ee6a-989e-4086-be03-e20e994ba789 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Flavor limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 825.419124] env[61957]: DEBUG nova.virt.hardware [None req-9387ee6a-989e-4086-be03-e20e994ba789 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Image limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 825.419327] env[61957]: DEBUG nova.virt.hardware [None req-9387ee6a-989e-4086-be03-e20e994ba789 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Flavor pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 825.419451] env[61957]: DEBUG nova.virt.hardware [None req-9387ee6a-989e-4086-be03-e20e994ba789 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Image pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 825.419618] env[61957]: DEBUG nova.virt.hardware [None req-9387ee6a-989e-4086-be03-e20e994ba789 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 825.419841] env[61957]: DEBUG nova.virt.hardware [None req-9387ee6a-989e-4086-be03-e20e994ba789 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 825.419999] env[61957]: DEBUG nova.virt.hardware [None req-9387ee6a-989e-4086-be03-e20e994ba789 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 825.420528] env[61957]: DEBUG nova.virt.hardware [None req-9387ee6a-989e-4086-be03-e20e994ba789 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Got 1 possible topologies {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 825.420737] env[61957]: DEBUG nova.virt.hardware [None req-9387ee6a-989e-4086-be03-e20e994ba789 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 825.420884] env[61957]: DEBUG nova.virt.hardware [None req-9387ee6a-989e-4086-be03-e20e994ba789 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 825.421960] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-936b088f-fcc6-4c3c-91cf-154a41aa4ac1 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.431011] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5ffc528-f0aa-48f3-bfcb-b83aad3dc14e {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.445852] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-9387ee6a-989e-4086-be03-e20e994ba789 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] [instance: 17f00b20-9d3b-45e6-919d-6fab9999ec77] Instance VIF info [] {{(pid=61957) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 825.455159] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-9387ee6a-989e-4086-be03-e20e994ba789 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Creating folder: Project (6cd7313ee475455b9dad79903c857497). Parent ref: group-v274445. {{(pid=61957) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 825.455826] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-951b7e1a-8197-44ac-be89-8e3c9fe153ef {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.466453] env[61957]: INFO nova.virt.vmwareapi.vm_util [None req-9387ee6a-989e-4086-be03-e20e994ba789 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Created folder: Project (6cd7313ee475455b9dad79903c857497) in parent group-v274445. [ 825.466664] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-9387ee6a-989e-4086-be03-e20e994ba789 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Creating folder: Instances. Parent ref: group-v274486. {{(pid=61957) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 825.466902] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f8cba9de-53a6-4735-b5c3-ff8955607308 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.476701] env[61957]: INFO nova.virt.vmwareapi.vm_util [None req-9387ee6a-989e-4086-be03-e20e994ba789 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Created folder: Instances in parent group-v274486. [ 825.477081] env[61957]: DEBUG oslo.service.loopingcall [None req-9387ee6a-989e-4086-be03-e20e994ba789 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 825.477161] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 17f00b20-9d3b-45e6-919d-6fab9999ec77] Creating VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 825.481017] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7cf459bb-83d7-4f33-8917-56828acb792e {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.500594] env[61957]: DEBUG oslo_vmware.api [None req-83fd723b-e45f-4910-94da-c61980bb25f7 tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Task: {'id': task-1277419, 'name': CloneVM_Task} progress is 94%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.504646] env[61957]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 825.504646] env[61957]: value = "task-1277424" [ 825.504646] env[61957]: _type = "Task" [ 825.504646] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.510690] env[61957]: DEBUG oslo_vmware.api [None req-dce171c8-b166-439d-a4ea-36b63ebfc3eb tempest-ServerGroupTestJSON-950728618 tempest-ServerGroupTestJSON-950728618-project-member] Task: {'id': task-1277420, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064416} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.511223] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-dce171c8-b166-439d-a4ea-36b63ebfc3eb tempest-ServerGroupTestJSON-950728618 tempest-ServerGroupTestJSON-950728618-project-member] [instance: e9ac118f-08b3-430b-848c-461c2b2e3e02] Extended root virtual disk {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 825.511915] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3bde917-4748-41cc-83db-e0c864c4f2c5 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.516747] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277424, 'name': CreateVM_Task} progress is 6%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.536061] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-dce171c8-b166-439d-a4ea-36b63ebfc3eb tempest-ServerGroupTestJSON-950728618 tempest-ServerGroupTestJSON-950728618-project-member] [instance: e9ac118f-08b3-430b-848c-461c2b2e3e02] Reconfiguring VM instance instance-00000035 to attach disk [datastore1] e9ac118f-08b3-430b-848c-461c2b2e3e02/e9ac118f-08b3-430b-848c-461c2b2e3e02.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 825.538859] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6a512987-99df-45f0-8e4d-73b818b946cf {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.559069] env[61957]: DEBUG oslo_vmware.api [None req-dce171c8-b166-439d-a4ea-36b63ebfc3eb tempest-ServerGroupTestJSON-950728618 tempest-ServerGroupTestJSON-950728618-project-member] Waiting for the task: (returnval){ [ 825.559069] env[61957]: value = "task-1277425" [ 825.559069] env[61957]: _type = "Task" [ 825.559069] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.569783] env[61957]: DEBUG oslo_vmware.api [None req-dce171c8-b166-439d-a4ea-36b63ebfc3eb tempest-ServerGroupTestJSON-950728618 tempest-ServerGroupTestJSON-950728618-project-member] Task: {'id': task-1277425, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.576235] env[61957]: DEBUG oslo_concurrency.lockutils [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 825.584725] env[61957]: DEBUG oslo_vmware.api [None req-bc6cc00e-6b7d-4787-8382-d5cf71a4537c tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] Task: {'id': task-1277421, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068907} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.584991] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-bc6cc00e-6b7d-4787-8382-d5cf71a4537c tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] [instance: 615dbbed-2b02-4351-9e03-8c13f424a133] Extended root virtual disk {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 825.585790] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fb8f750-0542-438b-8e97-8ba160d48e1e {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.610576] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-bc6cc00e-6b7d-4787-8382-d5cf71a4537c tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] [instance: 615dbbed-2b02-4351-9e03-8c13f424a133] Reconfiguring VM instance instance-00000034 to attach disk [datastore2] 615dbbed-2b02-4351-9e03-8c13f424a133/615dbbed-2b02-4351-9e03-8c13f424a133.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 825.610994] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f558fc43-c277-4c2d-abbb-a53cee44cf2f {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.631420] env[61957]: DEBUG oslo_vmware.api [None req-bc6cc00e-6b7d-4787-8382-d5cf71a4537c tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] Waiting for the task: (returnval){ [ 825.631420] env[61957]: value = "task-1277426" [ 825.631420] env[61957]: _type = "Task" [ 825.631420] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.642085] env[61957]: DEBUG oslo_vmware.api [None req-bc6cc00e-6b7d-4787-8382-d5cf71a4537c tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] Task: {'id': task-1277426, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.754814] env[61957]: DEBUG oslo_vmware.api [None req-36e797a2-06a5-481b-9dd9-66f4ec9dadf3 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]5238c74d-82e7-6286-7aa4-8921d91593e7, 'name': SearchDatastore_Task, 'duration_secs': 0.011604} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.755127] env[61957]: DEBUG oslo_concurrency.lockutils [None req-36e797a2-06a5-481b-9dd9-66f4ec9dadf3 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 825.755386] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-36e797a2-06a5-481b-9dd9-66f4ec9dadf3 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore2] 14ba830d-4e2c-4e9a-a059-3c86209f0127/14ba830d-4e2c-4e9a-a059-3c86209f0127.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 825.756463] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4469146d-d9d4-43d9-8e86-eb3119393882 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.765113] env[61957]: DEBUG oslo_vmware.api [None req-36e797a2-06a5-481b-9dd9-66f4ec9dadf3 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Waiting for the task: (returnval){ [ 825.765113] env[61957]: value = "task-1277427" [ 825.765113] env[61957]: _type = "Task" [ 825.765113] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.773867] env[61957]: DEBUG oslo_vmware.api [None req-36e797a2-06a5-481b-9dd9-66f4ec9dadf3 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1277427, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.891446] env[61957]: DEBUG oslo_concurrency.lockutils [None req-7bbc9489-a498-401d-a2af-2f28d461499e tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.537s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 825.891747] env[61957]: DEBUG nova.compute.manager [None req-7bbc9489-a498-401d-a2af-2f28d461499e tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 4951132e-7247-4772-8f88-3664c6a7e61e] Start building networks asynchronously for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 825.895817] env[61957]: DEBUG oslo_concurrency.lockutils [None req-3c3df299-fd1f-460a-8cc2-979af1db30d1 tempest-ServersTestJSON-1093633677 tempest-ServersTestJSON-1093633677-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.930s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 825.897345] env[61957]: INFO nova.compute.claims [None req-3c3df299-fd1f-460a-8cc2-979af1db30d1 tempest-ServersTestJSON-1093633677 tempest-ServersTestJSON-1093633677-project-member] [instance: ead6aae1-36b5-4f57-9129-3bb02cf103ce] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 826.002400] env[61957]: DEBUG oslo_vmware.api [None req-83fd723b-e45f-4910-94da-c61980bb25f7 tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Task: {'id': task-1277419, 'name': CloneVM_Task} progress is 94%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.013874] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277424, 'name': CreateVM_Task, 'duration_secs': 0.305842} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.014064] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 17f00b20-9d3b-45e6-919d-6fab9999ec77] Created VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 826.014993] env[61957]: DEBUG oslo_concurrency.lockutils [None req-9387ee6a-989e-4086-be03-e20e994ba789 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 826.015075] env[61957]: DEBUG oslo_concurrency.lockutils [None req-9387ee6a-989e-4086-be03-e20e994ba789 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 826.015399] env[61957]: DEBUG oslo_concurrency.lockutils [None req-9387ee6a-989e-4086-be03-e20e994ba789 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 826.015656] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a942ec2b-828a-4b41-8634-1eed75569669 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.020231] env[61957]: DEBUG oslo_vmware.api [None req-9387ee6a-989e-4086-be03-e20e994ba789 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Waiting for the task: (returnval){ [ 826.020231] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52a0fc75-9185-7657-b9e7-aff21319fe3d" [ 826.020231] env[61957]: _type = "Task" [ 826.020231] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.027490] env[61957]: DEBUG oslo_vmware.api [None req-9387ee6a-989e-4086-be03-e20e994ba789 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52a0fc75-9185-7657-b9e7-aff21319fe3d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.071508] env[61957]: DEBUG oslo_vmware.api [None req-dce171c8-b166-439d-a4ea-36b63ebfc3eb tempest-ServerGroupTestJSON-950728618 tempest-ServerGroupTestJSON-950728618-project-member] Task: {'id': task-1277425, 'name': ReconfigVM_Task, 'duration_secs': 0.284465} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.071796] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-dce171c8-b166-439d-a4ea-36b63ebfc3eb tempest-ServerGroupTestJSON-950728618 tempest-ServerGroupTestJSON-950728618-project-member] [instance: e9ac118f-08b3-430b-848c-461c2b2e3e02] Reconfigured VM instance instance-00000035 to attach disk [datastore1] e9ac118f-08b3-430b-848c-461c2b2e3e02/e9ac118f-08b3-430b-848c-461c2b2e3e02.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 826.072629] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-fa1b9312-0ced-4105-86f0-597e530eb6a9 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.078400] env[61957]: DEBUG oslo_vmware.api [None req-dce171c8-b166-439d-a4ea-36b63ebfc3eb tempest-ServerGroupTestJSON-950728618 tempest-ServerGroupTestJSON-950728618-project-member] Waiting for the task: (returnval){ [ 826.078400] env[61957]: value = "task-1277428" [ 826.078400] env[61957]: _type = "Task" [ 826.078400] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.086576] env[61957]: DEBUG oslo_vmware.api [None req-dce171c8-b166-439d-a4ea-36b63ebfc3eb tempest-ServerGroupTestJSON-950728618 tempest-ServerGroupTestJSON-950728618-project-member] Task: {'id': task-1277428, 'name': Rename_Task} progress is 5%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.143765] env[61957]: DEBUG oslo_vmware.api [None req-bc6cc00e-6b7d-4787-8382-d5cf71a4537c tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] Task: {'id': task-1277426, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.274820] env[61957]: DEBUG oslo_vmware.api [None req-36e797a2-06a5-481b-9dd9-66f4ec9dadf3 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1277427, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.406548] env[61957]: DEBUG nova.compute.utils [None req-7bbc9489-a498-401d-a2af-2f28d461499e tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Using /dev/sd instead of None {{(pid=61957) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 826.410615] env[61957]: DEBUG nova.compute.manager [None req-7bbc9489-a498-401d-a2af-2f28d461499e tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 4951132e-7247-4772-8f88-3664c6a7e61e] Allocating IP information in the background. {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 826.410615] env[61957]: DEBUG nova.network.neutron [None req-7bbc9489-a498-401d-a2af-2f28d461499e tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 4951132e-7247-4772-8f88-3664c6a7e61e] allocate_for_instance() {{(pid=61957) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 826.448873] env[61957]: DEBUG nova.policy [None req-7bbc9489-a498-401d-a2af-2f28d461499e tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6fd1c19dc3a44212ada44445e0919106', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8713aa35bcb24b86ad0b58ca9fc991ba', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61957) authorize /opt/stack/nova/nova/policy.py:203}} [ 826.503614] env[61957]: DEBUG oslo_vmware.api [None req-83fd723b-e45f-4910-94da-c61980bb25f7 tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Task: {'id': task-1277419, 'name': CloneVM_Task} progress is 94%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.530424] env[61957]: DEBUG oslo_vmware.api [None req-9387ee6a-989e-4086-be03-e20e994ba789 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52a0fc75-9185-7657-b9e7-aff21319fe3d, 'name': SearchDatastore_Task, 'duration_secs': 0.012354} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.530784] env[61957]: DEBUG oslo_concurrency.lockutils [None req-9387ee6a-989e-4086-be03-e20e994ba789 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 826.531059] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-9387ee6a-989e-4086-be03-e20e994ba789 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] [instance: 17f00b20-9d3b-45e6-919d-6fab9999ec77] Processing image 11c76a2c-f705-470a-ba9d-4657858bab38 {{(pid=61957) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 826.531297] env[61957]: DEBUG oslo_concurrency.lockutils [None req-9387ee6a-989e-4086-be03-e20e994ba789 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 826.531442] env[61957]: DEBUG oslo_concurrency.lockutils [None req-9387ee6a-989e-4086-be03-e20e994ba789 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 826.531627] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-9387ee6a-989e-4086-be03-e20e994ba789 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 826.531897] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-565595fe-714c-4888-aafa-800abbda7244 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.545927] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-9387ee6a-989e-4086-be03-e20e994ba789 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 826.546060] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-9387ee6a-989e-4086-be03-e20e994ba789 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61957) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 826.546772] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fea7383b-b8ba-4f74-8f24-da416f8f7223 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.552506] env[61957]: DEBUG oslo_vmware.api [None req-9387ee6a-989e-4086-be03-e20e994ba789 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Waiting for the task: (returnval){ [ 826.552506] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52e07a6d-8b77-f12a-8632-2b0c24a74ece" [ 826.552506] env[61957]: _type = "Task" [ 826.552506] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.561745] env[61957]: DEBUG oslo_vmware.api [None req-9387ee6a-989e-4086-be03-e20e994ba789 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52e07a6d-8b77-f12a-8632-2b0c24a74ece, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.598377] env[61957]: DEBUG oslo_vmware.api [None req-dce171c8-b166-439d-a4ea-36b63ebfc3eb tempest-ServerGroupTestJSON-950728618 tempest-ServerGroupTestJSON-950728618-project-member] Task: {'id': task-1277428, 'name': Rename_Task, 'duration_secs': 0.150382} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.598882] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-dce171c8-b166-439d-a4ea-36b63ebfc3eb tempest-ServerGroupTestJSON-950728618 tempest-ServerGroupTestJSON-950728618-project-member] [instance: e9ac118f-08b3-430b-848c-461c2b2e3e02] Powering on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 826.598882] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3ff0e015-f3a6-4e64-a1bc-455a69d07399 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.605628] env[61957]: DEBUG oslo_vmware.api [None req-dce171c8-b166-439d-a4ea-36b63ebfc3eb tempest-ServerGroupTestJSON-950728618 tempest-ServerGroupTestJSON-950728618-project-member] Waiting for the task: (returnval){ [ 826.605628] env[61957]: value = "task-1277429" [ 826.605628] env[61957]: _type = "Task" [ 826.605628] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.613366] env[61957]: DEBUG oslo_vmware.api [None req-dce171c8-b166-439d-a4ea-36b63ebfc3eb tempest-ServerGroupTestJSON-950728618 tempest-ServerGroupTestJSON-950728618-project-member] Task: {'id': task-1277429, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.640491] env[61957]: DEBUG oslo_vmware.api [None req-bc6cc00e-6b7d-4787-8382-d5cf71a4537c tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] Task: {'id': task-1277426, 'name': ReconfigVM_Task, 'duration_secs': 0.623776} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.640811] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-bc6cc00e-6b7d-4787-8382-d5cf71a4537c tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] [instance: 615dbbed-2b02-4351-9e03-8c13f424a133] Reconfigured VM instance instance-00000034 to attach disk [datastore2] 615dbbed-2b02-4351-9e03-8c13f424a133/615dbbed-2b02-4351-9e03-8c13f424a133.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 826.641685] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f831576b-bcf2-4a06-8d4e-46deb67b872f {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.650977] env[61957]: DEBUG oslo_vmware.api [None req-bc6cc00e-6b7d-4787-8382-d5cf71a4537c tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] Waiting for the task: (returnval){ [ 826.650977] env[61957]: value = "task-1277430" [ 826.650977] env[61957]: _type = "Task" [ 826.650977] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.656856] env[61957]: DEBUG oslo_vmware.api [None req-bc6cc00e-6b7d-4787-8382-d5cf71a4537c tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] Task: {'id': task-1277430, 'name': Rename_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.765663] env[61957]: DEBUG nova.network.neutron [None req-7bbc9489-a498-401d-a2af-2f28d461499e tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 4951132e-7247-4772-8f88-3664c6a7e61e] Successfully created port: 3016c8eb-2ad3-40b1-8d88-eaec447de9ea {{(pid=61957) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 826.777258] env[61957]: DEBUG oslo_vmware.api [None req-36e797a2-06a5-481b-9dd9-66f4ec9dadf3 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1277427, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.704554} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.777534] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-36e797a2-06a5-481b-9dd9-66f4ec9dadf3 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore2] 14ba830d-4e2c-4e9a-a059-3c86209f0127/14ba830d-4e2c-4e9a-a059-3c86209f0127.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 826.777765] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-36e797a2-06a5-481b-9dd9-66f4ec9dadf3 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 14ba830d-4e2c-4e9a-a059-3c86209f0127] Extending root virtual disk to 1048576 {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 826.778216] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4e416f14-2ddf-4bec-951e-dc37a440aa99 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.784451] env[61957]: DEBUG oslo_vmware.api [None req-36e797a2-06a5-481b-9dd9-66f4ec9dadf3 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Waiting for the task: (returnval){ [ 826.784451] env[61957]: value = "task-1277431" [ 826.784451] env[61957]: _type = "Task" [ 826.784451] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.793124] env[61957]: DEBUG oslo_vmware.api [None req-36e797a2-06a5-481b-9dd9-66f4ec9dadf3 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1277431, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.912825] env[61957]: DEBUG nova.compute.manager [None req-7bbc9489-a498-401d-a2af-2f28d461499e tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 4951132e-7247-4772-8f88-3664c6a7e61e] Start building block device mappings for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 827.004627] env[61957]: DEBUG oslo_vmware.api [None req-83fd723b-e45f-4910-94da-c61980bb25f7 tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Task: {'id': task-1277419, 'name': CloneVM_Task} progress is 100%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.064866] env[61957]: DEBUG oslo_vmware.api [None req-9387ee6a-989e-4086-be03-e20e994ba789 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52e07a6d-8b77-f12a-8632-2b0c24a74ece, 'name': SearchDatastore_Task, 'duration_secs': 0.040788} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.065813] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e5496298-7a87-47c1-9f02-00b37b3d896f {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.074005] env[61957]: DEBUG oslo_vmware.api [None req-9387ee6a-989e-4086-be03-e20e994ba789 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Waiting for the task: (returnval){ [ 827.074005] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]520c3c75-478b-292a-7099-0fc46b3eba7c" [ 827.074005] env[61957]: _type = "Task" [ 827.074005] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.082505] env[61957]: DEBUG oslo_vmware.api [None req-9387ee6a-989e-4086-be03-e20e994ba789 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]520c3c75-478b-292a-7099-0fc46b3eba7c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.121016] env[61957]: DEBUG oslo_vmware.api [None req-dce171c8-b166-439d-a4ea-36b63ebfc3eb tempest-ServerGroupTestJSON-950728618 tempest-ServerGroupTestJSON-950728618-project-member] Task: {'id': task-1277429, 'name': PowerOnVM_Task, 'duration_secs': 0.489993} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.121016] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-dce171c8-b166-439d-a4ea-36b63ebfc3eb tempest-ServerGroupTestJSON-950728618 tempest-ServerGroupTestJSON-950728618-project-member] [instance: e9ac118f-08b3-430b-848c-461c2b2e3e02] Powered on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 827.121016] env[61957]: INFO nova.compute.manager [None req-dce171c8-b166-439d-a4ea-36b63ebfc3eb tempest-ServerGroupTestJSON-950728618 tempest-ServerGroupTestJSON-950728618-project-member] [instance: e9ac118f-08b3-430b-848c-461c2b2e3e02] Took 7.71 seconds to spawn the instance on the hypervisor. [ 827.121016] env[61957]: DEBUG nova.compute.manager [None req-dce171c8-b166-439d-a4ea-36b63ebfc3eb tempest-ServerGroupTestJSON-950728618 tempest-ServerGroupTestJSON-950728618-project-member] [instance: e9ac118f-08b3-430b-848c-461c2b2e3e02] Checking state {{(pid=61957) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 827.121016] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0854c8ba-abb5-4290-8cf2-09247ac1fb07 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.162029] env[61957]: DEBUG oslo_vmware.api [None req-bc6cc00e-6b7d-4787-8382-d5cf71a4537c tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] Task: {'id': task-1277430, 'name': Rename_Task, 'duration_secs': 0.244938} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.162029] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-bc6cc00e-6b7d-4787-8382-d5cf71a4537c tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] [instance: 615dbbed-2b02-4351-9e03-8c13f424a133] Powering on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 827.162029] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a736cde7-0dd5-4fcb-8152-c1f67da88629 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.166790] env[61957]: DEBUG oslo_vmware.api [None req-bc6cc00e-6b7d-4787-8382-d5cf71a4537c tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] Waiting for the task: (returnval){ [ 827.166790] env[61957]: value = "task-1277432" [ 827.166790] env[61957]: _type = "Task" [ 827.166790] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.177415] env[61957]: DEBUG oslo_vmware.api [None req-bc6cc00e-6b7d-4787-8382-d5cf71a4537c tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] Task: {'id': task-1277432, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.273244] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6042266-8568-427a-b393-071122e67c5b {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.282617] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5fca881-d74b-41cb-a634-8be6c07384be {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.295584] env[61957]: DEBUG oslo_vmware.api [None req-36e797a2-06a5-481b-9dd9-66f4ec9dadf3 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1277431, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06369} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.295937] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-36e797a2-06a5-481b-9dd9-66f4ec9dadf3 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 14ba830d-4e2c-4e9a-a059-3c86209f0127] Extended root virtual disk {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 827.296806] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f434231-9d74-48d8-b4e1-d9ebaf7d361d {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.324745] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-525ce397-838c-48a5-8611-4ec7744bdfe3 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.344353] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-36e797a2-06a5-481b-9dd9-66f4ec9dadf3 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 14ba830d-4e2c-4e9a-a059-3c86209f0127] Reconfiguring VM instance instance-00000036 to attach disk [datastore2] 14ba830d-4e2c-4e9a-a059-3c86209f0127/14ba830d-4e2c-4e9a-a059-3c86209f0127.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 827.345011] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e3dc2269-6a90-4a91-b5c8-78a8eab8e90b {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.362722] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd16c4bd-0b81-4eb9-9c21-258ead0c95a1 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.367874] env[61957]: DEBUG oslo_vmware.api [None req-36e797a2-06a5-481b-9dd9-66f4ec9dadf3 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Waiting for the task: (returnval){ [ 827.367874] env[61957]: value = "task-1277433" [ 827.367874] env[61957]: _type = "Task" [ 827.367874] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.378334] env[61957]: DEBUG nova.compute.provider_tree [None req-3c3df299-fd1f-460a-8cc2-979af1db30d1 tempest-ServersTestJSON-1093633677 tempest-ServersTestJSON-1093633677-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 827.387019] env[61957]: DEBUG oslo_vmware.api [None req-36e797a2-06a5-481b-9dd9-66f4ec9dadf3 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1277433, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.411788] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ca2fb2e2-e99f-422a-be65-983bd81b6e15 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Acquiring lock "736eabcf-70f8-4e1e-9830-270d9971fc9a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 827.412090] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ca2fb2e2-e99f-422a-be65-983bd81b6e15 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Lock "736eabcf-70f8-4e1e-9830-270d9971fc9a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 827.506281] env[61957]: DEBUG oslo_vmware.api [None req-83fd723b-e45f-4910-94da-c61980bb25f7 tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Task: {'id': task-1277419, 'name': CloneVM_Task, 'duration_secs': 2.106969} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.506572] env[61957]: INFO nova.virt.vmwareapi.vmops [None req-83fd723b-e45f-4910-94da-c61980bb25f7 tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] [instance: e0249936-d616-4ffb-8f77-d8107633c42a] Created linked-clone VM from snapshot [ 827.507332] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52819a21-129b-4b76-a152-978674471fdf {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.514444] env[61957]: DEBUG nova.virt.vmwareapi.images [None req-83fd723b-e45f-4910-94da-c61980bb25f7 tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] [instance: e0249936-d616-4ffb-8f77-d8107633c42a] Uploading image c5bc6925-6597-4ca2-8c16-f44d676d30b2 {{(pid=61957) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 827.543803] env[61957]: DEBUG oslo_vmware.rw_handles [None req-83fd723b-e45f-4910-94da-c61980bb25f7 tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 827.543803] env[61957]: value = "vm-274485" [ 827.543803] env[61957]: _type = "VirtualMachine" [ 827.543803] env[61957]: }. {{(pid=61957) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 827.544632] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-763595d1-844b-4269-8888-9dd6f23e94f0 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.555881] env[61957]: DEBUG oslo_vmware.rw_handles [None req-83fd723b-e45f-4910-94da-c61980bb25f7 tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Lease: (returnval){ [ 827.555881] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52c171ae-6959-08e0-a357-88cf1848f25a" [ 827.555881] env[61957]: _type = "HttpNfcLease" [ 827.555881] env[61957]: } obtained for exporting VM: (result){ [ 827.555881] env[61957]: value = "vm-274485" [ 827.555881] env[61957]: _type = "VirtualMachine" [ 827.555881] env[61957]: }. {{(pid=61957) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 827.555881] env[61957]: DEBUG oslo_vmware.api [None req-83fd723b-e45f-4910-94da-c61980bb25f7 tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Waiting for the lease: (returnval){ [ 827.555881] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52c171ae-6959-08e0-a357-88cf1848f25a" [ 827.555881] env[61957]: _type = "HttpNfcLease" [ 827.555881] env[61957]: } to be ready. {{(pid=61957) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 827.558547] env[61957]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 827.558547] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52c171ae-6959-08e0-a357-88cf1848f25a" [ 827.558547] env[61957]: _type = "HttpNfcLease" [ 827.558547] env[61957]: } is initializing. {{(pid=61957) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 827.584111] env[61957]: DEBUG oslo_vmware.api [None req-9387ee6a-989e-4086-be03-e20e994ba789 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]520c3c75-478b-292a-7099-0fc46b3eba7c, 'name': SearchDatastore_Task, 'duration_secs': 0.044813} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.584377] env[61957]: DEBUG oslo_concurrency.lockutils [None req-9387ee6a-989e-4086-be03-e20e994ba789 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 827.584631] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-9387ee6a-989e-4086-be03-e20e994ba789 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore2] 17f00b20-9d3b-45e6-919d-6fab9999ec77/17f00b20-9d3b-45e6-919d-6fab9999ec77.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 827.584896] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f3a5dde2-ecce-480c-9b08-e49cb8314b5b {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.591293] env[61957]: DEBUG oslo_vmware.api [None req-9387ee6a-989e-4086-be03-e20e994ba789 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Waiting for the task: (returnval){ [ 827.591293] env[61957]: value = "task-1277435" [ 827.591293] env[61957]: _type = "Task" [ 827.591293] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.599263] env[61957]: DEBUG oslo_vmware.api [None req-9387ee6a-989e-4086-be03-e20e994ba789 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Task: {'id': task-1277435, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.642091] env[61957]: INFO nova.compute.manager [None req-dce171c8-b166-439d-a4ea-36b63ebfc3eb tempest-ServerGroupTestJSON-950728618 tempest-ServerGroupTestJSON-950728618-project-member] [instance: e9ac118f-08b3-430b-848c-461c2b2e3e02] Took 40.04 seconds to build instance. [ 827.679462] env[61957]: DEBUG oslo_vmware.api [None req-bc6cc00e-6b7d-4787-8382-d5cf71a4537c tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] Task: {'id': task-1277432, 'name': PowerOnVM_Task} progress is 64%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.881985] env[61957]: DEBUG oslo_vmware.api [None req-36e797a2-06a5-481b-9dd9-66f4ec9dadf3 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1277433, 'name': ReconfigVM_Task, 'duration_secs': 0.344558} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.882888] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-36e797a2-06a5-481b-9dd9-66f4ec9dadf3 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 14ba830d-4e2c-4e9a-a059-3c86209f0127] Reconfigured VM instance instance-00000036 to attach disk [datastore2] 14ba830d-4e2c-4e9a-a059-3c86209f0127/14ba830d-4e2c-4e9a-a059-3c86209f0127.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 827.883648] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-51b8e6de-9afa-4460-9530-8cc533e0f58e {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.885965] env[61957]: DEBUG nova.scheduler.client.report [None req-3c3df299-fd1f-460a-8cc2-979af1db30d1 tempest-ServersTestJSON-1093633677 tempest-ServersTestJSON-1093633677-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 827.895796] env[61957]: DEBUG oslo_vmware.api [None req-36e797a2-06a5-481b-9dd9-66f4ec9dadf3 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Waiting for the task: (returnval){ [ 827.895796] env[61957]: value = "task-1277436" [ 827.895796] env[61957]: _type = "Task" [ 827.895796] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.908614] env[61957]: DEBUG oslo_vmware.api [None req-36e797a2-06a5-481b-9dd9-66f4ec9dadf3 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1277436, 'name': Rename_Task} progress is 5%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.925810] env[61957]: DEBUG nova.compute.manager [None req-7bbc9489-a498-401d-a2af-2f28d461499e tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 4951132e-7247-4772-8f88-3664c6a7e61e] Start spawning the instance on the hypervisor. {{(pid=61957) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 827.955854] env[61957]: DEBUG nova.virt.hardware [None req-7bbc9489-a498-401d-a2af-2f28d461499e tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T17:22:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T17:21:50Z,direct_url=,disk_format='vmdk',id=11c76a2c-f705-470a-ba9d-4657858bab38,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='81d3096d0f094373913fc3dc8f5c3c6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T17:21:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 827.956553] env[61957]: DEBUG nova.virt.hardware [None req-7bbc9489-a498-401d-a2af-2f28d461499e tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Flavor limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 827.956553] env[61957]: DEBUG nova.virt.hardware [None req-7bbc9489-a498-401d-a2af-2f28d461499e tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Image limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 827.956553] env[61957]: DEBUG nova.virt.hardware [None req-7bbc9489-a498-401d-a2af-2f28d461499e tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Flavor pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 827.956695] env[61957]: DEBUG nova.virt.hardware [None req-7bbc9489-a498-401d-a2af-2f28d461499e tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Image pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 827.956802] env[61957]: DEBUG nova.virt.hardware [None req-7bbc9489-a498-401d-a2af-2f28d461499e tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 827.957025] env[61957]: DEBUG nova.virt.hardware [None req-7bbc9489-a498-401d-a2af-2f28d461499e tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 827.957188] env[61957]: DEBUG nova.virt.hardware [None req-7bbc9489-a498-401d-a2af-2f28d461499e tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 827.957355] env[61957]: DEBUG nova.virt.hardware [None req-7bbc9489-a498-401d-a2af-2f28d461499e tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Got 1 possible topologies {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 827.957540] env[61957]: DEBUG nova.virt.hardware [None req-7bbc9489-a498-401d-a2af-2f28d461499e tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 827.957767] env[61957]: DEBUG nova.virt.hardware [None req-7bbc9489-a498-401d-a2af-2f28d461499e tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 827.958720] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9683e5b4-3c89-4dc3-8603-5b17a8594d54 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.968287] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bff0b04-8e77-40e2-a6d1-341fe168a052 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.064149] env[61957]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 828.064149] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52c171ae-6959-08e0-a357-88cf1848f25a" [ 828.064149] env[61957]: _type = "HttpNfcLease" [ 828.064149] env[61957]: } is ready. {{(pid=61957) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 828.065034] env[61957]: DEBUG oslo_vmware.rw_handles [None req-83fd723b-e45f-4910-94da-c61980bb25f7 tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 828.065034] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52c171ae-6959-08e0-a357-88cf1848f25a" [ 828.065034] env[61957]: _type = "HttpNfcLease" [ 828.065034] env[61957]: }. {{(pid=61957) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 828.065467] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d516a4e2-d44e-4650-8ee4-0a388990a8c9 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.077364] env[61957]: DEBUG oslo_vmware.rw_handles [None req-83fd723b-e45f-4910-94da-c61980bb25f7 tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5271ff36-84ad-9f57-797f-de6dfd662268/disk-0.vmdk from lease info. {{(pid=61957) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 828.077572] env[61957]: DEBUG oslo_vmware.rw_handles [None req-83fd723b-e45f-4910-94da-c61980bb25f7 tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5271ff36-84ad-9f57-797f-de6dfd662268/disk-0.vmdk for reading. {{(pid=61957) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 828.144105] env[61957]: DEBUG oslo_concurrency.lockutils [None req-dce171c8-b166-439d-a4ea-36b63ebfc3eb tempest-ServerGroupTestJSON-950728618 tempest-ServerGroupTestJSON-950728618-project-member] Lock "e9ac118f-08b3-430b-848c-461c2b2e3e02" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 115.229s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 828.151172] env[61957]: DEBUG oslo_vmware.api [None req-9387ee6a-989e-4086-be03-e20e994ba789 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Task: {'id': task-1277435, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.183097] env[61957]: DEBUG oslo_vmware.api [None req-bc6cc00e-6b7d-4787-8382-d5cf71a4537c tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] Task: {'id': task-1277432, 'name': PowerOnVM_Task, 'duration_secs': 1.002286} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.183898] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-bc6cc00e-6b7d-4787-8382-d5cf71a4537c tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] [instance: 615dbbed-2b02-4351-9e03-8c13f424a133] Powered on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 828.183898] env[61957]: INFO nova.compute.manager [None req-bc6cc00e-6b7d-4787-8382-d5cf71a4537c tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] [instance: 615dbbed-2b02-4351-9e03-8c13f424a133] Took 12.24 seconds to spawn the instance on the hypervisor. [ 828.183898] env[61957]: DEBUG nova.compute.manager [None req-bc6cc00e-6b7d-4787-8382-d5cf71a4537c tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] [instance: 615dbbed-2b02-4351-9e03-8c13f424a133] Checking state {{(pid=61957) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 828.184598] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e0130c1-70e3-4a0e-89cf-0ba19c87ef71 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.228717] env[61957]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-2f092b47-d358-4c30-b404-d3c3b5d5668e {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.353905] env[61957]: DEBUG nova.network.neutron [None req-7bbc9489-a498-401d-a2af-2f28d461499e tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 4951132e-7247-4772-8f88-3664c6a7e61e] Successfully updated port: 3016c8eb-2ad3-40b1-8d88-eaec447de9ea {{(pid=61957) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 828.377666] env[61957]: DEBUG nova.compute.manager [req-5524c599-e9e9-4665-95dd-a019f08755bb req-b5c8f91a-d4ea-4d24-9514-c8f0dfdda7ff service nova] [instance: 4951132e-7247-4772-8f88-3664c6a7e61e] Received event network-vif-plugged-3016c8eb-2ad3-40b1-8d88-eaec447de9ea {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 828.377666] env[61957]: DEBUG oslo_concurrency.lockutils [req-5524c599-e9e9-4665-95dd-a019f08755bb req-b5c8f91a-d4ea-4d24-9514-c8f0dfdda7ff service nova] Acquiring lock "4951132e-7247-4772-8f88-3664c6a7e61e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 828.377848] env[61957]: DEBUG oslo_concurrency.lockutils [req-5524c599-e9e9-4665-95dd-a019f08755bb req-b5c8f91a-d4ea-4d24-9514-c8f0dfdda7ff service nova] Lock "4951132e-7247-4772-8f88-3664c6a7e61e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 828.378049] env[61957]: DEBUG oslo_concurrency.lockutils [req-5524c599-e9e9-4665-95dd-a019f08755bb req-b5c8f91a-d4ea-4d24-9514-c8f0dfdda7ff service nova] Lock "4951132e-7247-4772-8f88-3664c6a7e61e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 828.378223] env[61957]: DEBUG nova.compute.manager [req-5524c599-e9e9-4665-95dd-a019f08755bb req-b5c8f91a-d4ea-4d24-9514-c8f0dfdda7ff service nova] [instance: 4951132e-7247-4772-8f88-3664c6a7e61e] No waiting events found dispatching network-vif-plugged-3016c8eb-2ad3-40b1-8d88-eaec447de9ea {{(pid=61957) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 828.378392] env[61957]: WARNING nova.compute.manager [req-5524c599-e9e9-4665-95dd-a019f08755bb req-b5c8f91a-d4ea-4d24-9514-c8f0dfdda7ff service nova] [instance: 4951132e-7247-4772-8f88-3664c6a7e61e] Received unexpected event network-vif-plugged-3016c8eb-2ad3-40b1-8d88-eaec447de9ea for instance with vm_state building and task_state spawning. [ 828.394477] env[61957]: DEBUG oslo_concurrency.lockutils [None req-3c3df299-fd1f-460a-8cc2-979af1db30d1 tempest-ServersTestJSON-1093633677 tempest-ServersTestJSON-1093633677-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.499s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 828.395045] env[61957]: DEBUG nova.compute.manager [None req-3c3df299-fd1f-460a-8cc2-979af1db30d1 tempest-ServersTestJSON-1093633677 tempest-ServersTestJSON-1093633677-project-member] [instance: ead6aae1-36b5-4f57-9129-3bb02cf103ce] Start building networks asynchronously for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 828.398410] env[61957]: DEBUG oslo_concurrency.lockutils [None req-be584a65-2286-47ec-9ff6-29a7835a7d50 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.493s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 828.400702] env[61957]: INFO nova.compute.claims [None req-be584a65-2286-47ec-9ff6-29a7835a7d50 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: 37ce46a1-8f9a-4d15-bd81-e40845a0e48a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 828.415141] env[61957]: DEBUG oslo_vmware.api [None req-36e797a2-06a5-481b-9dd9-66f4ec9dadf3 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1277436, 'name': Rename_Task, 'duration_secs': 0.168212} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.415395] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-36e797a2-06a5-481b-9dd9-66f4ec9dadf3 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 14ba830d-4e2c-4e9a-a059-3c86209f0127] Powering on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 828.415628] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1a487d8a-8128-4a69-9844-75904c9a6bf3 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.421791] env[61957]: DEBUG oslo_vmware.api [None req-36e797a2-06a5-481b-9dd9-66f4ec9dadf3 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Waiting for the task: (returnval){ [ 828.421791] env[61957]: value = "task-1277437" [ 828.421791] env[61957]: _type = "Task" [ 828.421791] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.433102] env[61957]: DEBUG oslo_vmware.api [None req-36e797a2-06a5-481b-9dd9-66f4ec9dadf3 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1277437, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.650556] env[61957]: DEBUG oslo_vmware.api [None req-9387ee6a-989e-4086-be03-e20e994ba789 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Task: {'id': task-1277435, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.616079} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.650880] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-9387ee6a-989e-4086-be03-e20e994ba789 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore2] 17f00b20-9d3b-45e6-919d-6fab9999ec77/17f00b20-9d3b-45e6-919d-6fab9999ec77.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 828.651133] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-9387ee6a-989e-4086-be03-e20e994ba789 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] [instance: 17f00b20-9d3b-45e6-919d-6fab9999ec77] Extending root virtual disk to 1048576 {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 828.652104] env[61957]: DEBUG nova.compute.manager [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: f66db265-887e-4d61-b848-c609e5c884cb] Starting instance... {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 828.655896] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3c508990-4272-40f6-a833-b84228fa058a {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.664068] env[61957]: DEBUG oslo_vmware.api [None req-9387ee6a-989e-4086-be03-e20e994ba789 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Waiting for the task: (returnval){ [ 828.664068] env[61957]: value = "task-1277438" [ 828.664068] env[61957]: _type = "Task" [ 828.664068] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.673769] env[61957]: DEBUG oslo_vmware.api [None req-9387ee6a-989e-4086-be03-e20e994ba789 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Task: {'id': task-1277438, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.709266] env[61957]: INFO nova.compute.manager [None req-bc6cc00e-6b7d-4787-8382-d5cf71a4537c tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] [instance: 615dbbed-2b02-4351-9e03-8c13f424a133] Took 43.05 seconds to build instance. [ 828.789017] env[61957]: DEBUG oslo_concurrency.lockutils [None req-5a8f4b06-9979-4e23-adff-500ed36e9a0d tempest-ServerGroupTestJSON-950728618 tempest-ServerGroupTestJSON-950728618-project-member] Acquiring lock "e9ac118f-08b3-430b-848c-461c2b2e3e02" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 828.789321] env[61957]: DEBUG oslo_concurrency.lockutils [None req-5a8f4b06-9979-4e23-adff-500ed36e9a0d tempest-ServerGroupTestJSON-950728618 tempest-ServerGroupTestJSON-950728618-project-member] Lock "e9ac118f-08b3-430b-848c-461c2b2e3e02" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 828.789732] env[61957]: DEBUG oslo_concurrency.lockutils [None req-5a8f4b06-9979-4e23-adff-500ed36e9a0d tempest-ServerGroupTestJSON-950728618 tempest-ServerGroupTestJSON-950728618-project-member] Acquiring lock "e9ac118f-08b3-430b-848c-461c2b2e3e02-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 828.790049] env[61957]: DEBUG oslo_concurrency.lockutils [None req-5a8f4b06-9979-4e23-adff-500ed36e9a0d tempest-ServerGroupTestJSON-950728618 tempest-ServerGroupTestJSON-950728618-project-member] Lock "e9ac118f-08b3-430b-848c-461c2b2e3e02-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 828.790317] env[61957]: DEBUG oslo_concurrency.lockutils [None req-5a8f4b06-9979-4e23-adff-500ed36e9a0d tempest-ServerGroupTestJSON-950728618 tempest-ServerGroupTestJSON-950728618-project-member] Lock "e9ac118f-08b3-430b-848c-461c2b2e3e02-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 828.793442] env[61957]: INFO nova.compute.manager [None req-5a8f4b06-9979-4e23-adff-500ed36e9a0d tempest-ServerGroupTestJSON-950728618 tempest-ServerGroupTestJSON-950728618-project-member] [instance: e9ac118f-08b3-430b-848c-461c2b2e3e02] Terminating instance [ 828.796263] env[61957]: DEBUG nova.compute.manager [None req-5a8f4b06-9979-4e23-adff-500ed36e9a0d tempest-ServerGroupTestJSON-950728618 tempest-ServerGroupTestJSON-950728618-project-member] [instance: e9ac118f-08b3-430b-848c-461c2b2e3e02] Start destroying the instance on the hypervisor. {{(pid=61957) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 828.796651] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-5a8f4b06-9979-4e23-adff-500ed36e9a0d tempest-ServerGroupTestJSON-950728618 tempest-ServerGroupTestJSON-950728618-project-member] [instance: e9ac118f-08b3-430b-848c-461c2b2e3e02] Destroying instance {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 828.797806] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4e4a7d0-393e-4e83-9a1d-71bed3441e0f {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.807766] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a8f4b06-9979-4e23-adff-500ed36e9a0d tempest-ServerGroupTestJSON-950728618 tempest-ServerGroupTestJSON-950728618-project-member] [instance: e9ac118f-08b3-430b-848c-461c2b2e3e02] Powering off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 828.809722] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-27f7e5b6-e0dc-478c-a849-6e3e7d1fec6e {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.820980] env[61957]: DEBUG oslo_vmware.api [None req-5a8f4b06-9979-4e23-adff-500ed36e9a0d tempest-ServerGroupTestJSON-950728618 tempest-ServerGroupTestJSON-950728618-project-member] Waiting for the task: (returnval){ [ 828.820980] env[61957]: value = "task-1277439" [ 828.820980] env[61957]: _type = "Task" [ 828.820980] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.831512] env[61957]: DEBUG oslo_vmware.api [None req-5a8f4b06-9979-4e23-adff-500ed36e9a0d tempest-ServerGroupTestJSON-950728618 tempest-ServerGroupTestJSON-950728618-project-member] Task: {'id': task-1277439, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.856540] env[61957]: DEBUG oslo_concurrency.lockutils [None req-7bbc9489-a498-401d-a2af-2f28d461499e tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Acquiring lock "refresh_cache-4951132e-7247-4772-8f88-3664c6a7e61e" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 828.856782] env[61957]: DEBUG oslo_concurrency.lockutils [None req-7bbc9489-a498-401d-a2af-2f28d461499e tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Acquired lock "refresh_cache-4951132e-7247-4772-8f88-3664c6a7e61e" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 828.856992] env[61957]: DEBUG nova.network.neutron [None req-7bbc9489-a498-401d-a2af-2f28d461499e tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 4951132e-7247-4772-8f88-3664c6a7e61e] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 828.900290] env[61957]: DEBUG nova.compute.utils [None req-3c3df299-fd1f-460a-8cc2-979af1db30d1 tempest-ServersTestJSON-1093633677 tempest-ServersTestJSON-1093633677-project-member] Using /dev/sd instead of None {{(pid=61957) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 828.903422] env[61957]: DEBUG nova.compute.manager [None req-3c3df299-fd1f-460a-8cc2-979af1db30d1 tempest-ServersTestJSON-1093633677 tempest-ServersTestJSON-1093633677-project-member] [instance: ead6aae1-36b5-4f57-9129-3bb02cf103ce] Allocating IP information in the background. {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 828.903769] env[61957]: DEBUG nova.network.neutron [None req-3c3df299-fd1f-460a-8cc2-979af1db30d1 tempest-ServersTestJSON-1093633677 tempest-ServersTestJSON-1093633677-project-member] [instance: ead6aae1-36b5-4f57-9129-3bb02cf103ce] allocate_for_instance() {{(pid=61957) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 828.932678] env[61957]: DEBUG oslo_vmware.api [None req-36e797a2-06a5-481b-9dd9-66f4ec9dadf3 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1277437, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.958958] env[61957]: DEBUG nova.policy [None req-3c3df299-fd1f-460a-8cc2-979af1db30d1 tempest-ServersTestJSON-1093633677 tempest-ServersTestJSON-1093633677-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b41929f6e4734039b032088fabf5693c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1a127f9f057c43d2bddf261d2d875b63', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61957) authorize /opt/stack/nova/nova/policy.py:203}} [ 829.177553] env[61957]: DEBUG oslo_vmware.api [None req-9387ee6a-989e-4086-be03-e20e994ba789 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Task: {'id': task-1277438, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073009} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.178820] env[61957]: DEBUG oslo_concurrency.lockutils [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 829.179162] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-9387ee6a-989e-4086-be03-e20e994ba789 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] [instance: 17f00b20-9d3b-45e6-919d-6fab9999ec77] Extended root virtual disk {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 829.180429] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1321422-985b-4c61-9b61-55ca21868464 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.201299] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-9387ee6a-989e-4086-be03-e20e994ba789 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] [instance: 17f00b20-9d3b-45e6-919d-6fab9999ec77] Reconfiguring VM instance instance-00000037 to attach disk [datastore2] 17f00b20-9d3b-45e6-919d-6fab9999ec77/17f00b20-9d3b-45e6-919d-6fab9999ec77.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 829.201708] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1c88d78c-4086-483a-95ce-2b277e65f7d0 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.219093] env[61957]: DEBUG oslo_concurrency.lockutils [None req-bc6cc00e-6b7d-4787-8382-d5cf71a4537c tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] Lock "615dbbed-2b02-4351-9e03-8c13f424a133" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 156.059s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 829.227135] env[61957]: DEBUG oslo_vmware.api [None req-9387ee6a-989e-4086-be03-e20e994ba789 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Waiting for the task: (returnval){ [ 829.227135] env[61957]: value = "task-1277440" [ 829.227135] env[61957]: _type = "Task" [ 829.227135] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.237251] env[61957]: DEBUG oslo_vmware.api [None req-9387ee6a-989e-4086-be03-e20e994ba789 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Task: {'id': task-1277440, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.315214] env[61957]: DEBUG nova.network.neutron [None req-3c3df299-fd1f-460a-8cc2-979af1db30d1 tempest-ServersTestJSON-1093633677 tempest-ServersTestJSON-1093633677-project-member] [instance: ead6aae1-36b5-4f57-9129-3bb02cf103ce] Successfully created port: 606b786e-f948-4d1c-a109-6390d9349ea8 {{(pid=61957) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 829.316896] env[61957]: DEBUG oslo_concurrency.lockutils [None req-47642d83-3473-4b34-8e43-c1548504523c tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] Acquiring lock "615dbbed-2b02-4351-9e03-8c13f424a133" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 829.318061] env[61957]: DEBUG oslo_concurrency.lockutils [None req-47642d83-3473-4b34-8e43-c1548504523c tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] Lock "615dbbed-2b02-4351-9e03-8c13f424a133" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 829.318061] env[61957]: DEBUG oslo_concurrency.lockutils [None req-47642d83-3473-4b34-8e43-c1548504523c tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] Acquiring lock "615dbbed-2b02-4351-9e03-8c13f424a133-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 829.318061] env[61957]: DEBUG oslo_concurrency.lockutils [None req-47642d83-3473-4b34-8e43-c1548504523c tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] Lock "615dbbed-2b02-4351-9e03-8c13f424a133-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 829.318896] env[61957]: DEBUG oslo_concurrency.lockutils [None req-47642d83-3473-4b34-8e43-c1548504523c tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] Lock "615dbbed-2b02-4351-9e03-8c13f424a133-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 829.321642] env[61957]: INFO nova.compute.manager [None req-47642d83-3473-4b34-8e43-c1548504523c tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] [instance: 615dbbed-2b02-4351-9e03-8c13f424a133] Terminating instance [ 829.327107] env[61957]: DEBUG nova.compute.manager [None req-47642d83-3473-4b34-8e43-c1548504523c tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] [instance: 615dbbed-2b02-4351-9e03-8c13f424a133] Start destroying the instance on the hypervisor. {{(pid=61957) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 829.327347] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-47642d83-3473-4b34-8e43-c1548504523c tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] [instance: 615dbbed-2b02-4351-9e03-8c13f424a133] Destroying instance {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 829.328323] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8feffd9-4d3c-4850-b4f3-ee6792117bc1 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.337295] env[61957]: DEBUG oslo_vmware.api [None req-5a8f4b06-9979-4e23-adff-500ed36e9a0d tempest-ServerGroupTestJSON-950728618 tempest-ServerGroupTestJSON-950728618-project-member] Task: {'id': task-1277439, 'name': PowerOffVM_Task, 'duration_secs': 0.344937} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.339492] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a8f4b06-9979-4e23-adff-500ed36e9a0d tempest-ServerGroupTestJSON-950728618 tempest-ServerGroupTestJSON-950728618-project-member] [instance: e9ac118f-08b3-430b-848c-461c2b2e3e02] Powered off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 829.340389] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-5a8f4b06-9979-4e23-adff-500ed36e9a0d tempest-ServerGroupTestJSON-950728618 tempest-ServerGroupTestJSON-950728618-project-member] [instance: e9ac118f-08b3-430b-848c-461c2b2e3e02] Unregistering the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 829.340892] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-47642d83-3473-4b34-8e43-c1548504523c tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] [instance: 615dbbed-2b02-4351-9e03-8c13f424a133] Powering off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 829.341142] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5b718256-a842-4fab-ac7f-6502b7a2b59c {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.342998] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c66ffa88-39aa-4172-955e-2adc65fee207 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.350988] env[61957]: DEBUG oslo_vmware.api [None req-47642d83-3473-4b34-8e43-c1548504523c tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] Waiting for the task: (returnval){ [ 829.350988] env[61957]: value = "task-1277442" [ 829.350988] env[61957]: _type = "Task" [ 829.350988] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.362279] env[61957]: DEBUG oslo_vmware.api [None req-47642d83-3473-4b34-8e43-c1548504523c tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] Task: {'id': task-1277442, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.404091] env[61957]: DEBUG nova.compute.manager [None req-3c3df299-fd1f-460a-8cc2-979af1db30d1 tempest-ServersTestJSON-1093633677 tempest-ServersTestJSON-1093633677-project-member] [instance: ead6aae1-36b5-4f57-9129-3bb02cf103ce] Start building block device mappings for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 829.409211] env[61957]: DEBUG nova.network.neutron [None req-7bbc9489-a498-401d-a2af-2f28d461499e tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 4951132e-7247-4772-8f88-3664c6a7e61e] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 829.439116] env[61957]: DEBUG oslo_vmware.api [None req-36e797a2-06a5-481b-9dd9-66f4ec9dadf3 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1277437, 'name': PowerOnVM_Task, 'duration_secs': 0.961064} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.439479] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-36e797a2-06a5-481b-9dd9-66f4ec9dadf3 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 14ba830d-4e2c-4e9a-a059-3c86209f0127] Powered on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 829.439767] env[61957]: INFO nova.compute.manager [None req-36e797a2-06a5-481b-9dd9-66f4ec9dadf3 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 14ba830d-4e2c-4e9a-a059-3c86209f0127] Took 8.66 seconds to spawn the instance on the hypervisor. [ 829.440117] env[61957]: DEBUG nova.compute.manager [None req-36e797a2-06a5-481b-9dd9-66f4ec9dadf3 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 14ba830d-4e2c-4e9a-a059-3c86209f0127] Checking state {{(pid=61957) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 829.441191] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22954a3d-7517-493b-8fa6-20916f8ef3bc {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.542873] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-5a8f4b06-9979-4e23-adff-500ed36e9a0d tempest-ServerGroupTestJSON-950728618 tempest-ServerGroupTestJSON-950728618-project-member] [instance: e9ac118f-08b3-430b-848c-461c2b2e3e02] Unregistered the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 829.543412] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-5a8f4b06-9979-4e23-adff-500ed36e9a0d tempest-ServerGroupTestJSON-950728618 tempest-ServerGroupTestJSON-950728618-project-member] [instance: e9ac118f-08b3-430b-848c-461c2b2e3e02] Deleting contents of the VM from datastore datastore1 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 829.543715] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-5a8f4b06-9979-4e23-adff-500ed36e9a0d tempest-ServerGroupTestJSON-950728618 tempest-ServerGroupTestJSON-950728618-project-member] Deleting the datastore file [datastore1] e9ac118f-08b3-430b-848c-461c2b2e3e02 {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 829.544126] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b17befe6-14f3-4783-b356-88a0d28363da {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.554945] env[61957]: DEBUG oslo_vmware.api [None req-5a8f4b06-9979-4e23-adff-500ed36e9a0d tempest-ServerGroupTestJSON-950728618 tempest-ServerGroupTestJSON-950728618-project-member] Waiting for the task: (returnval){ [ 829.554945] env[61957]: value = "task-1277443" [ 829.554945] env[61957]: _type = "Task" [ 829.554945] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.566935] env[61957]: DEBUG oslo_vmware.api [None req-5a8f4b06-9979-4e23-adff-500ed36e9a0d tempest-ServerGroupTestJSON-950728618 tempest-ServerGroupTestJSON-950728618-project-member] Task: {'id': task-1277443, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.670754] env[61957]: DEBUG nova.network.neutron [None req-7bbc9489-a498-401d-a2af-2f28d461499e tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 4951132e-7247-4772-8f88-3664c6a7e61e] Updating instance_info_cache with network_info: [{"id": "3016c8eb-2ad3-40b1-8d88-eaec447de9ea", "address": "fa:16:3e:59:19:f2", "network": {"id": "364cabcb-046d-4f91-b1ee-5dd3adfae6d7", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1432396967-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8713aa35bcb24b86ad0b58ca9fc991ba", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b7a73c01-1bb9-4612-a1a7-16d71b732e81", "external-id": "nsx-vlan-transportzone-711", "segmentation_id": 711, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3016c8eb-2a", "ovs_interfaceid": "3016c8eb-2ad3-40b1-8d88-eaec447de9ea", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 829.724474] env[61957]: DEBUG nova.compute.manager [None req-ac5c4b48-cfd7-420e-9aa1-225385350d11 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] [instance: c23141ee-0cbb-4d1b-8390-c3073fe354f1] Starting instance... {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 829.739463] env[61957]: DEBUG oslo_vmware.api [None req-9387ee6a-989e-4086-be03-e20e994ba789 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Task: {'id': task-1277440, 'name': ReconfigVM_Task, 'duration_secs': 0.465953} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.739918] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-9387ee6a-989e-4086-be03-e20e994ba789 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] [instance: 17f00b20-9d3b-45e6-919d-6fab9999ec77] Reconfigured VM instance instance-00000037 to attach disk [datastore2] 17f00b20-9d3b-45e6-919d-6fab9999ec77/17f00b20-9d3b-45e6-919d-6fab9999ec77.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 829.742717] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-dbea7d4d-5f39-45c3-8190-b64986714a0d {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.750013] env[61957]: DEBUG oslo_vmware.api [None req-9387ee6a-989e-4086-be03-e20e994ba789 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Waiting for the task: (returnval){ [ 829.750013] env[61957]: value = "task-1277444" [ 829.750013] env[61957]: _type = "Task" [ 829.750013] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.759840] env[61957]: DEBUG oslo_vmware.api [None req-9387ee6a-989e-4086-be03-e20e994ba789 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Task: {'id': task-1277444, 'name': Rename_Task} progress is 6%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.863506] env[61957]: DEBUG oslo_vmware.api [None req-47642d83-3473-4b34-8e43-c1548504523c tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] Task: {'id': task-1277442, 'name': PowerOffVM_Task, 'duration_secs': 0.324308} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.863506] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-47642d83-3473-4b34-8e43-c1548504523c tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] [instance: 615dbbed-2b02-4351-9e03-8c13f424a133] Powered off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 829.863506] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-47642d83-3473-4b34-8e43-c1548504523c tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] [instance: 615dbbed-2b02-4351-9e03-8c13f424a133] Unregistering the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 829.863815] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6868a990-c3a8-407a-91c0-bd5b76e32b30 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.869132] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c832e20-df9a-442b-b3ba-246e35e4d9be {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.877813] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8972999c-e406-42b8-9199-dccd07cfef91 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.908499] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-218a2c98-d2b9-4a9d-a88d-0321870149d4 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.921126] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc4656a5-3639-4c4d-a8ca-cbaa3321dbcd {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.784396] env[61957]: DEBUG oslo_concurrency.lockutils [None req-7bbc9489-a498-401d-a2af-2f28d461499e tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Releasing lock "refresh_cache-4951132e-7247-4772-8f88-3664c6a7e61e" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 830.784396] env[61957]: DEBUG nova.compute.manager [None req-7bbc9489-a498-401d-a2af-2f28d461499e tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 4951132e-7247-4772-8f88-3664c6a7e61e] Instance network_info: |[{"id": "3016c8eb-2ad3-40b1-8d88-eaec447de9ea", "address": "fa:16:3e:59:19:f2", "network": {"id": "364cabcb-046d-4f91-b1ee-5dd3adfae6d7", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1432396967-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8713aa35bcb24b86ad0b58ca9fc991ba", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b7a73c01-1bb9-4612-a1a7-16d71b732e81", "external-id": "nsx-vlan-transportzone-711", "segmentation_id": 711, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3016c8eb-2a", "ovs_interfaceid": "3016c8eb-2ad3-40b1-8d88-eaec447de9ea", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 830.786844] env[61957]: DEBUG nova.compute.manager [None req-3c3df299-fd1f-460a-8cc2-979af1db30d1 tempest-ServersTestJSON-1093633677 tempest-ServersTestJSON-1093633677-project-member] [instance: ead6aae1-36b5-4f57-9129-3bb02cf103ce] Start spawning the instance on the hypervisor. {{(pid=61957) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 830.789153] env[61957]: DEBUG nova.compute.provider_tree [None req-be584a65-2286-47ec-9ff6-29a7835a7d50 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 830.801449] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-7bbc9489-a498-401d-a2af-2f28d461499e tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 4951132e-7247-4772-8f88-3664c6a7e61e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:59:19:f2', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'b7a73c01-1bb9-4612-a1a7-16d71b732e81', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3016c8eb-2ad3-40b1-8d88-eaec447de9ea', 'vif_model': 'vmxnet3'}] {{(pid=61957) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 830.808782] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-7bbc9489-a498-401d-a2af-2f28d461499e tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Creating folder: Project (8713aa35bcb24b86ad0b58ca9fc991ba). Parent ref: group-v274445. {{(pid=61957) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 830.812114] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-47642d83-3473-4b34-8e43-c1548504523c tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] [instance: 615dbbed-2b02-4351-9e03-8c13f424a133] Unregistered the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 830.812345] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-47642d83-3473-4b34-8e43-c1548504523c tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] [instance: 615dbbed-2b02-4351-9e03-8c13f424a133] Deleting contents of the VM from datastore datastore2 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 830.812526] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-47642d83-3473-4b34-8e43-c1548504523c tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] Deleting the datastore file [datastore2] 615dbbed-2b02-4351-9e03-8c13f424a133 {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 830.815922] env[61957]: DEBUG nova.compute.manager [req-b9471571-748b-42fe-a6cc-0b59509a9565 req-dff2e161-de29-4710-b6b5-3f027efee897 service nova] [instance: 4951132e-7247-4772-8f88-3664c6a7e61e] Received event network-changed-3016c8eb-2ad3-40b1-8d88-eaec447de9ea {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 830.816115] env[61957]: DEBUG nova.compute.manager [req-b9471571-748b-42fe-a6cc-0b59509a9565 req-dff2e161-de29-4710-b6b5-3f027efee897 service nova] [instance: 4951132e-7247-4772-8f88-3664c6a7e61e] Refreshing instance network info cache due to event network-changed-3016c8eb-2ad3-40b1-8d88-eaec447de9ea. {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 830.816336] env[61957]: DEBUG oslo_concurrency.lockutils [req-b9471571-748b-42fe-a6cc-0b59509a9565 req-dff2e161-de29-4710-b6b5-3f027efee897 service nova] Acquiring lock "refresh_cache-4951132e-7247-4772-8f88-3664c6a7e61e" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 830.816480] env[61957]: DEBUG oslo_concurrency.lockutils [req-b9471571-748b-42fe-a6cc-0b59509a9565 req-dff2e161-de29-4710-b6b5-3f027efee897 service nova] Acquired lock "refresh_cache-4951132e-7247-4772-8f88-3664c6a7e61e" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 830.816639] env[61957]: DEBUG nova.network.neutron [req-b9471571-748b-42fe-a6cc-0b59509a9565 req-dff2e161-de29-4710-b6b5-3f027efee897 service nova] [instance: 4951132e-7247-4772-8f88-3664c6a7e61e] Refreshing network info cache for port 3016c8eb-2ad3-40b1-8d88-eaec447de9ea {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 830.818411] env[61957]: INFO nova.compute.manager [None req-36e797a2-06a5-481b-9dd9-66f4ec9dadf3 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 14ba830d-4e2c-4e9a-a059-3c86209f0127] Took 29.44 seconds to build instance. [ 830.822277] env[61957]: DEBUG nova.scheduler.client.report [None req-be584a65-2286-47ec-9ff6-29a7835a7d50 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 830.824948] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-11be6f49-8edb-4f7b-acd2-ee5186cef008 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.826685] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6133b6d5-29d0-41c2-983a-6516ecc3bc08 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.836577] env[61957]: DEBUG oslo_vmware.api [None req-5a8f4b06-9979-4e23-adff-500ed36e9a0d tempest-ServerGroupTestJSON-950728618 tempest-ServerGroupTestJSON-950728618-project-member] Task: {'id': task-1277443, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.204329} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.840435] env[61957]: DEBUG nova.virt.hardware [None req-3c3df299-fd1f-460a-8cc2-979af1db30d1 tempest-ServersTestJSON-1093633677 tempest-ServersTestJSON-1093633677-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T17:22:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T17:21:50Z,direct_url=,disk_format='vmdk',id=11c76a2c-f705-470a-ba9d-4657858bab38,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='81d3096d0f094373913fc3dc8f5c3c6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T17:21:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 830.840798] env[61957]: DEBUG nova.virt.hardware [None req-3c3df299-fd1f-460a-8cc2-979af1db30d1 tempest-ServersTestJSON-1093633677 tempest-ServersTestJSON-1093633677-project-member] Flavor limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 830.840845] env[61957]: DEBUG nova.virt.hardware [None req-3c3df299-fd1f-460a-8cc2-979af1db30d1 tempest-ServersTestJSON-1093633677 tempest-ServersTestJSON-1093633677-project-member] Image limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 830.841019] env[61957]: DEBUG nova.virt.hardware [None req-3c3df299-fd1f-460a-8cc2-979af1db30d1 tempest-ServersTestJSON-1093633677 tempest-ServersTestJSON-1093633677-project-member] Flavor pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 830.841185] env[61957]: DEBUG nova.virt.hardware [None req-3c3df299-fd1f-460a-8cc2-979af1db30d1 tempest-ServersTestJSON-1093633677 tempest-ServersTestJSON-1093633677-project-member] Image pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 830.841371] env[61957]: DEBUG nova.virt.hardware [None req-3c3df299-fd1f-460a-8cc2-979af1db30d1 tempest-ServersTestJSON-1093633677 tempest-ServersTestJSON-1093633677-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 830.841536] env[61957]: DEBUG nova.virt.hardware [None req-3c3df299-fd1f-460a-8cc2-979af1db30d1 tempest-ServersTestJSON-1093633677 tempest-ServersTestJSON-1093633677-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 830.841699] env[61957]: DEBUG nova.virt.hardware [None req-3c3df299-fd1f-460a-8cc2-979af1db30d1 tempest-ServersTestJSON-1093633677 tempest-ServersTestJSON-1093633677-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 830.841863] env[61957]: DEBUG nova.virt.hardware [None req-3c3df299-fd1f-460a-8cc2-979af1db30d1 tempest-ServersTestJSON-1093633677 tempest-ServersTestJSON-1093633677-project-member] Got 1 possible topologies {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 830.842036] env[61957]: DEBUG nova.virt.hardware [None req-3c3df299-fd1f-460a-8cc2-979af1db30d1 tempest-ServersTestJSON-1093633677 tempest-ServersTestJSON-1093633677-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 830.842214] env[61957]: DEBUG nova.virt.hardware [None req-3c3df299-fd1f-460a-8cc2-979af1db30d1 tempest-ServersTestJSON-1093633677 tempest-ServersTestJSON-1093633677-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 830.845437] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-5a8f4b06-9979-4e23-adff-500ed36e9a0d tempest-ServerGroupTestJSON-950728618 tempest-ServerGroupTestJSON-950728618-project-member] Deleted the datastore file {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 830.845629] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-5a8f4b06-9979-4e23-adff-500ed36e9a0d tempest-ServerGroupTestJSON-950728618 tempest-ServerGroupTestJSON-950728618-project-member] [instance: e9ac118f-08b3-430b-848c-461c2b2e3e02] Deleted contents of the VM from datastore datastore1 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 830.845803] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-5a8f4b06-9979-4e23-adff-500ed36e9a0d tempest-ServerGroupTestJSON-950728618 tempest-ServerGroupTestJSON-950728618-project-member] [instance: e9ac118f-08b3-430b-848c-461c2b2e3e02] Instance destroyed {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 830.845969] env[61957]: INFO nova.compute.manager [None req-5a8f4b06-9979-4e23-adff-500ed36e9a0d tempest-ServerGroupTestJSON-950728618 tempest-ServerGroupTestJSON-950728618-project-member] [instance: e9ac118f-08b3-430b-848c-461c2b2e3e02] Took 2.05 seconds to destroy the instance on the hypervisor. [ 830.846228] env[61957]: DEBUG oslo.service.loopingcall [None req-5a8f4b06-9979-4e23-adff-500ed36e9a0d tempest-ServerGroupTestJSON-950728618 tempest-ServerGroupTestJSON-950728618-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 830.849213] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b768a91-ebc7-4aa1-949c-25453a22d9fa {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.851371] env[61957]: DEBUG oslo_vmware.api [None req-9387ee6a-989e-4086-be03-e20e994ba789 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Task: {'id': task-1277444, 'name': Rename_Task, 'duration_secs': 0.17293} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.854203] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ac5c4b48-cfd7-420e-9aa1-225385350d11 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 830.854469] env[61957]: DEBUG nova.compute.manager [-] [instance: e9ac118f-08b3-430b-848c-461c2b2e3e02] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 830.854568] env[61957]: DEBUG nova.network.neutron [-] [instance: e9ac118f-08b3-430b-848c-461c2b2e3e02] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 830.856274] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-9387ee6a-989e-4086-be03-e20e994ba789 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] [instance: 17f00b20-9d3b-45e6-919d-6fab9999ec77] Powering on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 830.856561] env[61957]: INFO nova.virt.vmwareapi.vm_util [None req-7bbc9489-a498-401d-a2af-2f28d461499e tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Created folder: Project (8713aa35bcb24b86ad0b58ca9fc991ba) in parent group-v274445. [ 830.856747] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-7bbc9489-a498-401d-a2af-2f28d461499e tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Creating folder: Instances. Parent ref: group-v274489. {{(pid=61957) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 830.858609] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-202dca2d-c14a-420e-8e02-aacfc9ed0c0a {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.860177] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e7fa9183-c84b-4268-a7cf-a9d1a94d2c7d {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.861771] env[61957]: DEBUG oslo_vmware.api [None req-47642d83-3473-4b34-8e43-c1548504523c tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] Waiting for the task: (returnval){ [ 830.861771] env[61957]: value = "task-1277447" [ 830.861771] env[61957]: _type = "Task" [ 830.861771] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.869137] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c709db85-bf99-48b7-92cf-dda3b56a6784 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.875302] env[61957]: DEBUG oslo_vmware.api [None req-9387ee6a-989e-4086-be03-e20e994ba789 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Waiting for the task: (returnval){ [ 830.875302] env[61957]: value = "task-1277448" [ 830.875302] env[61957]: _type = "Task" [ 830.875302] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.880109] env[61957]: INFO nova.virt.vmwareapi.vm_util [None req-7bbc9489-a498-401d-a2af-2f28d461499e tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Created folder: Instances in parent group-v274489. [ 830.880376] env[61957]: DEBUG oslo.service.loopingcall [None req-7bbc9489-a498-401d-a2af-2f28d461499e tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 830.894046] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4951132e-7247-4772-8f88-3664c6a7e61e] Creating VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 830.894972] env[61957]: DEBUG oslo_vmware.api [None req-9387ee6a-989e-4086-be03-e20e994ba789 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Task: {'id': task-1277448, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.895289] env[61957]: DEBUG oslo_vmware.api [None req-47642d83-3473-4b34-8e43-c1548504523c tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] Task: {'id': task-1277447, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.895543] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e433974b-adf5-40f4-ac58-e8396e25132a {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.914604] env[61957]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 830.914604] env[61957]: value = "task-1277450" [ 830.914604] env[61957]: _type = "Task" [ 830.914604] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.922698] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277450, 'name': CreateVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.180564] env[61957]: DEBUG nova.network.neutron [None req-3c3df299-fd1f-460a-8cc2-979af1db30d1 tempest-ServersTestJSON-1093633677 tempest-ServersTestJSON-1093633677-project-member] [instance: ead6aae1-36b5-4f57-9129-3bb02cf103ce] Successfully updated port: 606b786e-f948-4d1c-a109-6390d9349ea8 {{(pid=61957) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 831.324946] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d61a7353-b862-4c7b-8175-41d701129e22 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Acquiring lock "14ba830d-4e2c-4e9a-a059-3c86209f0127" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 831.330974] env[61957]: DEBUG oslo_concurrency.lockutils [None req-be584a65-2286-47ec-9ff6-29a7835a7d50 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.932s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 831.331679] env[61957]: DEBUG nova.compute.manager [None req-be584a65-2286-47ec-9ff6-29a7835a7d50 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: 37ce46a1-8f9a-4d15-bd81-e40845a0e48a] Start building networks asynchronously for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 831.335172] env[61957]: DEBUG oslo_concurrency.lockutils [None req-36e797a2-06a5-481b-9dd9-66f4ec9dadf3 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Lock "14ba830d-4e2c-4e9a-a059-3c86209f0127" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 115.180s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 831.335172] env[61957]: DEBUG oslo_concurrency.lockutils [None req-227f29c0-ef6f-447c-b621-3f65bc420b4e tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.978s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 831.337110] env[61957]: INFO nova.compute.claims [None req-227f29c0-ef6f-447c-b621-3f65bc420b4e tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] [instance: 37a9e1c1-9f6c-4047-9ff5-e141d0ca383e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 831.342230] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d61a7353-b862-4c7b-8175-41d701129e22 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Lock "14ba830d-4e2c-4e9a-a059-3c86209f0127" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.018s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 831.342230] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d61a7353-b862-4c7b-8175-41d701129e22 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Acquiring lock "14ba830d-4e2c-4e9a-a059-3c86209f0127-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 831.342230] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d61a7353-b862-4c7b-8175-41d701129e22 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Lock "14ba830d-4e2c-4e9a-a059-3c86209f0127-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 831.342230] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d61a7353-b862-4c7b-8175-41d701129e22 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Lock "14ba830d-4e2c-4e9a-a059-3c86209f0127-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 831.344009] env[61957]: INFO nova.compute.manager [None req-d61a7353-b862-4c7b-8175-41d701129e22 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 14ba830d-4e2c-4e9a-a059-3c86209f0127] Terminating instance [ 831.346092] env[61957]: DEBUG nova.compute.manager [None req-d61a7353-b862-4c7b-8175-41d701129e22 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 14ba830d-4e2c-4e9a-a059-3c86209f0127] Start destroying the instance on the hypervisor. {{(pid=61957) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 831.346276] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-d61a7353-b862-4c7b-8175-41d701129e22 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 14ba830d-4e2c-4e9a-a059-3c86209f0127] Destroying instance {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 831.347106] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a28b9b6c-fb6e-437e-959c-67c1bb628c18 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.355102] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-d61a7353-b862-4c7b-8175-41d701129e22 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 14ba830d-4e2c-4e9a-a059-3c86209f0127] Powering off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 831.355394] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4cf3f16d-f136-444b-ac8f-4fb9be9d36da {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.362354] env[61957]: DEBUG oslo_vmware.api [None req-d61a7353-b862-4c7b-8175-41d701129e22 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Waiting for the task: (returnval){ [ 831.362354] env[61957]: value = "task-1277451" [ 831.362354] env[61957]: _type = "Task" [ 831.362354] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.376059] env[61957]: DEBUG oslo_vmware.api [None req-d61a7353-b862-4c7b-8175-41d701129e22 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1277451, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.382937] env[61957]: DEBUG oslo_vmware.api [None req-47642d83-3473-4b34-8e43-c1548504523c tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] Task: {'id': task-1277447, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.248439} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.386499] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-47642d83-3473-4b34-8e43-c1548504523c tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] Deleted the datastore file {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 831.386732] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-47642d83-3473-4b34-8e43-c1548504523c tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] [instance: 615dbbed-2b02-4351-9e03-8c13f424a133] Deleted contents of the VM from datastore datastore2 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 831.386915] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-47642d83-3473-4b34-8e43-c1548504523c tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] [instance: 615dbbed-2b02-4351-9e03-8c13f424a133] Instance destroyed {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 831.387102] env[61957]: INFO nova.compute.manager [None req-47642d83-3473-4b34-8e43-c1548504523c tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] [instance: 615dbbed-2b02-4351-9e03-8c13f424a133] Took 2.06 seconds to destroy the instance on the hypervisor. [ 831.387380] env[61957]: DEBUG oslo.service.loopingcall [None req-47642d83-3473-4b34-8e43-c1548504523c tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 831.387636] env[61957]: DEBUG nova.compute.manager [-] [instance: 615dbbed-2b02-4351-9e03-8c13f424a133] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 831.387704] env[61957]: DEBUG nova.network.neutron [-] [instance: 615dbbed-2b02-4351-9e03-8c13f424a133] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 831.397082] env[61957]: DEBUG oslo_vmware.api [None req-9387ee6a-989e-4086-be03-e20e994ba789 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Task: {'id': task-1277448, 'name': PowerOnVM_Task} progress is 64%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.426811] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277450, 'name': CreateVM_Task, 'duration_secs': 0.321734} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.427028] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4951132e-7247-4772-8f88-3664c6a7e61e] Created VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 831.427685] env[61957]: DEBUG oslo_concurrency.lockutils [None req-7bbc9489-a498-401d-a2af-2f28d461499e tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 831.427859] env[61957]: DEBUG oslo_concurrency.lockutils [None req-7bbc9489-a498-401d-a2af-2f28d461499e tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Acquired lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 831.428371] env[61957]: DEBUG oslo_concurrency.lockutils [None req-7bbc9489-a498-401d-a2af-2f28d461499e tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 831.428473] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b169e380-8b7f-4551-a19e-73c3af8fff93 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.434261] env[61957]: DEBUG oslo_vmware.api [None req-7bbc9489-a498-401d-a2af-2f28d461499e tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Waiting for the task: (returnval){ [ 831.434261] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52270b4d-57ce-1625-55f6-c7d7ce7b4d69" [ 831.434261] env[61957]: _type = "Task" [ 831.434261] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.444073] env[61957]: DEBUG oslo_vmware.api [None req-7bbc9489-a498-401d-a2af-2f28d461499e tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52270b4d-57ce-1625-55f6-c7d7ce7b4d69, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.682979] env[61957]: DEBUG oslo_concurrency.lockutils [None req-3c3df299-fd1f-460a-8cc2-979af1db30d1 tempest-ServersTestJSON-1093633677 tempest-ServersTestJSON-1093633677-project-member] Acquiring lock "refresh_cache-ead6aae1-36b5-4f57-9129-3bb02cf103ce" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 831.683183] env[61957]: DEBUG oslo_concurrency.lockutils [None req-3c3df299-fd1f-460a-8cc2-979af1db30d1 tempest-ServersTestJSON-1093633677 tempest-ServersTestJSON-1093633677-project-member] Acquired lock "refresh_cache-ead6aae1-36b5-4f57-9129-3bb02cf103ce" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 831.683340] env[61957]: DEBUG nova.network.neutron [None req-3c3df299-fd1f-460a-8cc2-979af1db30d1 tempest-ServersTestJSON-1093633677 tempest-ServersTestJSON-1093633677-project-member] [instance: ead6aae1-36b5-4f57-9129-3bb02cf103ce] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 831.702822] env[61957]: DEBUG nova.network.neutron [req-b9471571-748b-42fe-a6cc-0b59509a9565 req-dff2e161-de29-4710-b6b5-3f027efee897 service nova] [instance: 4951132e-7247-4772-8f88-3664c6a7e61e] Updated VIF entry in instance network info cache for port 3016c8eb-2ad3-40b1-8d88-eaec447de9ea. {{(pid=61957) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 831.703282] env[61957]: DEBUG nova.network.neutron [req-b9471571-748b-42fe-a6cc-0b59509a9565 req-dff2e161-de29-4710-b6b5-3f027efee897 service nova] [instance: 4951132e-7247-4772-8f88-3664c6a7e61e] Updating instance_info_cache with network_info: [{"id": "3016c8eb-2ad3-40b1-8d88-eaec447de9ea", "address": "fa:16:3e:59:19:f2", "network": {"id": "364cabcb-046d-4f91-b1ee-5dd3adfae6d7", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1432396967-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8713aa35bcb24b86ad0b58ca9fc991ba", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b7a73c01-1bb9-4612-a1a7-16d71b732e81", "external-id": "nsx-vlan-transportzone-711", "segmentation_id": 711, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3016c8eb-2a", "ovs_interfaceid": "3016c8eb-2ad3-40b1-8d88-eaec447de9ea", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 831.765050] env[61957]: DEBUG nova.network.neutron [-] [instance: e9ac118f-08b3-430b-848c-461c2b2e3e02] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 831.841988] env[61957]: DEBUG nova.compute.utils [None req-be584a65-2286-47ec-9ff6-29a7835a7d50 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Using /dev/sd instead of None {{(pid=61957) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 831.847027] env[61957]: DEBUG nova.compute.manager [None req-be584a65-2286-47ec-9ff6-29a7835a7d50 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: 37ce46a1-8f9a-4d15-bd81-e40845a0e48a] Allocating IP information in the background. {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 831.847027] env[61957]: DEBUG nova.network.neutron [None req-be584a65-2286-47ec-9ff6-29a7835a7d50 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: 37ce46a1-8f9a-4d15-bd81-e40845a0e48a] allocate_for_instance() {{(pid=61957) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 831.850362] env[61957]: DEBUG nova.compute.manager [None req-09e8eb69-1030-4228-8cb0-915d5cee1a52 tempest-AttachInterfacesV270Test-2032541732 tempest-AttachInterfacesV270Test-2032541732-project-member] [instance: 281b9644-0e7d-48b9-a7b6-45fd6102d558] Starting instance... {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 831.873494] env[61957]: DEBUG oslo_vmware.api [None req-d61a7353-b862-4c7b-8175-41d701129e22 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1277451, 'name': PowerOffVM_Task, 'duration_secs': 0.201876} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.873954] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-d61a7353-b862-4c7b-8175-41d701129e22 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 14ba830d-4e2c-4e9a-a059-3c86209f0127] Powered off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 831.875698] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-d61a7353-b862-4c7b-8175-41d701129e22 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 14ba830d-4e2c-4e9a-a059-3c86209f0127] Unregistering the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 831.876323] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-750520e4-f230-44ce-9781-16c72b2cbb95 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.892171] env[61957]: DEBUG oslo_vmware.api [None req-9387ee6a-989e-4086-be03-e20e994ba789 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Task: {'id': task-1277448, 'name': PowerOnVM_Task} progress is 91%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.924141] env[61957]: DEBUG nova.policy [None req-be584a65-2286-47ec-9ff6-29a7835a7d50 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3136bf0d1db3411fb1ff1ebe9c6045f6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1dbdc699741a48af93c52068d88a5357', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61957) authorize /opt/stack/nova/nova/policy.py:203}} [ 831.939892] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-d61a7353-b862-4c7b-8175-41d701129e22 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 14ba830d-4e2c-4e9a-a059-3c86209f0127] Unregistered the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 831.940205] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-d61a7353-b862-4c7b-8175-41d701129e22 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 14ba830d-4e2c-4e9a-a059-3c86209f0127] Deleting contents of the VM from datastore datastore2 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 831.940368] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-d61a7353-b862-4c7b-8175-41d701129e22 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Deleting the datastore file [datastore2] 14ba830d-4e2c-4e9a-a059-3c86209f0127 {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 831.944132] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4ea889c8-7573-4c30-8674-e5769afdefb2 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.951836] env[61957]: DEBUG oslo_vmware.api [None req-7bbc9489-a498-401d-a2af-2f28d461499e tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52270b4d-57ce-1625-55f6-c7d7ce7b4d69, 'name': SearchDatastore_Task, 'duration_secs': 0.01132} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.953765] env[61957]: DEBUG oslo_concurrency.lockutils [None req-7bbc9489-a498-401d-a2af-2f28d461499e tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Releasing lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 831.954020] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-7bbc9489-a498-401d-a2af-2f28d461499e tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 4951132e-7247-4772-8f88-3664c6a7e61e] Processing image 11c76a2c-f705-470a-ba9d-4657858bab38 {{(pid=61957) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 831.954259] env[61957]: DEBUG oslo_concurrency.lockutils [None req-7bbc9489-a498-401d-a2af-2f28d461499e tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 831.954409] env[61957]: DEBUG oslo_concurrency.lockutils [None req-7bbc9489-a498-401d-a2af-2f28d461499e tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Acquired lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 831.954587] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-7bbc9489-a498-401d-a2af-2f28d461499e tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 831.954904] env[61957]: DEBUG oslo_vmware.api [None req-d61a7353-b862-4c7b-8175-41d701129e22 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Waiting for the task: (returnval){ [ 831.954904] env[61957]: value = "task-1277453" [ 831.954904] env[61957]: _type = "Task" [ 831.954904] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.955112] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-101a9957-0649-42b4-9b6c-b0fe20d17eb0 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.976217] env[61957]: DEBUG oslo_vmware.api [None req-d61a7353-b862-4c7b-8175-41d701129e22 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1277453, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.976217] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-7bbc9489-a498-401d-a2af-2f28d461499e tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 831.976217] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-7bbc9489-a498-401d-a2af-2f28d461499e tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61957) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 831.976570] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0dd598e4-9efd-4c46-862d-ab4cd823cae4 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.983028] env[61957]: DEBUG oslo_vmware.api [None req-7bbc9489-a498-401d-a2af-2f28d461499e tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Waiting for the task: (returnval){ [ 831.983028] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52b09cdc-22c9-7c3e-4cb9-9f8733512591" [ 831.983028] env[61957]: _type = "Task" [ 831.983028] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.989980] env[61957]: DEBUG oslo_vmware.api [None req-7bbc9489-a498-401d-a2af-2f28d461499e tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52b09cdc-22c9-7c3e-4cb9-9f8733512591, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.206053] env[61957]: DEBUG oslo_concurrency.lockutils [req-b9471571-748b-42fe-a6cc-0b59509a9565 req-dff2e161-de29-4710-b6b5-3f027efee897 service nova] Releasing lock "refresh_cache-4951132e-7247-4772-8f88-3664c6a7e61e" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 832.229388] env[61957]: DEBUG nova.network.neutron [None req-3c3df299-fd1f-460a-8cc2-979af1db30d1 tempest-ServersTestJSON-1093633677 tempest-ServersTestJSON-1093633677-project-member] [instance: ead6aae1-36b5-4f57-9129-3bb02cf103ce] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 832.267384] env[61957]: INFO nova.compute.manager [-] [instance: e9ac118f-08b3-430b-848c-461c2b2e3e02] Took 1.41 seconds to deallocate network for instance. [ 832.350015] env[61957]: DEBUG nova.compute.manager [None req-be584a65-2286-47ec-9ff6-29a7835a7d50 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: 37ce46a1-8f9a-4d15-bd81-e40845a0e48a] Start building block device mappings for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 832.352511] env[61957]: DEBUG nova.network.neutron [None req-be584a65-2286-47ec-9ff6-29a7835a7d50 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: 37ce46a1-8f9a-4d15-bd81-e40845a0e48a] Successfully created port: 7234f796-0abc-4e81-ba04-29607283618c {{(pid=61957) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 832.395054] env[61957]: DEBUG oslo_concurrency.lockutils [None req-09e8eb69-1030-4228-8cb0-915d5cee1a52 tempest-AttachInterfacesV270Test-2032541732 tempest-AttachInterfacesV270Test-2032541732-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 832.405816] env[61957]: DEBUG oslo_vmware.api [None req-9387ee6a-989e-4086-be03-e20e994ba789 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Task: {'id': task-1277448, 'name': PowerOnVM_Task, 'duration_secs': 1.048066} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.406213] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-9387ee6a-989e-4086-be03-e20e994ba789 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] [instance: 17f00b20-9d3b-45e6-919d-6fab9999ec77] Powered on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 832.406766] env[61957]: INFO nova.compute.manager [None req-9387ee6a-989e-4086-be03-e20e994ba789 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] [instance: 17f00b20-9d3b-45e6-919d-6fab9999ec77] Took 7.02 seconds to spawn the instance on the hypervisor. [ 832.406766] env[61957]: DEBUG nova.compute.manager [None req-9387ee6a-989e-4086-be03-e20e994ba789 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] [instance: 17f00b20-9d3b-45e6-919d-6fab9999ec77] Checking state {{(pid=61957) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 832.409817] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0df011f-60d6-4fa8-9ee5-5e92773b2573 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.471137] env[61957]: DEBUG oslo_vmware.api [None req-d61a7353-b862-4c7b-8175-41d701129e22 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1277453, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.231907} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.471525] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-d61a7353-b862-4c7b-8175-41d701129e22 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Deleted the datastore file {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 832.471657] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-d61a7353-b862-4c7b-8175-41d701129e22 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 14ba830d-4e2c-4e9a-a059-3c86209f0127] Deleted contents of the VM from datastore datastore2 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 832.471835] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-d61a7353-b862-4c7b-8175-41d701129e22 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 14ba830d-4e2c-4e9a-a059-3c86209f0127] Instance destroyed {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 832.472034] env[61957]: INFO nova.compute.manager [None req-d61a7353-b862-4c7b-8175-41d701129e22 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 14ba830d-4e2c-4e9a-a059-3c86209f0127] Took 1.13 seconds to destroy the instance on the hypervisor. [ 832.472310] env[61957]: DEBUG oslo.service.loopingcall [None req-d61a7353-b862-4c7b-8175-41d701129e22 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 832.474976] env[61957]: DEBUG nova.compute.manager [-] [instance: 14ba830d-4e2c-4e9a-a059-3c86209f0127] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 832.475039] env[61957]: DEBUG nova.network.neutron [-] [instance: 14ba830d-4e2c-4e9a-a059-3c86209f0127] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 832.493917] env[61957]: DEBUG oslo_vmware.api [None req-7bbc9489-a498-401d-a2af-2f28d461499e tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52b09cdc-22c9-7c3e-4cb9-9f8733512591, 'name': SearchDatastore_Task, 'duration_secs': 0.010286} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.497174] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7bd6b5c3-d68c-448c-837d-561a1aea9c5d {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.501435] env[61957]: DEBUG nova.compute.manager [req-281fef2d-8c0f-4d22-a9ec-1f6fc54ea0f4 req-0737baf9-2e1b-432e-ab35-c708e26c4298 service nova] [instance: ead6aae1-36b5-4f57-9129-3bb02cf103ce] Received event network-vif-plugged-606b786e-f948-4d1c-a109-6390d9349ea8 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 832.501630] env[61957]: DEBUG oslo_concurrency.lockutils [req-281fef2d-8c0f-4d22-a9ec-1f6fc54ea0f4 req-0737baf9-2e1b-432e-ab35-c708e26c4298 service nova] Acquiring lock "ead6aae1-36b5-4f57-9129-3bb02cf103ce-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 832.501839] env[61957]: DEBUG oslo_concurrency.lockutils [req-281fef2d-8c0f-4d22-a9ec-1f6fc54ea0f4 req-0737baf9-2e1b-432e-ab35-c708e26c4298 service nova] Lock "ead6aae1-36b5-4f57-9129-3bb02cf103ce-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 832.502016] env[61957]: DEBUG oslo_concurrency.lockutils [req-281fef2d-8c0f-4d22-a9ec-1f6fc54ea0f4 req-0737baf9-2e1b-432e-ab35-c708e26c4298 service nova] Lock "ead6aae1-36b5-4f57-9129-3bb02cf103ce-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 832.502188] env[61957]: DEBUG nova.compute.manager [req-281fef2d-8c0f-4d22-a9ec-1f6fc54ea0f4 req-0737baf9-2e1b-432e-ab35-c708e26c4298 service nova] [instance: ead6aae1-36b5-4f57-9129-3bb02cf103ce] No waiting events found dispatching network-vif-plugged-606b786e-f948-4d1c-a109-6390d9349ea8 {{(pid=61957) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 832.502354] env[61957]: WARNING nova.compute.manager [req-281fef2d-8c0f-4d22-a9ec-1f6fc54ea0f4 req-0737baf9-2e1b-432e-ab35-c708e26c4298 service nova] [instance: ead6aae1-36b5-4f57-9129-3bb02cf103ce] Received unexpected event network-vif-plugged-606b786e-f948-4d1c-a109-6390d9349ea8 for instance with vm_state building and task_state spawning. [ 832.502510] env[61957]: DEBUG nova.compute.manager [req-281fef2d-8c0f-4d22-a9ec-1f6fc54ea0f4 req-0737baf9-2e1b-432e-ab35-c708e26c4298 service nova] [instance: ead6aae1-36b5-4f57-9129-3bb02cf103ce] Received event network-changed-606b786e-f948-4d1c-a109-6390d9349ea8 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 832.502659] env[61957]: DEBUG nova.compute.manager [req-281fef2d-8c0f-4d22-a9ec-1f6fc54ea0f4 req-0737baf9-2e1b-432e-ab35-c708e26c4298 service nova] [instance: ead6aae1-36b5-4f57-9129-3bb02cf103ce] Refreshing instance network info cache due to event network-changed-606b786e-f948-4d1c-a109-6390d9349ea8. {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 832.502837] env[61957]: DEBUG oslo_concurrency.lockutils [req-281fef2d-8c0f-4d22-a9ec-1f6fc54ea0f4 req-0737baf9-2e1b-432e-ab35-c708e26c4298 service nova] Acquiring lock "refresh_cache-ead6aae1-36b5-4f57-9129-3bb02cf103ce" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 832.510268] env[61957]: DEBUG oslo_vmware.api [None req-7bbc9489-a498-401d-a2af-2f28d461499e tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Waiting for the task: (returnval){ [ 832.510268] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]5205a5b6-8336-a553-de33-6b8e56c44716" [ 832.510268] env[61957]: _type = "Task" [ 832.510268] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.518770] env[61957]: DEBUG oslo_vmware.api [None req-7bbc9489-a498-401d-a2af-2f28d461499e tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]5205a5b6-8336-a553-de33-6b8e56c44716, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.530323] env[61957]: DEBUG nova.network.neutron [None req-3c3df299-fd1f-460a-8cc2-979af1db30d1 tempest-ServersTestJSON-1093633677 tempest-ServersTestJSON-1093633677-project-member] [instance: ead6aae1-36b5-4f57-9129-3bb02cf103ce] Updating instance_info_cache with network_info: [{"id": "606b786e-f948-4d1c-a109-6390d9349ea8", "address": "fa:16:3e:fb:fd:15", "network": {"id": "200da3c5-aacb-42e6-b638-65442010d6c0", "bridge": "br-int", "label": "tempest-ServersTestJSON-1424742350-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1a127f9f057c43d2bddf261d2d875b63", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dacd109c-2442-41b8-b612-7ed3efbdaa94", "external-id": "nsx-vlan-transportzone-940", "segmentation_id": 940, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap606b786e-f9", "ovs_interfaceid": "606b786e-f948-4d1c-a109-6390d9349ea8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 832.605800] env[61957]: DEBUG nova.network.neutron [-] [instance: 615dbbed-2b02-4351-9e03-8c13f424a133] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 832.776398] env[61957]: DEBUG oslo_concurrency.lockutils [None req-5a8f4b06-9979-4e23-adff-500ed36e9a0d tempest-ServerGroupTestJSON-950728618 tempest-ServerGroupTestJSON-950728618-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 832.777292] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ce01cd8-2050-4ff3-8892-0b0454366117 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.785868] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e7ecaa7-c2be-46d2-a328-be71b4ad56d3 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.837479] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8994ecb5-c710-4098-b099-344f2f9fab24 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.846110] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-016796cc-e32a-41af-8af7-88257c7eb291 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.866466] env[61957]: DEBUG nova.compute.provider_tree [None req-227f29c0-ef6f-447c-b621-3f65bc420b4e tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 832.933932] env[61957]: INFO nova.compute.manager [None req-9387ee6a-989e-4086-be03-e20e994ba789 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] [instance: 17f00b20-9d3b-45e6-919d-6fab9999ec77] Took 27.22 seconds to build instance. [ 833.021132] env[61957]: DEBUG oslo_vmware.api [None req-7bbc9489-a498-401d-a2af-2f28d461499e tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]5205a5b6-8336-a553-de33-6b8e56c44716, 'name': SearchDatastore_Task, 'duration_secs': 0.012451} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.021482] env[61957]: DEBUG oslo_concurrency.lockutils [None req-7bbc9489-a498-401d-a2af-2f28d461499e tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Releasing lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 833.021690] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-7bbc9489-a498-401d-a2af-2f28d461499e tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore1] 4951132e-7247-4772-8f88-3664c6a7e61e/4951132e-7247-4772-8f88-3664c6a7e61e.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 833.021997] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d46ca39e-6ec6-44b0-a9bb-373e1750acf8 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.029035] env[61957]: DEBUG oslo_vmware.api [None req-7bbc9489-a498-401d-a2af-2f28d461499e tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Waiting for the task: (returnval){ [ 833.029035] env[61957]: value = "task-1277454" [ 833.029035] env[61957]: _type = "Task" [ 833.029035] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.036232] env[61957]: DEBUG oslo_concurrency.lockutils [None req-3c3df299-fd1f-460a-8cc2-979af1db30d1 tempest-ServersTestJSON-1093633677 tempest-ServersTestJSON-1093633677-project-member] Releasing lock "refresh_cache-ead6aae1-36b5-4f57-9129-3bb02cf103ce" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 833.036536] env[61957]: DEBUG nova.compute.manager [None req-3c3df299-fd1f-460a-8cc2-979af1db30d1 tempest-ServersTestJSON-1093633677 tempest-ServersTestJSON-1093633677-project-member] [instance: ead6aae1-36b5-4f57-9129-3bb02cf103ce] Instance network_info: |[{"id": "606b786e-f948-4d1c-a109-6390d9349ea8", "address": "fa:16:3e:fb:fd:15", "network": {"id": "200da3c5-aacb-42e6-b638-65442010d6c0", "bridge": "br-int", "label": "tempest-ServersTestJSON-1424742350-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1a127f9f057c43d2bddf261d2d875b63", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dacd109c-2442-41b8-b612-7ed3efbdaa94", "external-id": "nsx-vlan-transportzone-940", "segmentation_id": 940, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap606b786e-f9", "ovs_interfaceid": "606b786e-f948-4d1c-a109-6390d9349ea8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 833.036821] env[61957]: DEBUG oslo_vmware.api [None req-7bbc9489-a498-401d-a2af-2f28d461499e tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': task-1277454, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.037861] env[61957]: DEBUG oslo_concurrency.lockutils [req-281fef2d-8c0f-4d22-a9ec-1f6fc54ea0f4 req-0737baf9-2e1b-432e-ab35-c708e26c4298 service nova] Acquired lock "refresh_cache-ead6aae1-36b5-4f57-9129-3bb02cf103ce" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 833.037861] env[61957]: DEBUG nova.network.neutron [req-281fef2d-8c0f-4d22-a9ec-1f6fc54ea0f4 req-0737baf9-2e1b-432e-ab35-c708e26c4298 service nova] [instance: ead6aae1-36b5-4f57-9129-3bb02cf103ce] Refreshing network info cache for port 606b786e-f948-4d1c-a109-6390d9349ea8 {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 833.038439] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-3c3df299-fd1f-460a-8cc2-979af1db30d1 tempest-ServersTestJSON-1093633677 tempest-ServersTestJSON-1093633677-project-member] [instance: ead6aae1-36b5-4f57-9129-3bb02cf103ce] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:fb:fd:15', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'dacd109c-2442-41b8-b612-7ed3efbdaa94', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '606b786e-f948-4d1c-a109-6390d9349ea8', 'vif_model': 'vmxnet3'}] {{(pid=61957) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 833.045704] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c3df299-fd1f-460a-8cc2-979af1db30d1 tempest-ServersTestJSON-1093633677 tempest-ServersTestJSON-1093633677-project-member] Creating folder: Project (1a127f9f057c43d2bddf261d2d875b63). Parent ref: group-v274445. {{(pid=61957) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 833.046219] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0da70e9a-c466-4c15-96a4-db8bc83e7c8c {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.055938] env[61957]: INFO nova.virt.vmwareapi.vm_util [None req-3c3df299-fd1f-460a-8cc2-979af1db30d1 tempest-ServersTestJSON-1093633677 tempest-ServersTestJSON-1093633677-project-member] Created folder: Project (1a127f9f057c43d2bddf261d2d875b63) in parent group-v274445. [ 833.056148] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c3df299-fd1f-460a-8cc2-979af1db30d1 tempest-ServersTestJSON-1093633677 tempest-ServersTestJSON-1093633677-project-member] Creating folder: Instances. Parent ref: group-v274492. {{(pid=61957) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 833.056894] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9af0f7ca-b98e-4903-b201-e418911d9880 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.065305] env[61957]: INFO nova.virt.vmwareapi.vm_util [None req-3c3df299-fd1f-460a-8cc2-979af1db30d1 tempest-ServersTestJSON-1093633677 tempest-ServersTestJSON-1093633677-project-member] Created folder: Instances in parent group-v274492. [ 833.065542] env[61957]: DEBUG oslo.service.loopingcall [None req-3c3df299-fd1f-460a-8cc2-979af1db30d1 tempest-ServersTestJSON-1093633677 tempest-ServersTestJSON-1093633677-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 833.065738] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ead6aae1-36b5-4f57-9129-3bb02cf103ce] Creating VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 833.065948] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-026d35c3-c953-49b8-b787-23fd24baa424 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.084234] env[61957]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 833.084234] env[61957]: value = "task-1277457" [ 833.084234] env[61957]: _type = "Task" [ 833.084234] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.091860] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277457, 'name': CreateVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.110431] env[61957]: INFO nova.compute.manager [-] [instance: 615dbbed-2b02-4351-9e03-8c13f424a133] Took 1.72 seconds to deallocate network for instance. [ 833.369681] env[61957]: DEBUG nova.compute.manager [None req-be584a65-2286-47ec-9ff6-29a7835a7d50 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: 37ce46a1-8f9a-4d15-bd81-e40845a0e48a] Start spawning the instance on the hypervisor. {{(pid=61957) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 833.373084] env[61957]: DEBUG nova.scheduler.client.report [None req-227f29c0-ef6f-447c-b621-3f65bc420b4e tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 833.396035] env[61957]: DEBUG nova.network.neutron [-] [instance: 14ba830d-4e2c-4e9a-a059-3c86209f0127] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 833.401164] env[61957]: DEBUG nova.virt.hardware [None req-be584a65-2286-47ec-9ff6-29a7835a7d50 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T17:22:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T17:21:50Z,direct_url=,disk_format='vmdk',id=11c76a2c-f705-470a-ba9d-4657858bab38,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='81d3096d0f094373913fc3dc8f5c3c6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T17:21:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 833.401382] env[61957]: DEBUG nova.virt.hardware [None req-be584a65-2286-47ec-9ff6-29a7835a7d50 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Flavor limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 833.401580] env[61957]: DEBUG nova.virt.hardware [None req-be584a65-2286-47ec-9ff6-29a7835a7d50 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Image limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 833.401783] env[61957]: DEBUG nova.virt.hardware [None req-be584a65-2286-47ec-9ff6-29a7835a7d50 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Flavor pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 833.401939] env[61957]: DEBUG nova.virt.hardware [None req-be584a65-2286-47ec-9ff6-29a7835a7d50 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Image pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 833.402169] env[61957]: DEBUG nova.virt.hardware [None req-be584a65-2286-47ec-9ff6-29a7835a7d50 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 833.402513] env[61957]: DEBUG nova.virt.hardware [None req-be584a65-2286-47ec-9ff6-29a7835a7d50 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 833.402718] env[61957]: DEBUG nova.virt.hardware [None req-be584a65-2286-47ec-9ff6-29a7835a7d50 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 833.402919] env[61957]: DEBUG nova.virt.hardware [None req-be584a65-2286-47ec-9ff6-29a7835a7d50 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Got 1 possible topologies {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 833.403169] env[61957]: DEBUG nova.virt.hardware [None req-be584a65-2286-47ec-9ff6-29a7835a7d50 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 833.403443] env[61957]: DEBUG nova.virt.hardware [None req-be584a65-2286-47ec-9ff6-29a7835a7d50 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 833.404894] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af8f1875-5f85-404a-ac3c-9eea1834ae71 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.418315] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09b30b50-05f1-4cb7-8dc1-1e645e77d911 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.436940] env[61957]: DEBUG oslo_concurrency.lockutils [None req-9387ee6a-989e-4086-be03-e20e994ba789 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Lock "17f00b20-9d3b-45e6-919d-6fab9999ec77" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 107.544s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 833.541358] env[61957]: DEBUG oslo_vmware.api [None req-7bbc9489-a498-401d-a2af-2f28d461499e tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': task-1277454, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.595973] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277457, 'name': CreateVM_Task} progress is 99%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.617502] env[61957]: DEBUG oslo_concurrency.lockutils [None req-47642d83-3473-4b34-8e43-c1548504523c tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 833.780108] env[61957]: DEBUG nova.network.neutron [req-281fef2d-8c0f-4d22-a9ec-1f6fc54ea0f4 req-0737baf9-2e1b-432e-ab35-c708e26c4298 service nova] [instance: ead6aae1-36b5-4f57-9129-3bb02cf103ce] Updated VIF entry in instance network info cache for port 606b786e-f948-4d1c-a109-6390d9349ea8. {{(pid=61957) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 833.780496] env[61957]: DEBUG nova.network.neutron [req-281fef2d-8c0f-4d22-a9ec-1f6fc54ea0f4 req-0737baf9-2e1b-432e-ab35-c708e26c4298 service nova] [instance: ead6aae1-36b5-4f57-9129-3bb02cf103ce] Updating instance_info_cache with network_info: [{"id": "606b786e-f948-4d1c-a109-6390d9349ea8", "address": "fa:16:3e:fb:fd:15", "network": {"id": "200da3c5-aacb-42e6-b638-65442010d6c0", "bridge": "br-int", "label": "tempest-ServersTestJSON-1424742350-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1a127f9f057c43d2bddf261d2d875b63", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dacd109c-2442-41b8-b612-7ed3efbdaa94", "external-id": "nsx-vlan-transportzone-940", "segmentation_id": 940, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap606b786e-f9", "ovs_interfaceid": "606b786e-f948-4d1c-a109-6390d9349ea8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 833.881024] env[61957]: DEBUG oslo_concurrency.lockutils [None req-227f29c0-ef6f-447c-b621-3f65bc420b4e tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.543s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 833.881024] env[61957]: DEBUG nova.compute.manager [None req-227f29c0-ef6f-447c-b621-3f65bc420b4e tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] [instance: 37a9e1c1-9f6c-4047-9ff5-e141d0ca383e] Start building networks asynchronously for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 833.882828] env[61957]: DEBUG oslo_concurrency.lockutils [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 17.407s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 833.898626] env[61957]: INFO nova.compute.manager [-] [instance: 14ba830d-4e2c-4e9a-a059-3c86209f0127] Took 1.42 seconds to deallocate network for instance. [ 833.940416] env[61957]: DEBUG nova.compute.manager [None req-902199c7-5b38-4a0e-a88f-80539210637b tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] [instance: 526ff179-62a6-4763-ab25-797617c4ed57] Starting instance... {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 834.042640] env[61957]: DEBUG oslo_vmware.api [None req-7bbc9489-a498-401d-a2af-2f28d461499e tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': task-1277454, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.608366} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.042931] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-7bbc9489-a498-401d-a2af-2f28d461499e tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore1] 4951132e-7247-4772-8f88-3664c6a7e61e/4951132e-7247-4772-8f88-3664c6a7e61e.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 834.043176] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-7bbc9489-a498-401d-a2af-2f28d461499e tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 4951132e-7247-4772-8f88-3664c6a7e61e] Extending root virtual disk to 1048576 {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 834.043821] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-18a69662-cc38-4b7c-98cf-faa8a6d39c55 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.050746] env[61957]: DEBUG oslo_vmware.api [None req-7bbc9489-a498-401d-a2af-2f28d461499e tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Waiting for the task: (returnval){ [ 834.050746] env[61957]: value = "task-1277458" [ 834.050746] env[61957]: _type = "Task" [ 834.050746] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.059079] env[61957]: DEBUG oslo_vmware.api [None req-7bbc9489-a498-401d-a2af-2f28d461499e tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': task-1277458, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.096180] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277457, 'name': CreateVM_Task, 'duration_secs': 0.582417} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.096420] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ead6aae1-36b5-4f57-9129-3bb02cf103ce] Created VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 834.097162] env[61957]: DEBUG oslo_concurrency.lockutils [None req-3c3df299-fd1f-460a-8cc2-979af1db30d1 tempest-ServersTestJSON-1093633677 tempest-ServersTestJSON-1093633677-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 834.097343] env[61957]: DEBUG oslo_concurrency.lockutils [None req-3c3df299-fd1f-460a-8cc2-979af1db30d1 tempest-ServersTestJSON-1093633677 tempest-ServersTestJSON-1093633677-project-member] Acquired lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 834.097867] env[61957]: DEBUG oslo_concurrency.lockutils [None req-3c3df299-fd1f-460a-8cc2-979af1db30d1 tempest-ServersTestJSON-1093633677 tempest-ServersTestJSON-1093633677-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 834.098152] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f1a4d402-77df-4c27-a7cd-4379cf2eb96e {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.103651] env[61957]: DEBUG oslo_vmware.api [None req-3c3df299-fd1f-460a-8cc2-979af1db30d1 tempest-ServersTestJSON-1093633677 tempest-ServersTestJSON-1093633677-project-member] Waiting for the task: (returnval){ [ 834.103651] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]5202e991-92eb-7450-80fc-527770146bb3" [ 834.103651] env[61957]: _type = "Task" [ 834.103651] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.114563] env[61957]: DEBUG oslo_vmware.api [None req-3c3df299-fd1f-460a-8cc2-979af1db30d1 tempest-ServersTestJSON-1093633677 tempest-ServersTestJSON-1093633677-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]5202e991-92eb-7450-80fc-527770146bb3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.121616] env[61957]: DEBUG nova.network.neutron [None req-be584a65-2286-47ec-9ff6-29a7835a7d50 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: 37ce46a1-8f9a-4d15-bd81-e40845a0e48a] Successfully updated port: 7234f796-0abc-4e81-ba04-29607283618c {{(pid=61957) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 834.283019] env[61957]: DEBUG oslo_concurrency.lockutils [req-281fef2d-8c0f-4d22-a9ec-1f6fc54ea0f4 req-0737baf9-2e1b-432e-ab35-c708e26c4298 service nova] Releasing lock "refresh_cache-ead6aae1-36b5-4f57-9129-3bb02cf103ce" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 834.283440] env[61957]: DEBUG nova.compute.manager [req-281fef2d-8c0f-4d22-a9ec-1f6fc54ea0f4 req-0737baf9-2e1b-432e-ab35-c708e26c4298 service nova] [instance: e9ac118f-08b3-430b-848c-461c2b2e3e02] Received event network-vif-deleted-d6407909-c811-4b71-9ac5-f6128992d3f0 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 834.283679] env[61957]: DEBUG nova.compute.manager [req-281fef2d-8c0f-4d22-a9ec-1f6fc54ea0f4 req-0737baf9-2e1b-432e-ab35-c708e26c4298 service nova] [instance: 615dbbed-2b02-4351-9e03-8c13f424a133] Received event network-vif-deleted-bbe103a8-3b31-4432-a8d8-34647bec0527 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 834.283874] env[61957]: INFO nova.compute.manager [req-281fef2d-8c0f-4d22-a9ec-1f6fc54ea0f4 req-0737baf9-2e1b-432e-ab35-c708e26c4298 service nova] [instance: 615dbbed-2b02-4351-9e03-8c13f424a133] Neutron deleted interface bbe103a8-3b31-4432-a8d8-34647bec0527; detaching it from the instance and deleting it from the info cache [ 834.284208] env[61957]: DEBUG nova.network.neutron [req-281fef2d-8c0f-4d22-a9ec-1f6fc54ea0f4 req-0737baf9-2e1b-432e-ab35-c708e26c4298 service nova] [instance: 615dbbed-2b02-4351-9e03-8c13f424a133] Updating instance_info_cache with network_info: [{"id": "31850256-5668-4fd1-9758-bf86c83c33e4", "address": "fa:16:3e:e0:80:a6", "network": {"id": "20aa8adb-b6a9-42c4-94c0-17673dfad27a", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-865860169", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.116", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8363e21fa47b46b8b2538b741564cb60", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "40859343-2baa-45fd-88e3-ebf8aaed2b19", "external-id": "nsx-vlan-transportzone-10", "segmentation_id": 10, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap31850256-56", "ovs_interfaceid": "31850256-5668-4fd1-9758-bf86c83c33e4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 834.311944] env[61957]: INFO nova.compute.manager [None req-82a5645a-fb19-4ad1-aa47-e12b7f048b77 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] [instance: 17f00b20-9d3b-45e6-919d-6fab9999ec77] Rebuilding instance [ 834.358360] env[61957]: DEBUG nova.compute.manager [None req-82a5645a-fb19-4ad1-aa47-e12b7f048b77 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] [instance: 17f00b20-9d3b-45e6-919d-6fab9999ec77] Checking state {{(pid=61957) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 834.359298] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77541dff-b357-4e14-a76a-90e3b4943189 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.386721] env[61957]: DEBUG nova.compute.utils [None req-227f29c0-ef6f-447c-b621-3f65bc420b4e tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] Using /dev/sd instead of None {{(pid=61957) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 834.393544] env[61957]: DEBUG nova.compute.manager [None req-227f29c0-ef6f-447c-b621-3f65bc420b4e tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] [instance: 37a9e1c1-9f6c-4047-9ff5-e141d0ca383e] Allocating IP information in the background. {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 834.393740] env[61957]: DEBUG nova.network.neutron [None req-227f29c0-ef6f-447c-b621-3f65bc420b4e tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] [instance: 37a9e1c1-9f6c-4047-9ff5-e141d0ca383e] allocate_for_instance() {{(pid=61957) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 834.405555] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d61a7353-b862-4c7b-8175-41d701129e22 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 834.443816] env[61957]: DEBUG nova.policy [None req-227f29c0-ef6f-447c-b621-3f65bc420b4e tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2908dce4d5134e158bbed442a09c1df9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '21351f437ea54c9580753dfddd7f47be', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61957) authorize /opt/stack/nova/nova/policy.py:203}} [ 834.462633] env[61957]: DEBUG oslo_concurrency.lockutils [None req-902199c7-5b38-4a0e-a88f-80539210637b tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 834.533967] env[61957]: DEBUG nova.compute.manager [req-d53dc9e8-04fb-472d-b49b-f034d99beb21 req-3734b50b-b5ce-4f59-ac1a-722f2491140f service nova] [instance: 14ba830d-4e2c-4e9a-a059-3c86209f0127] Received event network-vif-deleted-6b5aa7fe-7f6f-4305-b6b9-eef2a2c90d69 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 834.534300] env[61957]: DEBUG nova.compute.manager [req-d53dc9e8-04fb-472d-b49b-f034d99beb21 req-3734b50b-b5ce-4f59-ac1a-722f2491140f service nova] [instance: 37ce46a1-8f9a-4d15-bd81-e40845a0e48a] Received event network-vif-plugged-7234f796-0abc-4e81-ba04-29607283618c {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 834.534549] env[61957]: DEBUG oslo_concurrency.lockutils [req-d53dc9e8-04fb-472d-b49b-f034d99beb21 req-3734b50b-b5ce-4f59-ac1a-722f2491140f service nova] Acquiring lock "37ce46a1-8f9a-4d15-bd81-e40845a0e48a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 834.534857] env[61957]: DEBUG oslo_concurrency.lockutils [req-d53dc9e8-04fb-472d-b49b-f034d99beb21 req-3734b50b-b5ce-4f59-ac1a-722f2491140f service nova] Lock "37ce46a1-8f9a-4d15-bd81-e40845a0e48a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 834.535114] env[61957]: DEBUG oslo_concurrency.lockutils [req-d53dc9e8-04fb-472d-b49b-f034d99beb21 req-3734b50b-b5ce-4f59-ac1a-722f2491140f service nova] Lock "37ce46a1-8f9a-4d15-bd81-e40845a0e48a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 834.535324] env[61957]: DEBUG nova.compute.manager [req-d53dc9e8-04fb-472d-b49b-f034d99beb21 req-3734b50b-b5ce-4f59-ac1a-722f2491140f service nova] [instance: 37ce46a1-8f9a-4d15-bd81-e40845a0e48a] No waiting events found dispatching network-vif-plugged-7234f796-0abc-4e81-ba04-29607283618c {{(pid=61957) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 834.535583] env[61957]: WARNING nova.compute.manager [req-d53dc9e8-04fb-472d-b49b-f034d99beb21 req-3734b50b-b5ce-4f59-ac1a-722f2491140f service nova] [instance: 37ce46a1-8f9a-4d15-bd81-e40845a0e48a] Received unexpected event network-vif-plugged-7234f796-0abc-4e81-ba04-29607283618c for instance with vm_state building and task_state spawning. [ 834.535792] env[61957]: DEBUG nova.compute.manager [req-d53dc9e8-04fb-472d-b49b-f034d99beb21 req-3734b50b-b5ce-4f59-ac1a-722f2491140f service nova] [instance: 37ce46a1-8f9a-4d15-bd81-e40845a0e48a] Received event network-changed-7234f796-0abc-4e81-ba04-29607283618c {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 834.536016] env[61957]: DEBUG nova.compute.manager [req-d53dc9e8-04fb-472d-b49b-f034d99beb21 req-3734b50b-b5ce-4f59-ac1a-722f2491140f service nova] [instance: 37ce46a1-8f9a-4d15-bd81-e40845a0e48a] Refreshing instance network info cache due to event network-changed-7234f796-0abc-4e81-ba04-29607283618c. {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 834.536302] env[61957]: DEBUG oslo_concurrency.lockutils [req-d53dc9e8-04fb-472d-b49b-f034d99beb21 req-3734b50b-b5ce-4f59-ac1a-722f2491140f service nova] Acquiring lock "refresh_cache-37ce46a1-8f9a-4d15-bd81-e40845a0e48a" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 834.536477] env[61957]: DEBUG oslo_concurrency.lockutils [req-d53dc9e8-04fb-472d-b49b-f034d99beb21 req-3734b50b-b5ce-4f59-ac1a-722f2491140f service nova] Acquired lock "refresh_cache-37ce46a1-8f9a-4d15-bd81-e40845a0e48a" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 834.536672] env[61957]: DEBUG nova.network.neutron [req-d53dc9e8-04fb-472d-b49b-f034d99beb21 req-3734b50b-b5ce-4f59-ac1a-722f2491140f service nova] [instance: 37ce46a1-8f9a-4d15-bd81-e40845a0e48a] Refreshing network info cache for port 7234f796-0abc-4e81-ba04-29607283618c {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 834.560981] env[61957]: DEBUG oslo_vmware.api [None req-7bbc9489-a498-401d-a2af-2f28d461499e tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': task-1277458, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068282} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.561277] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-7bbc9489-a498-401d-a2af-2f28d461499e tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 4951132e-7247-4772-8f88-3664c6a7e61e] Extended root virtual disk {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 834.562160] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93208a8f-4f16-4af5-8ac0-47715a7e94a0 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.585501] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-7bbc9489-a498-401d-a2af-2f28d461499e tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 4951132e-7247-4772-8f88-3664c6a7e61e] Reconfiguring VM instance instance-00000038 to attach disk [datastore1] 4951132e-7247-4772-8f88-3664c6a7e61e/4951132e-7247-4772-8f88-3664c6a7e61e.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 834.586170] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-56b44086-afa1-4b56-a040-3b1c31728c1d {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.608592] env[61957]: DEBUG oslo_vmware.api [None req-7bbc9489-a498-401d-a2af-2f28d461499e tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Waiting for the task: (returnval){ [ 834.608592] env[61957]: value = "task-1277459" [ 834.608592] env[61957]: _type = "Task" [ 834.608592] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.615942] env[61957]: DEBUG oslo_vmware.api [None req-3c3df299-fd1f-460a-8cc2-979af1db30d1 tempest-ServersTestJSON-1093633677 tempest-ServersTestJSON-1093633677-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]5202e991-92eb-7450-80fc-527770146bb3, 'name': SearchDatastore_Task, 'duration_secs': 0.012504} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.616653] env[61957]: DEBUG oslo_concurrency.lockutils [None req-3c3df299-fd1f-460a-8cc2-979af1db30d1 tempest-ServersTestJSON-1093633677 tempest-ServersTestJSON-1093633677-project-member] Releasing lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 834.616955] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-3c3df299-fd1f-460a-8cc2-979af1db30d1 tempest-ServersTestJSON-1093633677 tempest-ServersTestJSON-1093633677-project-member] [instance: ead6aae1-36b5-4f57-9129-3bb02cf103ce] Processing image 11c76a2c-f705-470a-ba9d-4657858bab38 {{(pid=61957) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 834.617220] env[61957]: DEBUG oslo_concurrency.lockutils [None req-3c3df299-fd1f-460a-8cc2-979af1db30d1 tempest-ServersTestJSON-1093633677 tempest-ServersTestJSON-1093633677-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 834.617376] env[61957]: DEBUG oslo_concurrency.lockutils [None req-3c3df299-fd1f-460a-8cc2-979af1db30d1 tempest-ServersTestJSON-1093633677 tempest-ServersTestJSON-1093633677-project-member] Acquired lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 834.617609] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-3c3df299-fd1f-460a-8cc2-979af1db30d1 tempest-ServersTestJSON-1093633677 tempest-ServersTestJSON-1093633677-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 834.620927] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-105030a4-c85e-42e2-aeed-b10f63a91fe3 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.623668] env[61957]: DEBUG oslo_vmware.api [None req-7bbc9489-a498-401d-a2af-2f28d461499e tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': task-1277459, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.624233] env[61957]: DEBUG oslo_concurrency.lockutils [None req-be584a65-2286-47ec-9ff6-29a7835a7d50 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Acquiring lock "refresh_cache-37ce46a1-8f9a-4d15-bd81-e40845a0e48a" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 834.631383] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-3c3df299-fd1f-460a-8cc2-979af1db30d1 tempest-ServersTestJSON-1093633677 tempest-ServersTestJSON-1093633677-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 834.631602] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-3c3df299-fd1f-460a-8cc2-979af1db30d1 tempest-ServersTestJSON-1093633677 tempest-ServersTestJSON-1093633677-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61957) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 834.632402] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-92cfc21b-88de-4ee6-b837-e861a7293143 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.639903] env[61957]: DEBUG oslo_vmware.api [None req-3c3df299-fd1f-460a-8cc2-979af1db30d1 tempest-ServersTestJSON-1093633677 tempest-ServersTestJSON-1093633677-project-member] Waiting for the task: (returnval){ [ 834.639903] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]524143b8-faea-9b7c-7dba-c2e59361b23a" [ 834.639903] env[61957]: _type = "Task" [ 834.639903] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.648634] env[61957]: DEBUG oslo_vmware.api [None req-3c3df299-fd1f-460a-8cc2-979af1db30d1 tempest-ServersTestJSON-1093633677 tempest-ServersTestJSON-1093633677-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]524143b8-faea-9b7c-7dba-c2e59361b23a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.789313] env[61957]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c94adc40-e105-4887-989f-4cd0ece22580 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.803374] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34cf8c64-7505-461a-a155-8dc6fa073c6a {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.834243] env[61957]: DEBUG nova.compute.manager [req-281fef2d-8c0f-4d22-a9ec-1f6fc54ea0f4 req-0737baf9-2e1b-432e-ab35-c708e26c4298 service nova] [instance: 615dbbed-2b02-4351-9e03-8c13f424a133] Detach interface failed, port_id=bbe103a8-3b31-4432-a8d8-34647bec0527, reason: Instance 615dbbed-2b02-4351-9e03-8c13f424a133 could not be found. {{(pid=61957) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 834.834502] env[61957]: DEBUG nova.compute.manager [req-281fef2d-8c0f-4d22-a9ec-1f6fc54ea0f4 req-0737baf9-2e1b-432e-ab35-c708e26c4298 service nova] [instance: 615dbbed-2b02-4351-9e03-8c13f424a133] Received event network-vif-deleted-31850256-5668-4fd1-9758-bf86c83c33e4 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 834.834683] env[61957]: INFO nova.compute.manager [req-281fef2d-8c0f-4d22-a9ec-1f6fc54ea0f4 req-0737baf9-2e1b-432e-ab35-c708e26c4298 service nova] [instance: 615dbbed-2b02-4351-9e03-8c13f424a133] Neutron deleted interface 31850256-5668-4fd1-9758-bf86c83c33e4; detaching it from the instance and deleting it from the info cache [ 834.834860] env[61957]: DEBUG nova.network.neutron [req-281fef2d-8c0f-4d22-a9ec-1f6fc54ea0f4 req-0737baf9-2e1b-432e-ab35-c708e26c4298 service nova] [instance: 615dbbed-2b02-4351-9e03-8c13f424a133] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 834.872494] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-82a5645a-fb19-4ad1-aa47-e12b7f048b77 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] [instance: 17f00b20-9d3b-45e6-919d-6fab9999ec77] Powering off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 834.872494] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ace78250-28cc-4a7e-9b59-874ceb90b714 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.881306] env[61957]: DEBUG oslo_vmware.api [None req-82a5645a-fb19-4ad1-aa47-e12b7f048b77 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Waiting for the task: (returnval){ [ 834.881306] env[61957]: value = "task-1277460" [ 834.881306] env[61957]: _type = "Task" [ 834.881306] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.891493] env[61957]: DEBUG oslo_vmware.api [None req-82a5645a-fb19-4ad1-aa47-e12b7f048b77 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Task: {'id': task-1277460, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.894653] env[61957]: DEBUG nova.compute.manager [None req-227f29c0-ef6f-447c-b621-3f65bc420b4e tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] [instance: 37a9e1c1-9f6c-4047-9ff5-e141d0ca383e] Start building block device mappings for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 834.905838] env[61957]: DEBUG nova.network.neutron [None req-227f29c0-ef6f-447c-b621-3f65bc420b4e tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] [instance: 37a9e1c1-9f6c-4047-9ff5-e141d0ca383e] Successfully created port: aee17432-3f8d-4d87-863f-20a333a4c92f {{(pid=61957) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 834.922810] env[61957]: WARNING nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Instance 6e7bd89c-2c2a-450a-9858-3526d96c28ab is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 834.922974] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Instance e0249936-d616-4ffb-8f77-d8107633c42a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61957) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 834.923130] env[61957]: WARNING nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Instance 615dbbed-2b02-4351-9e03-8c13f424a133 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 834.923260] env[61957]: WARNING nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Instance e9ac118f-08b3-430b-848c-461c2b2e3e02 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 834.923377] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Instance 14ba830d-4e2c-4e9a-a059-3c86209f0127 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61957) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 834.923496] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Instance 17f00b20-9d3b-45e6-919d-6fab9999ec77 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61957) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 834.923610] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Instance 4951132e-7247-4772-8f88-3664c6a7e61e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61957) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 834.923718] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Instance ead6aae1-36b5-4f57-9129-3bb02cf103ce actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61957) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 834.923830] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Instance 37ce46a1-8f9a-4d15-bd81-e40845a0e48a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61957) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 834.923939] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Instance 37a9e1c1-9f6c-4047-9ff5-e141d0ca383e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61957) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 835.087924] env[61957]: DEBUG nova.network.neutron [req-d53dc9e8-04fb-472d-b49b-f034d99beb21 req-3734b50b-b5ce-4f59-ac1a-722f2491140f service nova] [instance: 37ce46a1-8f9a-4d15-bd81-e40845a0e48a] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 835.119769] env[61957]: DEBUG oslo_vmware.api [None req-7bbc9489-a498-401d-a2af-2f28d461499e tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': task-1277459, 'name': ReconfigVM_Task, 'duration_secs': 0.477643} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.120016] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-7bbc9489-a498-401d-a2af-2f28d461499e tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 4951132e-7247-4772-8f88-3664c6a7e61e] Reconfigured VM instance instance-00000038 to attach disk [datastore1] 4951132e-7247-4772-8f88-3664c6a7e61e/4951132e-7247-4772-8f88-3664c6a7e61e.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 835.120720] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-05570790-687b-448b-90a0-cf9da7aa8660 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.126643] env[61957]: DEBUG oslo_vmware.api [None req-7bbc9489-a498-401d-a2af-2f28d461499e tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Waiting for the task: (returnval){ [ 835.126643] env[61957]: value = "task-1277461" [ 835.126643] env[61957]: _type = "Task" [ 835.126643] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.134856] env[61957]: DEBUG oslo_vmware.api [None req-7bbc9489-a498-401d-a2af-2f28d461499e tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': task-1277461, 'name': Rename_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.150036] env[61957]: DEBUG oslo_vmware.api [None req-3c3df299-fd1f-460a-8cc2-979af1db30d1 tempest-ServersTestJSON-1093633677 tempest-ServersTestJSON-1093633677-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]524143b8-faea-9b7c-7dba-c2e59361b23a, 'name': SearchDatastore_Task, 'duration_secs': 0.011212} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.150948] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6b9e2f5c-06b7-4501-99d8-b744fd918a24 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.156890] env[61957]: DEBUG oslo_vmware.api [None req-3c3df299-fd1f-460a-8cc2-979af1db30d1 tempest-ServersTestJSON-1093633677 tempest-ServersTestJSON-1093633677-project-member] Waiting for the task: (returnval){ [ 835.156890] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]522e95b4-1990-aa03-a928-5879725a406d" [ 835.156890] env[61957]: _type = "Task" [ 835.156890] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.165931] env[61957]: DEBUG oslo_vmware.api [None req-3c3df299-fd1f-460a-8cc2-979af1db30d1 tempest-ServersTestJSON-1093633677 tempest-ServersTestJSON-1093633677-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]522e95b4-1990-aa03-a928-5879725a406d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.239882] env[61957]: DEBUG nova.network.neutron [req-d53dc9e8-04fb-472d-b49b-f034d99beb21 req-3734b50b-b5ce-4f59-ac1a-722f2491140f service nova] [instance: 37ce46a1-8f9a-4d15-bd81-e40845a0e48a] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 835.340167] env[61957]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a8ca9fe0-825e-4dc8-aa76-40cd254703f7 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.349528] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84a35de4-d6e2-4e23-8ae7-7f90331ee94d {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.373810] env[61957]: DEBUG nova.compute.manager [req-281fef2d-8c0f-4d22-a9ec-1f6fc54ea0f4 req-0737baf9-2e1b-432e-ab35-c708e26c4298 service nova] [instance: 615dbbed-2b02-4351-9e03-8c13f424a133] Detach interface failed, port_id=31850256-5668-4fd1-9758-bf86c83c33e4, reason: Instance 615dbbed-2b02-4351-9e03-8c13f424a133 could not be found. {{(pid=61957) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 835.389981] env[61957]: DEBUG oslo_vmware.api [None req-82a5645a-fb19-4ad1-aa47-e12b7f048b77 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Task: {'id': task-1277460, 'name': PowerOffVM_Task, 'duration_secs': 0.132365} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.390281] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-82a5645a-fb19-4ad1-aa47-e12b7f048b77 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] [instance: 17f00b20-9d3b-45e6-919d-6fab9999ec77] Powered off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 835.390502] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-82a5645a-fb19-4ad1-aa47-e12b7f048b77 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] [instance: 17f00b20-9d3b-45e6-919d-6fab9999ec77] Destroying instance {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 835.391363] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f024a43f-2962-41bc-9940-b2395271edba {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.397724] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-82a5645a-fb19-4ad1-aa47-e12b7f048b77 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] [instance: 17f00b20-9d3b-45e6-919d-6fab9999ec77] Unregistering the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 835.397992] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5abc7ad0-106a-437e-8795-438be9aede83 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.424839] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-82a5645a-fb19-4ad1-aa47-e12b7f048b77 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] [instance: 17f00b20-9d3b-45e6-919d-6fab9999ec77] Unregistered the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 835.424839] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-82a5645a-fb19-4ad1-aa47-e12b7f048b77 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] [instance: 17f00b20-9d3b-45e6-919d-6fab9999ec77] Deleting contents of the VM from datastore datastore2 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 835.424839] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-82a5645a-fb19-4ad1-aa47-e12b7f048b77 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Deleting the datastore file [datastore2] 17f00b20-9d3b-45e6-919d-6fab9999ec77 {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 835.424839] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-af5015a3-a095-46b7-8055-83373cb574ba {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.428992] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Instance 0adae8e1-8c2f-4110-805b-1f286debc833 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61957) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 835.432036] env[61957]: DEBUG oslo_vmware.api [None req-82a5645a-fb19-4ad1-aa47-e12b7f048b77 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Waiting for the task: (returnval){ [ 835.432036] env[61957]: value = "task-1277463" [ 835.432036] env[61957]: _type = "Task" [ 835.432036] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.440656] env[61957]: DEBUG oslo_vmware.api [None req-82a5645a-fb19-4ad1-aa47-e12b7f048b77 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Task: {'id': task-1277463, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.636227] env[61957]: DEBUG oslo_vmware.api [None req-7bbc9489-a498-401d-a2af-2f28d461499e tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': task-1277461, 'name': Rename_Task, 'duration_secs': 0.210312} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.636539] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-7bbc9489-a498-401d-a2af-2f28d461499e tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 4951132e-7247-4772-8f88-3664c6a7e61e] Powering on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 835.636798] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b072207c-bbe3-4977-acd8-943b32e6a9c8 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.643032] env[61957]: DEBUG oslo_vmware.api [None req-7bbc9489-a498-401d-a2af-2f28d461499e tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Waiting for the task: (returnval){ [ 835.643032] env[61957]: value = "task-1277464" [ 835.643032] env[61957]: _type = "Task" [ 835.643032] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.651017] env[61957]: DEBUG oslo_vmware.api [None req-7bbc9489-a498-401d-a2af-2f28d461499e tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': task-1277464, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.665421] env[61957]: DEBUG oslo_vmware.api [None req-3c3df299-fd1f-460a-8cc2-979af1db30d1 tempest-ServersTestJSON-1093633677 tempest-ServersTestJSON-1093633677-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]522e95b4-1990-aa03-a928-5879725a406d, 'name': SearchDatastore_Task, 'duration_secs': 0.012041} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.665687] env[61957]: DEBUG oslo_concurrency.lockutils [None req-3c3df299-fd1f-460a-8cc2-979af1db30d1 tempest-ServersTestJSON-1093633677 tempest-ServersTestJSON-1093633677-project-member] Releasing lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 835.665947] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c3df299-fd1f-460a-8cc2-979af1db30d1 tempest-ServersTestJSON-1093633677 tempest-ServersTestJSON-1093633677-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore1] ead6aae1-36b5-4f57-9129-3bb02cf103ce/ead6aae1-36b5-4f57-9129-3bb02cf103ce.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 835.666254] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2eb2feda-feea-44a6-8585-74278d7511d8 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.671820] env[61957]: DEBUG oslo_vmware.api [None req-3c3df299-fd1f-460a-8cc2-979af1db30d1 tempest-ServersTestJSON-1093633677 tempest-ServersTestJSON-1093633677-project-member] Waiting for the task: (returnval){ [ 835.671820] env[61957]: value = "task-1277465" [ 835.671820] env[61957]: _type = "Task" [ 835.671820] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.679565] env[61957]: DEBUG oslo_vmware.api [None req-3c3df299-fd1f-460a-8cc2-979af1db30d1 tempest-ServersTestJSON-1093633677 tempest-ServersTestJSON-1093633677-project-member] Task: {'id': task-1277465, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.742972] env[61957]: DEBUG oslo_concurrency.lockutils [req-d53dc9e8-04fb-472d-b49b-f034d99beb21 req-3734b50b-b5ce-4f59-ac1a-722f2491140f service nova] Releasing lock "refresh_cache-37ce46a1-8f9a-4d15-bd81-e40845a0e48a" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 835.745319] env[61957]: DEBUG oslo_concurrency.lockutils [None req-be584a65-2286-47ec-9ff6-29a7835a7d50 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Acquired lock "refresh_cache-37ce46a1-8f9a-4d15-bd81-e40845a0e48a" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 835.745495] env[61957]: DEBUG nova.network.neutron [None req-be584a65-2286-47ec-9ff6-29a7835a7d50 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: 37ce46a1-8f9a-4d15-bd81-e40845a0e48a] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 835.753832] env[61957]: DEBUG oslo_vmware.rw_handles [None req-83fd723b-e45f-4910-94da-c61980bb25f7 tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5271ff36-84ad-9f57-797f-de6dfd662268/disk-0.vmdk. {{(pid=61957) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 835.754829] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e44b6df6-4c61-4994-a13c-a4967ee02d34 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.761866] env[61957]: DEBUG oslo_vmware.rw_handles [None req-83fd723b-e45f-4910-94da-c61980bb25f7 tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5271ff36-84ad-9f57-797f-de6dfd662268/disk-0.vmdk is in state: ready. {{(pid=61957) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 835.762112] env[61957]: ERROR oslo_vmware.rw_handles [None req-83fd723b-e45f-4910-94da-c61980bb25f7 tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5271ff36-84ad-9f57-797f-de6dfd662268/disk-0.vmdk due to incomplete transfer. [ 835.762572] env[61957]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-e5c0b79c-3ae3-4049-be28-14482e2849f0 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.769234] env[61957]: DEBUG oslo_vmware.rw_handles [None req-83fd723b-e45f-4910-94da-c61980bb25f7 tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5271ff36-84ad-9f57-797f-de6dfd662268/disk-0.vmdk. {{(pid=61957) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 835.769539] env[61957]: DEBUG nova.virt.vmwareapi.images [None req-83fd723b-e45f-4910-94da-c61980bb25f7 tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] [instance: e0249936-d616-4ffb-8f77-d8107633c42a] Uploaded image c5bc6925-6597-4ca2-8c16-f44d676d30b2 to the Glance image server {{(pid=61957) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 835.772106] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-83fd723b-e45f-4910-94da-c61980bb25f7 tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] [instance: e0249936-d616-4ffb-8f77-d8107633c42a] Destroying the VM {{(pid=61957) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 835.772379] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-d70fb2ec-9d52-4011-9c92-d4e314bc7a4b {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.779222] env[61957]: DEBUG oslo_vmware.api [None req-83fd723b-e45f-4910-94da-c61980bb25f7 tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Waiting for the task: (returnval){ [ 835.779222] env[61957]: value = "task-1277466" [ 835.779222] env[61957]: _type = "Task" [ 835.779222] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.786981] env[61957]: DEBUG oslo_vmware.api [None req-83fd723b-e45f-4910-94da-c61980bb25f7 tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Task: {'id': task-1277466, 'name': Destroy_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.906108] env[61957]: DEBUG nova.compute.manager [None req-227f29c0-ef6f-447c-b621-3f65bc420b4e tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] [instance: 37a9e1c1-9f6c-4047-9ff5-e141d0ca383e] Start spawning the instance on the hypervisor. {{(pid=61957) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 835.927824] env[61957]: DEBUG nova.virt.hardware [None req-227f29c0-ef6f-447c-b621-3f65bc420b4e tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T17:22:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T17:21:50Z,direct_url=,disk_format='vmdk',id=11c76a2c-f705-470a-ba9d-4657858bab38,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='81d3096d0f094373913fc3dc8f5c3c6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T17:21:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 835.928276] env[61957]: DEBUG nova.virt.hardware [None req-227f29c0-ef6f-447c-b621-3f65bc420b4e tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] Flavor limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 835.928322] env[61957]: DEBUG nova.virt.hardware [None req-227f29c0-ef6f-447c-b621-3f65bc420b4e tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] Image limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 835.928517] env[61957]: DEBUG nova.virt.hardware [None req-227f29c0-ef6f-447c-b621-3f65bc420b4e tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] Flavor pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 835.928668] env[61957]: DEBUG nova.virt.hardware [None req-227f29c0-ef6f-447c-b621-3f65bc420b4e tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] Image pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 835.928823] env[61957]: DEBUG nova.virt.hardware [None req-227f29c0-ef6f-447c-b621-3f65bc420b4e tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 835.929052] env[61957]: DEBUG nova.virt.hardware [None req-227f29c0-ef6f-447c-b621-3f65bc420b4e tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 835.929224] env[61957]: DEBUG nova.virt.hardware [None req-227f29c0-ef6f-447c-b621-3f65bc420b4e tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 835.929396] env[61957]: DEBUG nova.virt.hardware [None req-227f29c0-ef6f-447c-b621-3f65bc420b4e tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] Got 1 possible topologies {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 835.929559] env[61957]: DEBUG nova.virt.hardware [None req-227f29c0-ef6f-447c-b621-3f65bc420b4e tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 835.929783] env[61957]: DEBUG nova.virt.hardware [None req-227f29c0-ef6f-447c-b621-3f65bc420b4e tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 835.930662] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73db8ccb-3c6b-4ba1-b04e-a88645d798d9 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.933719] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Instance 6438fe40-046c-45d5-9986-8f182ecde49f has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61957) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 835.946130] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebb6f802-2062-4ab2-a368-eb733cc79bcc {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.950043] env[61957]: DEBUG oslo_vmware.api [None req-82a5645a-fb19-4ad1-aa47-e12b7f048b77 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Task: {'id': task-1277463, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.154932} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.950927] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-82a5645a-fb19-4ad1-aa47-e12b7f048b77 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Deleted the datastore file {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 835.951146] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-82a5645a-fb19-4ad1-aa47-e12b7f048b77 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] [instance: 17f00b20-9d3b-45e6-919d-6fab9999ec77] Deleted contents of the VM from datastore datastore2 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 835.951330] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-82a5645a-fb19-4ad1-aa47-e12b7f048b77 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] [instance: 17f00b20-9d3b-45e6-919d-6fab9999ec77] Instance destroyed {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 836.155979] env[61957]: DEBUG oslo_vmware.api [None req-7bbc9489-a498-401d-a2af-2f28d461499e tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': task-1277464, 'name': PowerOnVM_Task, 'duration_secs': 0.490869} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.156417] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-7bbc9489-a498-401d-a2af-2f28d461499e tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 4951132e-7247-4772-8f88-3664c6a7e61e] Powered on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 836.156780] env[61957]: INFO nova.compute.manager [None req-7bbc9489-a498-401d-a2af-2f28d461499e tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 4951132e-7247-4772-8f88-3664c6a7e61e] Took 8.23 seconds to spawn the instance on the hypervisor. [ 836.157058] env[61957]: DEBUG nova.compute.manager [None req-7bbc9489-a498-401d-a2af-2f28d461499e tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 4951132e-7247-4772-8f88-3664c6a7e61e] Checking state {{(pid=61957) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 836.157965] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8acddd33-163e-4afb-81f1-fbc3d378eac8 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.181116] env[61957]: DEBUG oslo_vmware.api [None req-3c3df299-fd1f-460a-8cc2-979af1db30d1 tempest-ServersTestJSON-1093633677 tempest-ServersTestJSON-1093633677-project-member] Task: {'id': task-1277465, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.294529] env[61957]: DEBUG oslo_vmware.api [None req-83fd723b-e45f-4910-94da-c61980bb25f7 tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Task: {'id': task-1277466, 'name': Destroy_Task} progress is 33%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.295348] env[61957]: DEBUG nova.network.neutron [None req-be584a65-2286-47ec-9ff6-29a7835a7d50 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: 37ce46a1-8f9a-4d15-bd81-e40845a0e48a] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 836.439615] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Instance f66db265-887e-4d61-b848-c609e5c884cb has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61957) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 836.511433] env[61957]: DEBUG nova.network.neutron [None req-be584a65-2286-47ec-9ff6-29a7835a7d50 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: 37ce46a1-8f9a-4d15-bd81-e40845a0e48a] Updating instance_info_cache with network_info: [{"id": "7234f796-0abc-4e81-ba04-29607283618c", "address": "fa:16:3e:26:32:6a", "network": {"id": "6fbccc17-6674-4cd5-8f1b-aa73570baac7", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-394544811-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1dbdc699741a48af93c52068d88a5357", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "20e3f794-c7a3-4696-9488-ecf34c570ef9", "external-id": "nsx-vlan-transportzone-509", "segmentation_id": 509, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7234f796-0a", "ovs_interfaceid": "7234f796-0abc-4e81-ba04-29607283618c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 836.612099] env[61957]: DEBUG nova.compute.manager [req-5ebc5b52-a51d-45a7-9ba7-cba57ef748dd req-66e630aa-c86d-49ca-80fe-2163b26f0d52 service nova] [instance: 37a9e1c1-9f6c-4047-9ff5-e141d0ca383e] Received event network-vif-plugged-aee17432-3f8d-4d87-863f-20a333a4c92f {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 836.612335] env[61957]: DEBUG oslo_concurrency.lockutils [req-5ebc5b52-a51d-45a7-9ba7-cba57ef748dd req-66e630aa-c86d-49ca-80fe-2163b26f0d52 service nova] Acquiring lock "37a9e1c1-9f6c-4047-9ff5-e141d0ca383e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 836.612546] env[61957]: DEBUG oslo_concurrency.lockutils [req-5ebc5b52-a51d-45a7-9ba7-cba57ef748dd req-66e630aa-c86d-49ca-80fe-2163b26f0d52 service nova] Lock "37a9e1c1-9f6c-4047-9ff5-e141d0ca383e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 836.612720] env[61957]: DEBUG oslo_concurrency.lockutils [req-5ebc5b52-a51d-45a7-9ba7-cba57ef748dd req-66e630aa-c86d-49ca-80fe-2163b26f0d52 service nova] Lock "37a9e1c1-9f6c-4047-9ff5-e141d0ca383e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 836.612887] env[61957]: DEBUG nova.compute.manager [req-5ebc5b52-a51d-45a7-9ba7-cba57ef748dd req-66e630aa-c86d-49ca-80fe-2163b26f0d52 service nova] [instance: 37a9e1c1-9f6c-4047-9ff5-e141d0ca383e] No waiting events found dispatching network-vif-plugged-aee17432-3f8d-4d87-863f-20a333a4c92f {{(pid=61957) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 836.613099] env[61957]: WARNING nova.compute.manager [req-5ebc5b52-a51d-45a7-9ba7-cba57ef748dd req-66e630aa-c86d-49ca-80fe-2163b26f0d52 service nova] [instance: 37a9e1c1-9f6c-4047-9ff5-e141d0ca383e] Received unexpected event network-vif-plugged-aee17432-3f8d-4d87-863f-20a333a4c92f for instance with vm_state building and task_state spawning. [ 836.652121] env[61957]: DEBUG nova.network.neutron [None req-227f29c0-ef6f-447c-b621-3f65bc420b4e tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] [instance: 37a9e1c1-9f6c-4047-9ff5-e141d0ca383e] Successfully updated port: aee17432-3f8d-4d87-863f-20a333a4c92f {{(pid=61957) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 836.675620] env[61957]: INFO nova.compute.manager [None req-7bbc9489-a498-401d-a2af-2f28d461499e tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 4951132e-7247-4772-8f88-3664c6a7e61e] Took 26.62 seconds to build instance. [ 836.686861] env[61957]: DEBUG oslo_vmware.api [None req-3c3df299-fd1f-460a-8cc2-979af1db30d1 tempest-ServersTestJSON-1093633677 tempest-ServersTestJSON-1093633677-project-member] Task: {'id': task-1277465, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.606804} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.687220] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c3df299-fd1f-460a-8cc2-979af1db30d1 tempest-ServersTestJSON-1093633677 tempest-ServersTestJSON-1093633677-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore1] ead6aae1-36b5-4f57-9129-3bb02cf103ce/ead6aae1-36b5-4f57-9129-3bb02cf103ce.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 836.687449] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-3c3df299-fd1f-460a-8cc2-979af1db30d1 tempest-ServersTestJSON-1093633677 tempest-ServersTestJSON-1093633677-project-member] [instance: ead6aae1-36b5-4f57-9129-3bb02cf103ce] Extending root virtual disk to 1048576 {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 836.687692] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-dbd61a7a-8560-460e-9779-d8ad08b0b922 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.694173] env[61957]: DEBUG oslo_vmware.api [None req-3c3df299-fd1f-460a-8cc2-979af1db30d1 tempest-ServersTestJSON-1093633677 tempest-ServersTestJSON-1093633677-project-member] Waiting for the task: (returnval){ [ 836.694173] env[61957]: value = "task-1277467" [ 836.694173] env[61957]: _type = "Task" [ 836.694173] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.702048] env[61957]: DEBUG oslo_vmware.api [None req-3c3df299-fd1f-460a-8cc2-979af1db30d1 tempest-ServersTestJSON-1093633677 tempest-ServersTestJSON-1093633677-project-member] Task: {'id': task-1277467, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.791711] env[61957]: DEBUG oslo_vmware.api [None req-83fd723b-e45f-4910-94da-c61980bb25f7 tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Task: {'id': task-1277466, 'name': Destroy_Task, 'duration_secs': 0.752034} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.792011] env[61957]: INFO nova.virt.vmwareapi.vm_util [None req-83fd723b-e45f-4910-94da-c61980bb25f7 tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] [instance: e0249936-d616-4ffb-8f77-d8107633c42a] Destroyed the VM [ 836.792351] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-83fd723b-e45f-4910-94da-c61980bb25f7 tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] [instance: e0249936-d616-4ffb-8f77-d8107633c42a] Deleting Snapshot of the VM instance {{(pid=61957) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 836.792620] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-caa95e9b-b409-48bc-abb0-a5df3fc285cb {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.798929] env[61957]: DEBUG oslo_vmware.api [None req-83fd723b-e45f-4910-94da-c61980bb25f7 tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Waiting for the task: (returnval){ [ 836.798929] env[61957]: value = "task-1277468" [ 836.798929] env[61957]: _type = "Task" [ 836.798929] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.806357] env[61957]: DEBUG oslo_vmware.api [None req-83fd723b-e45f-4910-94da-c61980bb25f7 tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Task: {'id': task-1277468, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.943108] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Instance c23141ee-0cbb-4d1b-8390-c3073fe354f1 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61957) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 836.993240] env[61957]: DEBUG nova.virt.hardware [None req-82a5645a-fb19-4ad1-aa47-e12b7f048b77 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T17:22:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T17:21:50Z,direct_url=,disk_format='vmdk',id=11c76a2c-f705-470a-ba9d-4657858bab38,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='81d3096d0f094373913fc3dc8f5c3c6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T17:21:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 836.993528] env[61957]: DEBUG nova.virt.hardware [None req-82a5645a-fb19-4ad1-aa47-e12b7f048b77 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Flavor limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 836.993763] env[61957]: DEBUG nova.virt.hardware [None req-82a5645a-fb19-4ad1-aa47-e12b7f048b77 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Image limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 836.993977] env[61957]: DEBUG nova.virt.hardware [None req-82a5645a-fb19-4ad1-aa47-e12b7f048b77 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Flavor pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 836.994904] env[61957]: DEBUG nova.virt.hardware [None req-82a5645a-fb19-4ad1-aa47-e12b7f048b77 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Image pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 836.995117] env[61957]: DEBUG nova.virt.hardware [None req-82a5645a-fb19-4ad1-aa47-e12b7f048b77 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 836.995348] env[61957]: DEBUG nova.virt.hardware [None req-82a5645a-fb19-4ad1-aa47-e12b7f048b77 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 836.995516] env[61957]: DEBUG nova.virt.hardware [None req-82a5645a-fb19-4ad1-aa47-e12b7f048b77 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 836.995693] env[61957]: DEBUG nova.virt.hardware [None req-82a5645a-fb19-4ad1-aa47-e12b7f048b77 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Got 1 possible topologies {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 836.995860] env[61957]: DEBUG nova.virt.hardware [None req-82a5645a-fb19-4ad1-aa47-e12b7f048b77 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 836.996052] env[61957]: DEBUG nova.virt.hardware [None req-82a5645a-fb19-4ad1-aa47-e12b7f048b77 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 836.997218] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c9d6017-ed4b-414f-9908-37cb109d686d {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.005044] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a3e79b0-00b2-4d39-a6db-7d30b4744c83 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.018461] env[61957]: DEBUG oslo_concurrency.lockutils [None req-be584a65-2286-47ec-9ff6-29a7835a7d50 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Releasing lock "refresh_cache-37ce46a1-8f9a-4d15-bd81-e40845a0e48a" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 837.018461] env[61957]: DEBUG nova.compute.manager [None req-be584a65-2286-47ec-9ff6-29a7835a7d50 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: 37ce46a1-8f9a-4d15-bd81-e40845a0e48a] Instance network_info: |[{"id": "7234f796-0abc-4e81-ba04-29607283618c", "address": "fa:16:3e:26:32:6a", "network": {"id": "6fbccc17-6674-4cd5-8f1b-aa73570baac7", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-394544811-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1dbdc699741a48af93c52068d88a5357", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "20e3f794-c7a3-4696-9488-ecf34c570ef9", "external-id": "nsx-vlan-transportzone-509", "segmentation_id": 509, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7234f796-0a", "ovs_interfaceid": "7234f796-0abc-4e81-ba04-29607283618c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 837.018730] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-82a5645a-fb19-4ad1-aa47-e12b7f048b77 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] [instance: 17f00b20-9d3b-45e6-919d-6fab9999ec77] Instance VIF info [] {{(pid=61957) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 837.024325] env[61957]: DEBUG oslo.service.loopingcall [None req-82a5645a-fb19-4ad1-aa47-e12b7f048b77 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 837.024636] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-be584a65-2286-47ec-9ff6-29a7835a7d50 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: 37ce46a1-8f9a-4d15-bd81-e40845a0e48a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:26:32:6a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '20e3f794-c7a3-4696-9488-ecf34c570ef9', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7234f796-0abc-4e81-ba04-29607283618c', 'vif_model': 'vmxnet3'}] {{(pid=61957) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 837.032142] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-be584a65-2286-47ec-9ff6-29a7835a7d50 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Creating folder: Project (1dbdc699741a48af93c52068d88a5357). Parent ref: group-v274445. {{(pid=61957) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 837.032387] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 17f00b20-9d3b-45e6-919d-6fab9999ec77] Creating VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 837.032904] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ebb4f7bc-140a-481f-b0cc-69c76fba194b {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.034510] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-80d73c7d-05a3-4942-bdba-65d2299a1560 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.052124] env[61957]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 837.052124] env[61957]: value = "task-1277470" [ 837.052124] env[61957]: _type = "Task" [ 837.052124] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.066017] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277470, 'name': CreateVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.067258] env[61957]: INFO nova.virt.vmwareapi.vm_util [None req-be584a65-2286-47ec-9ff6-29a7835a7d50 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Created folder: Project (1dbdc699741a48af93c52068d88a5357) in parent group-v274445. [ 837.067448] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-be584a65-2286-47ec-9ff6-29a7835a7d50 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Creating folder: Instances. Parent ref: group-v274495. {{(pid=61957) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 837.067689] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0ad03656-be13-4533-bf30-f9f4d6914b0c {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.077216] env[61957]: INFO nova.virt.vmwareapi.vm_util [None req-be584a65-2286-47ec-9ff6-29a7835a7d50 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Created folder: Instances in parent group-v274495. [ 837.077467] env[61957]: DEBUG oslo.service.loopingcall [None req-be584a65-2286-47ec-9ff6-29a7835a7d50 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 837.077681] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 37ce46a1-8f9a-4d15-bd81-e40845a0e48a] Creating VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 837.077975] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-69cdc78d-2795-460a-9f28-fbe24e8a3c9e {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.096413] env[61957]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 837.096413] env[61957]: value = "task-1277472" [ 837.096413] env[61957]: _type = "Task" [ 837.096413] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.104951] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277472, 'name': CreateVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.155541] env[61957]: DEBUG oslo_concurrency.lockutils [None req-227f29c0-ef6f-447c-b621-3f65bc420b4e tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] Acquiring lock "refresh_cache-37a9e1c1-9f6c-4047-9ff5-e141d0ca383e" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 837.155801] env[61957]: DEBUG oslo_concurrency.lockutils [None req-227f29c0-ef6f-447c-b621-3f65bc420b4e tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] Acquired lock "refresh_cache-37a9e1c1-9f6c-4047-9ff5-e141d0ca383e" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 837.156055] env[61957]: DEBUG nova.network.neutron [None req-227f29c0-ef6f-447c-b621-3f65bc420b4e tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] [instance: 37a9e1c1-9f6c-4047-9ff5-e141d0ca383e] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 837.177635] env[61957]: DEBUG oslo_concurrency.lockutils [None req-7bbc9489-a498-401d-a2af-2f28d461499e tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Lock "4951132e-7247-4772-8f88-3664c6a7e61e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 102.151s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 837.203927] env[61957]: DEBUG oslo_vmware.api [None req-3c3df299-fd1f-460a-8cc2-979af1db30d1 tempest-ServersTestJSON-1093633677 tempest-ServersTestJSON-1093633677-project-member] Task: {'id': task-1277467, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.079998} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.204229] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-3c3df299-fd1f-460a-8cc2-979af1db30d1 tempest-ServersTestJSON-1093633677 tempest-ServersTestJSON-1093633677-project-member] [instance: ead6aae1-36b5-4f57-9129-3bb02cf103ce] Extended root virtual disk {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 837.205059] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c7b8190-b716-4509-8730-a747a0084090 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.227816] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-3c3df299-fd1f-460a-8cc2-979af1db30d1 tempest-ServersTestJSON-1093633677 tempest-ServersTestJSON-1093633677-project-member] [instance: ead6aae1-36b5-4f57-9129-3bb02cf103ce] Reconfiguring VM instance instance-00000039 to attach disk [datastore1] ead6aae1-36b5-4f57-9129-3bb02cf103ce/ead6aae1-36b5-4f57-9129-3bb02cf103ce.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 837.228054] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e797bc47-df16-48ee-8c57-0869e795a70d {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.247813] env[61957]: DEBUG oslo_vmware.api [None req-3c3df299-fd1f-460a-8cc2-979af1db30d1 tempest-ServersTestJSON-1093633677 tempest-ServersTestJSON-1093633677-project-member] Waiting for the task: (returnval){ [ 837.247813] env[61957]: value = "task-1277473" [ 837.247813] env[61957]: _type = "Task" [ 837.247813] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.256778] env[61957]: DEBUG oslo_vmware.api [None req-3c3df299-fd1f-460a-8cc2-979af1db30d1 tempest-ServersTestJSON-1093633677 tempest-ServersTestJSON-1093633677-project-member] Task: {'id': task-1277473, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.311285] env[61957]: DEBUG oslo_vmware.api [None req-83fd723b-e45f-4910-94da-c61980bb25f7 tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Task: {'id': task-1277468, 'name': RemoveSnapshot_Task} progress is 80%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.449200] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Instance 281b9644-0e7d-48b9-a7b6-45fd6102d558 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61957) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 837.563156] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277470, 'name': CreateVM_Task, 'duration_secs': 0.408241} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.563572] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 17f00b20-9d3b-45e6-919d-6fab9999ec77] Created VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 837.564138] env[61957]: DEBUG oslo_concurrency.lockutils [None req-82a5645a-fb19-4ad1-aa47-e12b7f048b77 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 837.564419] env[61957]: DEBUG oslo_concurrency.lockutils [None req-82a5645a-fb19-4ad1-aa47-e12b7f048b77 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 837.564875] env[61957]: DEBUG oslo_concurrency.lockutils [None req-82a5645a-fb19-4ad1-aa47-e12b7f048b77 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 837.565234] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2161caf0-c554-4e1d-9189-a8e11feb080a {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.570158] env[61957]: DEBUG oslo_vmware.api [None req-82a5645a-fb19-4ad1-aa47-e12b7f048b77 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Waiting for the task: (returnval){ [ 837.570158] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]529d35a5-c03b-19dc-0c5f-f0a25d8b87ce" [ 837.570158] env[61957]: _type = "Task" [ 837.570158] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.579394] env[61957]: DEBUG oslo_vmware.api [None req-82a5645a-fb19-4ad1-aa47-e12b7f048b77 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]529d35a5-c03b-19dc-0c5f-f0a25d8b87ce, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.612645] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277472, 'name': CreateVM_Task, 'duration_secs': 0.435998} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.612835] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 37ce46a1-8f9a-4d15-bd81-e40845a0e48a] Created VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 837.613583] env[61957]: DEBUG oslo_concurrency.lockutils [None req-be584a65-2286-47ec-9ff6-29a7835a7d50 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 837.613755] env[61957]: DEBUG oslo_concurrency.lockutils [None req-be584a65-2286-47ec-9ff6-29a7835a7d50 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Acquired lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 837.614091] env[61957]: DEBUG oslo_concurrency.lockutils [None req-be584a65-2286-47ec-9ff6-29a7835a7d50 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 837.614436] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dd5bc60d-833a-47a4-bdfd-96f301f0af96 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.619866] env[61957]: DEBUG oslo_vmware.api [None req-be584a65-2286-47ec-9ff6-29a7835a7d50 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Waiting for the task: (returnval){ [ 837.619866] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]528e3a65-04c2-1b33-4e4e-2aba71b483ed" [ 837.619866] env[61957]: _type = "Task" [ 837.619866] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.629032] env[61957]: DEBUG oslo_vmware.api [None req-be584a65-2286-47ec-9ff6-29a7835a7d50 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]528e3a65-04c2-1b33-4e4e-2aba71b483ed, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.680970] env[61957]: DEBUG nova.compute.manager [None req-c13ea078-3782-4688-93ee-408e1ac828c0 tempest-ServerTagsTestJSON-969152625 tempest-ServerTagsTestJSON-969152625-project-member] [instance: 8e04a098-25de-4d57-9e3c-ea44d234d57e] Starting instance... {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 837.695580] env[61957]: DEBUG nova.network.neutron [None req-227f29c0-ef6f-447c-b621-3f65bc420b4e tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] [instance: 37a9e1c1-9f6c-4047-9ff5-e141d0ca383e] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 837.759190] env[61957]: DEBUG oslo_vmware.api [None req-3c3df299-fd1f-460a-8cc2-979af1db30d1 tempest-ServersTestJSON-1093633677 tempest-ServersTestJSON-1093633677-project-member] Task: {'id': task-1277473, 'name': ReconfigVM_Task, 'duration_secs': 0.405275} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.759460] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-3c3df299-fd1f-460a-8cc2-979af1db30d1 tempest-ServersTestJSON-1093633677 tempest-ServersTestJSON-1093633677-project-member] [instance: ead6aae1-36b5-4f57-9129-3bb02cf103ce] Reconfigured VM instance instance-00000039 to attach disk [datastore1] ead6aae1-36b5-4f57-9129-3bb02cf103ce/ead6aae1-36b5-4f57-9129-3bb02cf103ce.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 837.760186] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f5a2e06f-478b-4308-8bf1-2f2947abff7a {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.767259] env[61957]: DEBUG oslo_vmware.api [None req-3c3df299-fd1f-460a-8cc2-979af1db30d1 tempest-ServersTestJSON-1093633677 tempest-ServersTestJSON-1093633677-project-member] Waiting for the task: (returnval){ [ 837.767259] env[61957]: value = "task-1277474" [ 837.767259] env[61957]: _type = "Task" [ 837.767259] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.779880] env[61957]: DEBUG oslo_vmware.api [None req-3c3df299-fd1f-460a-8cc2-979af1db30d1 tempest-ServersTestJSON-1093633677 tempest-ServersTestJSON-1093633677-project-member] Task: {'id': task-1277474, 'name': Rename_Task} progress is 5%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.810300] env[61957]: DEBUG nova.compute.manager [req-c4581270-22f1-41a6-8f2b-7cbd0443931b req-f9f8d2d2-df39-48e3-a59e-12db11766021 service nova] [instance: 4951132e-7247-4772-8f88-3664c6a7e61e] Received event network-changed-3016c8eb-2ad3-40b1-8d88-eaec447de9ea {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 837.810400] env[61957]: DEBUG nova.compute.manager [req-c4581270-22f1-41a6-8f2b-7cbd0443931b req-f9f8d2d2-df39-48e3-a59e-12db11766021 service nova] [instance: 4951132e-7247-4772-8f88-3664c6a7e61e] Refreshing instance network info cache due to event network-changed-3016c8eb-2ad3-40b1-8d88-eaec447de9ea. {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 837.811031] env[61957]: DEBUG oslo_concurrency.lockutils [req-c4581270-22f1-41a6-8f2b-7cbd0443931b req-f9f8d2d2-df39-48e3-a59e-12db11766021 service nova] Acquiring lock "refresh_cache-4951132e-7247-4772-8f88-3664c6a7e61e" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 837.811199] env[61957]: DEBUG oslo_concurrency.lockutils [req-c4581270-22f1-41a6-8f2b-7cbd0443931b req-f9f8d2d2-df39-48e3-a59e-12db11766021 service nova] Acquired lock "refresh_cache-4951132e-7247-4772-8f88-3664c6a7e61e" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 837.811658] env[61957]: DEBUG nova.network.neutron [req-c4581270-22f1-41a6-8f2b-7cbd0443931b req-f9f8d2d2-df39-48e3-a59e-12db11766021 service nova] [instance: 4951132e-7247-4772-8f88-3664c6a7e61e] Refreshing network info cache for port 3016c8eb-2ad3-40b1-8d88-eaec447de9ea {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 837.816119] env[61957]: DEBUG oslo_vmware.api [None req-83fd723b-e45f-4910-94da-c61980bb25f7 tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Task: {'id': task-1277468, 'name': RemoveSnapshot_Task, 'duration_secs': 0.787198} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.816580] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-83fd723b-e45f-4910-94da-c61980bb25f7 tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] [instance: e0249936-d616-4ffb-8f77-d8107633c42a] Deleted Snapshot of the VM instance {{(pid=61957) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 837.817012] env[61957]: INFO nova.compute.manager [None req-83fd723b-e45f-4910-94da-c61980bb25f7 tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] [instance: e0249936-d616-4ffb-8f77-d8107633c42a] Took 14.93 seconds to snapshot the instance on the hypervisor. [ 837.876443] env[61957]: DEBUG nova.network.neutron [None req-227f29c0-ef6f-447c-b621-3f65bc420b4e tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] [instance: 37a9e1c1-9f6c-4047-9ff5-e141d0ca383e] Updating instance_info_cache with network_info: [{"id": "aee17432-3f8d-4d87-863f-20a333a4c92f", "address": "fa:16:3e:21:42:d0", "network": {"id": "8ee8bd88-2141-46db-b100-99dc2aaa8574", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-975448626-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "21351f437ea54c9580753dfddd7f47be", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b9aabc7c-0f6c-42eb-bd27-493a1496c0c8", "external-id": "nsx-vlan-transportzone-368", "segmentation_id": 368, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaee17432-3f", "ovs_interfaceid": "aee17432-3f8d-4d87-863f-20a333a4c92f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 837.953773] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Instance 526ff179-62a6-4763-ab25-797617c4ed57 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61957) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 838.081120] env[61957]: DEBUG oslo_vmware.api [None req-82a5645a-fb19-4ad1-aa47-e12b7f048b77 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]529d35a5-c03b-19dc-0c5f-f0a25d8b87ce, 'name': SearchDatastore_Task, 'duration_secs': 0.014} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.081448] env[61957]: DEBUG oslo_concurrency.lockutils [None req-82a5645a-fb19-4ad1-aa47-e12b7f048b77 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 838.081682] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-82a5645a-fb19-4ad1-aa47-e12b7f048b77 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] [instance: 17f00b20-9d3b-45e6-919d-6fab9999ec77] Processing image 11c76a2c-f705-470a-ba9d-4657858bab38 {{(pid=61957) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 838.081965] env[61957]: DEBUG oslo_concurrency.lockutils [None req-82a5645a-fb19-4ad1-aa47-e12b7f048b77 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 838.082158] env[61957]: DEBUG oslo_concurrency.lockutils [None req-82a5645a-fb19-4ad1-aa47-e12b7f048b77 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 838.082345] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-82a5645a-fb19-4ad1-aa47-e12b7f048b77 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 838.082609] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9be3efbf-ae29-4a55-9085-e21680e73220 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.090332] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-82a5645a-fb19-4ad1-aa47-e12b7f048b77 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 838.090508] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-82a5645a-fb19-4ad1-aa47-e12b7f048b77 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61957) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 838.091224] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-42abe1e4-8c1e-4634-a0f9-85d39516ecb9 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.096137] env[61957]: DEBUG oslo_vmware.api [None req-82a5645a-fb19-4ad1-aa47-e12b7f048b77 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Waiting for the task: (returnval){ [ 838.096137] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]523d7257-c966-6d50-0c5f-f047df2aa63d" [ 838.096137] env[61957]: _type = "Task" [ 838.096137] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.103606] env[61957]: DEBUG oslo_vmware.api [None req-82a5645a-fb19-4ad1-aa47-e12b7f048b77 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]523d7257-c966-6d50-0c5f-f047df2aa63d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.128972] env[61957]: DEBUG oslo_vmware.api [None req-be584a65-2286-47ec-9ff6-29a7835a7d50 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]528e3a65-04c2-1b33-4e4e-2aba71b483ed, 'name': SearchDatastore_Task, 'duration_secs': 0.037966} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.129290] env[61957]: DEBUG oslo_concurrency.lockutils [None req-be584a65-2286-47ec-9ff6-29a7835a7d50 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Releasing lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 838.129519] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-be584a65-2286-47ec-9ff6-29a7835a7d50 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: 37ce46a1-8f9a-4d15-bd81-e40845a0e48a] Processing image 11c76a2c-f705-470a-ba9d-4657858bab38 {{(pid=61957) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 838.129794] env[61957]: DEBUG oslo_concurrency.lockutils [None req-be584a65-2286-47ec-9ff6-29a7835a7d50 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 838.129951] env[61957]: DEBUG oslo_concurrency.lockutils [None req-be584a65-2286-47ec-9ff6-29a7835a7d50 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Acquired lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 838.130148] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-be584a65-2286-47ec-9ff6-29a7835a7d50 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 838.130410] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8794721d-06aa-4983-9f07-1db8840db00c {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.138390] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-be584a65-2286-47ec-9ff6-29a7835a7d50 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 838.138560] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-be584a65-2286-47ec-9ff6-29a7835a7d50 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61957) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 838.139313] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-46f98864-c40e-49b8-b648-30d790e43366 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.144096] env[61957]: DEBUG oslo_vmware.api [None req-be584a65-2286-47ec-9ff6-29a7835a7d50 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Waiting for the task: (returnval){ [ 838.144096] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52cdc3d3-7719-77f5-36c4-be31b971be25" [ 838.144096] env[61957]: _type = "Task" [ 838.144096] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.152316] env[61957]: DEBUG oslo_vmware.api [None req-be584a65-2286-47ec-9ff6-29a7835a7d50 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52cdc3d3-7719-77f5-36c4-be31b971be25, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.208802] env[61957]: DEBUG oslo_concurrency.lockutils [None req-c13ea078-3782-4688-93ee-408e1ac828c0 tempest-ServerTagsTestJSON-969152625 tempest-ServerTagsTestJSON-969152625-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 838.276719] env[61957]: DEBUG oslo_vmware.api [None req-3c3df299-fd1f-460a-8cc2-979af1db30d1 tempest-ServersTestJSON-1093633677 tempest-ServersTestJSON-1093633677-project-member] Task: {'id': task-1277474, 'name': Rename_Task} progress is 99%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.381076] env[61957]: DEBUG oslo_concurrency.lockutils [None req-227f29c0-ef6f-447c-b621-3f65bc420b4e tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] Releasing lock "refresh_cache-37a9e1c1-9f6c-4047-9ff5-e141d0ca383e" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 838.381421] env[61957]: DEBUG nova.compute.manager [None req-227f29c0-ef6f-447c-b621-3f65bc420b4e tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] [instance: 37a9e1c1-9f6c-4047-9ff5-e141d0ca383e] Instance network_info: |[{"id": "aee17432-3f8d-4d87-863f-20a333a4c92f", "address": "fa:16:3e:21:42:d0", "network": {"id": "8ee8bd88-2141-46db-b100-99dc2aaa8574", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-975448626-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "21351f437ea54c9580753dfddd7f47be", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b9aabc7c-0f6c-42eb-bd27-493a1496c0c8", "external-id": "nsx-vlan-transportzone-368", "segmentation_id": 368, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaee17432-3f", "ovs_interfaceid": "aee17432-3f8d-4d87-863f-20a333a4c92f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 838.381862] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-227f29c0-ef6f-447c-b621-3f65bc420b4e tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] [instance: 37a9e1c1-9f6c-4047-9ff5-e141d0ca383e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:21:42:d0', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'b9aabc7c-0f6c-42eb-bd27-493a1496c0c8', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'aee17432-3f8d-4d87-863f-20a333a4c92f', 'vif_model': 'vmxnet3'}] {{(pid=61957) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 838.389365] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-227f29c0-ef6f-447c-b621-3f65bc420b4e tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] Creating folder: Project (21351f437ea54c9580753dfddd7f47be). Parent ref: group-v274445. {{(pid=61957) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 838.391956] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-fa28121a-a008-4ead-b2b3-2eac6423d5eb {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.406642] env[61957]: INFO nova.virt.vmwareapi.vm_util [None req-227f29c0-ef6f-447c-b621-3f65bc420b4e tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] Created folder: Project (21351f437ea54c9580753dfddd7f47be) in parent group-v274445. [ 838.406861] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-227f29c0-ef6f-447c-b621-3f65bc420b4e tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] Creating folder: Instances. Parent ref: group-v274499. {{(pid=61957) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 838.407131] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a5cfd054-1fb5-43fb-9c9e-1a594529005a {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.417731] env[61957]: INFO nova.virt.vmwareapi.vm_util [None req-227f29c0-ef6f-447c-b621-3f65bc420b4e tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] Created folder: Instances in parent group-v274499. [ 838.417881] env[61957]: DEBUG oslo.service.loopingcall [None req-227f29c0-ef6f-447c-b621-3f65bc420b4e tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 838.418111] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 37a9e1c1-9f6c-4047-9ff5-e141d0ca383e] Creating VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 838.418323] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-34e9cb86-f5ee-48df-a3a3-6c67055bb3b2 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.438651] env[61957]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 838.438651] env[61957]: value = "task-1277477" [ 838.438651] env[61957]: _type = "Task" [ 838.438651] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.446523] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277477, 'name': CreateVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.457552] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Instance 8e04a098-25de-4d57-9e3c-ea44d234d57e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61957) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 838.559088] env[61957]: DEBUG nova.network.neutron [req-c4581270-22f1-41a6-8f2b-7cbd0443931b req-f9f8d2d2-df39-48e3-a59e-12db11766021 service nova] [instance: 4951132e-7247-4772-8f88-3664c6a7e61e] Updated VIF entry in instance network info cache for port 3016c8eb-2ad3-40b1-8d88-eaec447de9ea. {{(pid=61957) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 838.559485] env[61957]: DEBUG nova.network.neutron [req-c4581270-22f1-41a6-8f2b-7cbd0443931b req-f9f8d2d2-df39-48e3-a59e-12db11766021 service nova] [instance: 4951132e-7247-4772-8f88-3664c6a7e61e] Updating instance_info_cache with network_info: [{"id": "3016c8eb-2ad3-40b1-8d88-eaec447de9ea", "address": "fa:16:3e:59:19:f2", "network": {"id": "364cabcb-046d-4f91-b1ee-5dd3adfae6d7", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1432396967-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.193", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8713aa35bcb24b86ad0b58ca9fc991ba", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b7a73c01-1bb9-4612-a1a7-16d71b732e81", "external-id": "nsx-vlan-transportzone-711", "segmentation_id": 711, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3016c8eb-2a", "ovs_interfaceid": "3016c8eb-2ad3-40b1-8d88-eaec447de9ea", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 838.609932] env[61957]: DEBUG oslo_vmware.api [None req-82a5645a-fb19-4ad1-aa47-e12b7f048b77 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]523d7257-c966-6d50-0c5f-f047df2aa63d, 'name': SearchDatastore_Task, 'duration_secs': 0.007692} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.610861] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c5c9bf8c-b0c6-4364-852f-236f20c7bcdf {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.616330] env[61957]: DEBUG oslo_vmware.api [None req-82a5645a-fb19-4ad1-aa47-e12b7f048b77 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Waiting for the task: (returnval){ [ 838.616330] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]527847e5-4067-7cb7-37df-bc60a23af8a7" [ 838.616330] env[61957]: _type = "Task" [ 838.616330] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.624702] env[61957]: DEBUG oslo_vmware.api [None req-82a5645a-fb19-4ad1-aa47-e12b7f048b77 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]527847e5-4067-7cb7-37df-bc60a23af8a7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.638099] env[61957]: DEBUG nova.compute.manager [req-70e4c08b-85e3-4b41-ad92-bd626f0c83ff req-9b7a8cbb-18e2-4d16-b3f6-3d188d9e2db6 service nova] [instance: 37a9e1c1-9f6c-4047-9ff5-e141d0ca383e] Received event network-changed-aee17432-3f8d-4d87-863f-20a333a4c92f {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 838.638379] env[61957]: DEBUG nova.compute.manager [req-70e4c08b-85e3-4b41-ad92-bd626f0c83ff req-9b7a8cbb-18e2-4d16-b3f6-3d188d9e2db6 service nova] [instance: 37a9e1c1-9f6c-4047-9ff5-e141d0ca383e] Refreshing instance network info cache due to event network-changed-aee17432-3f8d-4d87-863f-20a333a4c92f. {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 838.638656] env[61957]: DEBUG oslo_concurrency.lockutils [req-70e4c08b-85e3-4b41-ad92-bd626f0c83ff req-9b7a8cbb-18e2-4d16-b3f6-3d188d9e2db6 service nova] Acquiring lock "refresh_cache-37a9e1c1-9f6c-4047-9ff5-e141d0ca383e" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 838.638831] env[61957]: DEBUG oslo_concurrency.lockutils [req-70e4c08b-85e3-4b41-ad92-bd626f0c83ff req-9b7a8cbb-18e2-4d16-b3f6-3d188d9e2db6 service nova] Acquired lock "refresh_cache-37a9e1c1-9f6c-4047-9ff5-e141d0ca383e" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 838.639117] env[61957]: DEBUG nova.network.neutron [req-70e4c08b-85e3-4b41-ad92-bd626f0c83ff req-9b7a8cbb-18e2-4d16-b3f6-3d188d9e2db6 service nova] [instance: 37a9e1c1-9f6c-4047-9ff5-e141d0ca383e] Refreshing network info cache for port aee17432-3f8d-4d87-863f-20a333a4c92f {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 838.655999] env[61957]: DEBUG oslo_vmware.api [None req-be584a65-2286-47ec-9ff6-29a7835a7d50 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52cdc3d3-7719-77f5-36c4-be31b971be25, 'name': SearchDatastore_Task, 'duration_secs': 0.008377} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.656637] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b276fa9b-bbeb-488c-a0ed-3756ebb7b3b5 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.663020] env[61957]: DEBUG oslo_vmware.api [None req-be584a65-2286-47ec-9ff6-29a7835a7d50 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Waiting for the task: (returnval){ [ 838.663020] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52cca221-c6bc-494d-bcc0-04ea0def8f3f" [ 838.663020] env[61957]: _type = "Task" [ 838.663020] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.672237] env[61957]: DEBUG oslo_vmware.api [None req-be584a65-2286-47ec-9ff6-29a7835a7d50 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52cca221-c6bc-494d-bcc0-04ea0def8f3f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.777471] env[61957]: DEBUG oslo_vmware.api [None req-3c3df299-fd1f-460a-8cc2-979af1db30d1 tempest-ServersTestJSON-1093633677 tempest-ServersTestJSON-1093633677-project-member] Task: {'id': task-1277474, 'name': Rename_Task} progress is 99%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.948699] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277477, 'name': CreateVM_Task, 'duration_secs': 0.325324} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.949198] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 37a9e1c1-9f6c-4047-9ff5-e141d0ca383e] Created VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 838.950075] env[61957]: DEBUG oslo_concurrency.lockutils [None req-227f29c0-ef6f-447c-b621-3f65bc420b4e tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 838.950669] env[61957]: DEBUG oslo_concurrency.lockutils [None req-227f29c0-ef6f-447c-b621-3f65bc420b4e tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 838.950760] env[61957]: DEBUG oslo_concurrency.lockutils [None req-227f29c0-ef6f-447c-b621-3f65bc420b4e tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 838.951272] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-426064c2-921f-47fa-89a3-5951af312980 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.955953] env[61957]: DEBUG oslo_vmware.api [None req-227f29c0-ef6f-447c-b621-3f65bc420b4e tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] Waiting for the task: (returnval){ [ 838.955953] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52e4da35-b7a7-dbb4-0fd9-970965a56464" [ 838.955953] env[61957]: _type = "Task" [ 838.955953] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.962051] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Instance 2e4a5344-600f-4b61-826e-c15f96b50af2 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61957) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 838.966152] env[61957]: DEBUG oslo_vmware.api [None req-227f29c0-ef6f-447c-b621-3f65bc420b4e tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52e4da35-b7a7-dbb4-0fd9-970965a56464, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.062752] env[61957]: DEBUG oslo_concurrency.lockutils [req-c4581270-22f1-41a6-8f2b-7cbd0443931b req-f9f8d2d2-df39-48e3-a59e-12db11766021 service nova] Releasing lock "refresh_cache-4951132e-7247-4772-8f88-3664c6a7e61e" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 839.126416] env[61957]: DEBUG oslo_vmware.api [None req-82a5645a-fb19-4ad1-aa47-e12b7f048b77 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]527847e5-4067-7cb7-37df-bc60a23af8a7, 'name': SearchDatastore_Task, 'duration_secs': 0.009716} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.126695] env[61957]: DEBUG oslo_concurrency.lockutils [None req-82a5645a-fb19-4ad1-aa47-e12b7f048b77 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 839.126980] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-82a5645a-fb19-4ad1-aa47-e12b7f048b77 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore2] 17f00b20-9d3b-45e6-919d-6fab9999ec77/17f00b20-9d3b-45e6-919d-6fab9999ec77.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 839.127329] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c86d6713-bea2-40e6-afd7-e21112aa1f64 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.133316] env[61957]: DEBUG oslo_vmware.api [None req-82a5645a-fb19-4ad1-aa47-e12b7f048b77 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Waiting for the task: (returnval){ [ 839.133316] env[61957]: value = "task-1277478" [ 839.133316] env[61957]: _type = "Task" [ 839.133316] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.143334] env[61957]: DEBUG oslo_vmware.api [None req-82a5645a-fb19-4ad1-aa47-e12b7f048b77 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Task: {'id': task-1277478, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.173281] env[61957]: DEBUG oslo_vmware.api [None req-be584a65-2286-47ec-9ff6-29a7835a7d50 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52cca221-c6bc-494d-bcc0-04ea0def8f3f, 'name': SearchDatastore_Task, 'duration_secs': 0.012493} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.175816] env[61957]: DEBUG oslo_concurrency.lockutils [None req-be584a65-2286-47ec-9ff6-29a7835a7d50 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Releasing lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 839.176143] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-be584a65-2286-47ec-9ff6-29a7835a7d50 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore1] 37ce46a1-8f9a-4d15-bd81-e40845a0e48a/37ce46a1-8f9a-4d15-bd81-e40845a0e48a.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 839.176832] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5ada732a-3b82-4883-a9f8-d4a957bedcbc {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.185073] env[61957]: DEBUG oslo_vmware.api [None req-be584a65-2286-47ec-9ff6-29a7835a7d50 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Waiting for the task: (returnval){ [ 839.185073] env[61957]: value = "task-1277479" [ 839.185073] env[61957]: _type = "Task" [ 839.185073] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.192405] env[61957]: DEBUG oslo_vmware.api [None req-be584a65-2286-47ec-9ff6-29a7835a7d50 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': task-1277479, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.278428] env[61957]: DEBUG oslo_vmware.api [None req-3c3df299-fd1f-460a-8cc2-979af1db30d1 tempest-ServersTestJSON-1093633677 tempest-ServersTestJSON-1093633677-project-member] Task: {'id': task-1277474, 'name': Rename_Task, 'duration_secs': 1.19596} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.278923] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c3df299-fd1f-460a-8cc2-979af1db30d1 tempest-ServersTestJSON-1093633677 tempest-ServersTestJSON-1093633677-project-member] [instance: ead6aae1-36b5-4f57-9129-3bb02cf103ce] Powering on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 839.279010] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b937a736-4556-426b-879e-e766f30404be {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.286628] env[61957]: DEBUG oslo_vmware.api [None req-3c3df299-fd1f-460a-8cc2-979af1db30d1 tempest-ServersTestJSON-1093633677 tempest-ServersTestJSON-1093633677-project-member] Waiting for the task: (returnval){ [ 839.286628] env[61957]: value = "task-1277480" [ 839.286628] env[61957]: _type = "Task" [ 839.286628] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.295193] env[61957]: DEBUG oslo_vmware.api [None req-3c3df299-fd1f-460a-8cc2-979af1db30d1 tempest-ServersTestJSON-1093633677 tempest-ServersTestJSON-1093633677-project-member] Task: {'id': task-1277480, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.392519] env[61957]: DEBUG nova.network.neutron [req-70e4c08b-85e3-4b41-ad92-bd626f0c83ff req-9b7a8cbb-18e2-4d16-b3f6-3d188d9e2db6 service nova] [instance: 37a9e1c1-9f6c-4047-9ff5-e141d0ca383e] Updated VIF entry in instance network info cache for port aee17432-3f8d-4d87-863f-20a333a4c92f. {{(pid=61957) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 839.392911] env[61957]: DEBUG nova.network.neutron [req-70e4c08b-85e3-4b41-ad92-bd626f0c83ff req-9b7a8cbb-18e2-4d16-b3f6-3d188d9e2db6 service nova] [instance: 37a9e1c1-9f6c-4047-9ff5-e141d0ca383e] Updating instance_info_cache with network_info: [{"id": "aee17432-3f8d-4d87-863f-20a333a4c92f", "address": "fa:16:3e:21:42:d0", "network": {"id": "8ee8bd88-2141-46db-b100-99dc2aaa8574", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-975448626-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "21351f437ea54c9580753dfddd7f47be", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b9aabc7c-0f6c-42eb-bd27-493a1496c0c8", "external-id": "nsx-vlan-transportzone-368", "segmentation_id": 368, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaee17432-3f", "ovs_interfaceid": "aee17432-3f8d-4d87-863f-20a333a4c92f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 839.468245] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Instance e40fa112-4648-428f-a403-b3e3b8319ea8 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61957) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 839.469997] env[61957]: DEBUG oslo_vmware.api [None req-227f29c0-ef6f-447c-b621-3f65bc420b4e tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52e4da35-b7a7-dbb4-0fd9-970965a56464, 'name': SearchDatastore_Task, 'duration_secs': 0.016744} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.471283] env[61957]: DEBUG oslo_concurrency.lockutils [None req-227f29c0-ef6f-447c-b621-3f65bc420b4e tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 839.471537] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-227f29c0-ef6f-447c-b621-3f65bc420b4e tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] [instance: 37a9e1c1-9f6c-4047-9ff5-e141d0ca383e] Processing image 11c76a2c-f705-470a-ba9d-4657858bab38 {{(pid=61957) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 839.471812] env[61957]: DEBUG oslo_concurrency.lockutils [None req-227f29c0-ef6f-447c-b621-3f65bc420b4e tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 839.471964] env[61957]: DEBUG oslo_concurrency.lockutils [None req-227f29c0-ef6f-447c-b621-3f65bc420b4e tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 839.472197] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-227f29c0-ef6f-447c-b621-3f65bc420b4e tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 839.472482] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-cec7fa67-55ac-4169-955d-a6a2562dd3ac {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.483188] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-227f29c0-ef6f-447c-b621-3f65bc420b4e tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 839.483287] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-227f29c0-ef6f-447c-b621-3f65bc420b4e tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61957) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 839.484259] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-456e501b-840d-4507-bad9-c69945914183 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.495996] env[61957]: DEBUG oslo_vmware.api [None req-227f29c0-ef6f-447c-b621-3f65bc420b4e tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] Waiting for the task: (returnval){ [ 839.495996] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]5287edc5-0b4d-44a9-772b-cad0c01d71cf" [ 839.495996] env[61957]: _type = "Task" [ 839.495996] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.513391] env[61957]: DEBUG oslo_vmware.api [None req-227f29c0-ef6f-447c-b621-3f65bc420b4e tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]5287edc5-0b4d-44a9-772b-cad0c01d71cf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.646942] env[61957]: DEBUG oslo_vmware.api [None req-82a5645a-fb19-4ad1-aa47-e12b7f048b77 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Task: {'id': task-1277478, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.695984] env[61957]: DEBUG oslo_vmware.api [None req-be584a65-2286-47ec-9ff6-29a7835a7d50 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': task-1277479, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.51} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.696406] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-be584a65-2286-47ec-9ff6-29a7835a7d50 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore1] 37ce46a1-8f9a-4d15-bd81-e40845a0e48a/37ce46a1-8f9a-4d15-bd81-e40845a0e48a.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 839.696649] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-be584a65-2286-47ec-9ff6-29a7835a7d50 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: 37ce46a1-8f9a-4d15-bd81-e40845a0e48a] Extending root virtual disk to 1048576 {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 839.696990] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f8969589-8cdf-471b-b2fe-b6be9ae39a11 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.704631] env[61957]: DEBUG oslo_vmware.api [None req-be584a65-2286-47ec-9ff6-29a7835a7d50 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Waiting for the task: (returnval){ [ 839.704631] env[61957]: value = "task-1277481" [ 839.704631] env[61957]: _type = "Task" [ 839.704631] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.714221] env[61957]: DEBUG oslo_vmware.api [None req-be584a65-2286-47ec-9ff6-29a7835a7d50 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': task-1277481, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.798075] env[61957]: DEBUG oslo_vmware.api [None req-3c3df299-fd1f-460a-8cc2-979af1db30d1 tempest-ServersTestJSON-1093633677 tempest-ServersTestJSON-1093633677-project-member] Task: {'id': task-1277480, 'name': PowerOnVM_Task} progress is 1%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.897483] env[61957]: DEBUG oslo_concurrency.lockutils [req-70e4c08b-85e3-4b41-ad92-bd626f0c83ff req-9b7a8cbb-18e2-4d16-b3f6-3d188d9e2db6 service nova] Releasing lock "refresh_cache-37a9e1c1-9f6c-4047-9ff5-e141d0ca383e" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 839.972085] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Instance 1a0327c2-5671-4970-9db7-c7cc912d8678 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61957) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 840.007228] env[61957]: DEBUG oslo_vmware.api [None req-227f29c0-ef6f-447c-b621-3f65bc420b4e tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]5287edc5-0b4d-44a9-772b-cad0c01d71cf, 'name': SearchDatastore_Task, 'duration_secs': 0.020774} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.008140] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-69495e8e-dc1f-4da4-958b-325f2a00ba22 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.014071] env[61957]: DEBUG oslo_vmware.api [None req-227f29c0-ef6f-447c-b621-3f65bc420b4e tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] Waiting for the task: (returnval){ [ 840.014071] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52c6d3e5-9e6e-279f-1003-464fb443abe5" [ 840.014071] env[61957]: _type = "Task" [ 840.014071] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.024824] env[61957]: DEBUG oslo_vmware.api [None req-227f29c0-ef6f-447c-b621-3f65bc420b4e tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52c6d3e5-9e6e-279f-1003-464fb443abe5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.143382] env[61957]: DEBUG oslo_vmware.api [None req-82a5645a-fb19-4ad1-aa47-e12b7f048b77 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Task: {'id': task-1277478, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.890217} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.143663] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-82a5645a-fb19-4ad1-aa47-e12b7f048b77 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore2] 17f00b20-9d3b-45e6-919d-6fab9999ec77/17f00b20-9d3b-45e6-919d-6fab9999ec77.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 840.143886] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-82a5645a-fb19-4ad1-aa47-e12b7f048b77 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] [instance: 17f00b20-9d3b-45e6-919d-6fab9999ec77] Extending root virtual disk to 1048576 {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 840.144179] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9cfbffa9-254b-4f1a-9b46-4bdba9a2c21d {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.151298] env[61957]: DEBUG oslo_vmware.api [None req-82a5645a-fb19-4ad1-aa47-e12b7f048b77 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Waiting for the task: (returnval){ [ 840.151298] env[61957]: value = "task-1277482" [ 840.151298] env[61957]: _type = "Task" [ 840.151298] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.160874] env[61957]: DEBUG oslo_vmware.api [None req-82a5645a-fb19-4ad1-aa47-e12b7f048b77 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Task: {'id': task-1277482, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.217469] env[61957]: DEBUG oslo_vmware.api [None req-be584a65-2286-47ec-9ff6-29a7835a7d50 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': task-1277481, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.093714} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.217743] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-be584a65-2286-47ec-9ff6-29a7835a7d50 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: 37ce46a1-8f9a-4d15-bd81-e40845a0e48a] Extended root virtual disk {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 840.218806] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bae874f6-788c-4eb1-9907-e5ce0b0717b9 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.242546] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-be584a65-2286-47ec-9ff6-29a7835a7d50 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: 37ce46a1-8f9a-4d15-bd81-e40845a0e48a] Reconfiguring VM instance instance-0000003a to attach disk [datastore1] 37ce46a1-8f9a-4d15-bd81-e40845a0e48a/37ce46a1-8f9a-4d15-bd81-e40845a0e48a.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 840.242924] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2c0bb37a-e850-4f1b-85fa-8cdf4eaf6553 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.263118] env[61957]: DEBUG oslo_vmware.api [None req-be584a65-2286-47ec-9ff6-29a7835a7d50 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Waiting for the task: (returnval){ [ 840.263118] env[61957]: value = "task-1277483" [ 840.263118] env[61957]: _type = "Task" [ 840.263118] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.271858] env[61957]: DEBUG oslo_vmware.api [None req-be584a65-2286-47ec-9ff6-29a7835a7d50 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': task-1277483, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.296923] env[61957]: DEBUG oslo_vmware.api [None req-3c3df299-fd1f-460a-8cc2-979af1db30d1 tempest-ServersTestJSON-1093633677 tempest-ServersTestJSON-1093633677-project-member] Task: {'id': task-1277480, 'name': PowerOnVM_Task} progress is 64%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.445525] env[61957]: DEBUG nova.compute.manager [None req-b1306d4e-743c-4d53-8fec-4c8136571e40 tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] [instance: e0249936-d616-4ffb-8f77-d8107633c42a] Checking state {{(pid=61957) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 840.446522] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-937f8879-a06f-4d88-97ae-146e842b37d0 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.481236] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Instance 28ab7d23-487f-4ae1-8fe4-58db55b59918 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61957) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 840.527267] env[61957]: DEBUG oslo_vmware.api [None req-227f29c0-ef6f-447c-b621-3f65bc420b4e tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52c6d3e5-9e6e-279f-1003-464fb443abe5, 'name': SearchDatastore_Task, 'duration_secs': 0.023355} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.527435] env[61957]: DEBUG oslo_concurrency.lockutils [None req-227f29c0-ef6f-447c-b621-3f65bc420b4e tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 840.527710] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-227f29c0-ef6f-447c-b621-3f65bc420b4e tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore2] 37a9e1c1-9f6c-4047-9ff5-e141d0ca383e/37a9e1c1-9f6c-4047-9ff5-e141d0ca383e.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 840.528036] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b0807b63-9552-4b5b-9b08-4c0d07ae280c {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.535256] env[61957]: DEBUG oslo_vmware.api [None req-227f29c0-ef6f-447c-b621-3f65bc420b4e tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] Waiting for the task: (returnval){ [ 840.535256] env[61957]: value = "task-1277484" [ 840.535256] env[61957]: _type = "Task" [ 840.535256] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.543706] env[61957]: DEBUG oslo_vmware.api [None req-227f29c0-ef6f-447c-b621-3f65bc420b4e tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] Task: {'id': task-1277484, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.663245] env[61957]: DEBUG oslo_vmware.api [None req-82a5645a-fb19-4ad1-aa47-e12b7f048b77 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Task: {'id': task-1277482, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.279677} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.663525] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-82a5645a-fb19-4ad1-aa47-e12b7f048b77 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] [instance: 17f00b20-9d3b-45e6-919d-6fab9999ec77] Extended root virtual disk {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 840.664732] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ba53018-db56-45c8-b167-a27a7c53729f {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.686130] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-82a5645a-fb19-4ad1-aa47-e12b7f048b77 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] [instance: 17f00b20-9d3b-45e6-919d-6fab9999ec77] Reconfiguring VM instance instance-00000037 to attach disk [datastore2] 17f00b20-9d3b-45e6-919d-6fab9999ec77/17f00b20-9d3b-45e6-919d-6fab9999ec77.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 840.686516] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-590481bd-ff23-481a-a0b5-24176eb0e902 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.707104] env[61957]: DEBUG oslo_vmware.api [None req-82a5645a-fb19-4ad1-aa47-e12b7f048b77 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Waiting for the task: (returnval){ [ 840.707104] env[61957]: value = "task-1277485" [ 840.707104] env[61957]: _type = "Task" [ 840.707104] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.715658] env[61957]: DEBUG oslo_vmware.api [None req-82a5645a-fb19-4ad1-aa47-e12b7f048b77 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Task: {'id': task-1277485, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.772556] env[61957]: DEBUG oslo_vmware.api [None req-be584a65-2286-47ec-9ff6-29a7835a7d50 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': task-1277483, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.799184] env[61957]: DEBUG oslo_vmware.api [None req-3c3df299-fd1f-460a-8cc2-979af1db30d1 tempest-ServersTestJSON-1093633677 tempest-ServersTestJSON-1093633677-project-member] Task: {'id': task-1277480, 'name': PowerOnVM_Task} progress is 68%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.958801] env[61957]: INFO nova.compute.manager [None req-b1306d4e-743c-4d53-8fec-4c8136571e40 tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] [instance: e0249936-d616-4ffb-8f77-d8107633c42a] instance snapshotting [ 840.962239] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2aa6ab88-d0e1-4a11-93cf-2b462dd3e763 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.988878] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Instance d5d10a31-0e78-4ed7-b944-9208138a4861 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61957) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 840.992190] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb5aca80-3628-4038-97da-682a5b9d5b29 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.054037] env[61957]: DEBUG oslo_vmware.api [None req-227f29c0-ef6f-447c-b621-3f65bc420b4e tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] Task: {'id': task-1277484, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.217530] env[61957]: DEBUG oslo_vmware.api [None req-82a5645a-fb19-4ad1-aa47-e12b7f048b77 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Task: {'id': task-1277485, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.272604] env[61957]: DEBUG oslo_vmware.api [None req-be584a65-2286-47ec-9ff6-29a7835a7d50 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': task-1277483, 'name': ReconfigVM_Task, 'duration_secs': 0.756239} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.272604] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-be584a65-2286-47ec-9ff6-29a7835a7d50 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: 37ce46a1-8f9a-4d15-bd81-e40845a0e48a] Reconfigured VM instance instance-0000003a to attach disk [datastore1] 37ce46a1-8f9a-4d15-bd81-e40845a0e48a/37ce46a1-8f9a-4d15-bd81-e40845a0e48a.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 841.273257] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e68a4976-0586-4456-b30d-3d97c394932a {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.279625] env[61957]: DEBUG oslo_vmware.api [None req-be584a65-2286-47ec-9ff6-29a7835a7d50 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Waiting for the task: (returnval){ [ 841.279625] env[61957]: value = "task-1277486" [ 841.279625] env[61957]: _type = "Task" [ 841.279625] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.287061] env[61957]: DEBUG oslo_vmware.api [None req-be584a65-2286-47ec-9ff6-29a7835a7d50 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': task-1277486, 'name': Rename_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.296386] env[61957]: DEBUG oslo_vmware.api [None req-3c3df299-fd1f-460a-8cc2-979af1db30d1 tempest-ServersTestJSON-1093633677 tempest-ServersTestJSON-1093633677-project-member] Task: {'id': task-1277480, 'name': PowerOnVM_Task, 'duration_secs': 1.821172} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.296627] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c3df299-fd1f-460a-8cc2-979af1db30d1 tempest-ServersTestJSON-1093633677 tempest-ServersTestJSON-1093633677-project-member] [instance: ead6aae1-36b5-4f57-9129-3bb02cf103ce] Powered on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 841.296828] env[61957]: INFO nova.compute.manager [None req-3c3df299-fd1f-460a-8cc2-979af1db30d1 tempest-ServersTestJSON-1093633677 tempest-ServersTestJSON-1093633677-project-member] [instance: ead6aae1-36b5-4f57-9129-3bb02cf103ce] Took 10.51 seconds to spawn the instance on the hypervisor. [ 841.297015] env[61957]: DEBUG nova.compute.manager [None req-3c3df299-fd1f-460a-8cc2-979af1db30d1 tempest-ServersTestJSON-1093633677 tempest-ServersTestJSON-1093633677-project-member] [instance: ead6aae1-36b5-4f57-9129-3bb02cf103ce] Checking state {{(pid=61957) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 841.297750] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-028da6fb-6126-4467-af8f-ebec36d71a27 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.496117] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Instance 0d0f2d34-de35-4e80-8d9f-12693add0786 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61957) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 841.503925] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-b1306d4e-743c-4d53-8fec-4c8136571e40 tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] [instance: e0249936-d616-4ffb-8f77-d8107633c42a] Creating Snapshot of the VM instance {{(pid=61957) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 841.504257] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-4166db68-e4c0-4707-a62e-253d6e39e7aa {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.512018] env[61957]: DEBUG oslo_vmware.api [None req-b1306d4e-743c-4d53-8fec-4c8136571e40 tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Waiting for the task: (returnval){ [ 841.512018] env[61957]: value = "task-1277487" [ 841.512018] env[61957]: _type = "Task" [ 841.512018] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.520830] env[61957]: DEBUG oslo_vmware.api [None req-b1306d4e-743c-4d53-8fec-4c8136571e40 tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Task: {'id': task-1277487, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.546628] env[61957]: DEBUG oslo_vmware.api [None req-227f29c0-ef6f-447c-b621-3f65bc420b4e tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] Task: {'id': task-1277484, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.707377} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.546961] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-227f29c0-ef6f-447c-b621-3f65bc420b4e tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore2] 37a9e1c1-9f6c-4047-9ff5-e141d0ca383e/37a9e1c1-9f6c-4047-9ff5-e141d0ca383e.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 841.547255] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-227f29c0-ef6f-447c-b621-3f65bc420b4e tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] [instance: 37a9e1c1-9f6c-4047-9ff5-e141d0ca383e] Extending root virtual disk to 1048576 {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 841.547534] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6e569a04-c19e-487b-87ed-44fb4659417e {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.554048] env[61957]: DEBUG oslo_vmware.api [None req-227f29c0-ef6f-447c-b621-3f65bc420b4e tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] Waiting for the task: (returnval){ [ 841.554048] env[61957]: value = "task-1277488" [ 841.554048] env[61957]: _type = "Task" [ 841.554048] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.562263] env[61957]: DEBUG oslo_vmware.api [None req-227f29c0-ef6f-447c-b621-3f65bc420b4e tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] Task: {'id': task-1277488, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.717439] env[61957]: DEBUG oslo_vmware.api [None req-82a5645a-fb19-4ad1-aa47-e12b7f048b77 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Task: {'id': task-1277485, 'name': ReconfigVM_Task, 'duration_secs': 0.620466} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.717763] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-82a5645a-fb19-4ad1-aa47-e12b7f048b77 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] [instance: 17f00b20-9d3b-45e6-919d-6fab9999ec77] Reconfigured VM instance instance-00000037 to attach disk [datastore2] 17f00b20-9d3b-45e6-919d-6fab9999ec77/17f00b20-9d3b-45e6-919d-6fab9999ec77.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 841.718438] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2cdbc1a5-ffc5-4a5a-9dd6-2afe5f4b4883 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.724577] env[61957]: DEBUG oslo_vmware.api [None req-82a5645a-fb19-4ad1-aa47-e12b7f048b77 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Waiting for the task: (returnval){ [ 841.724577] env[61957]: value = "task-1277489" [ 841.724577] env[61957]: _type = "Task" [ 841.724577] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.734086] env[61957]: DEBUG oslo_vmware.api [None req-82a5645a-fb19-4ad1-aa47-e12b7f048b77 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Task: {'id': task-1277489, 'name': Rename_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.788894] env[61957]: DEBUG oslo_vmware.api [None req-be584a65-2286-47ec-9ff6-29a7835a7d50 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': task-1277486, 'name': Rename_Task, 'duration_secs': 0.139988} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.789363] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-be584a65-2286-47ec-9ff6-29a7835a7d50 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: 37ce46a1-8f9a-4d15-bd81-e40845a0e48a] Powering on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 841.789700] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-76aabcbc-0d91-4926-bc5f-28bbc67039b1 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.796380] env[61957]: DEBUG oslo_vmware.api [None req-be584a65-2286-47ec-9ff6-29a7835a7d50 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Waiting for the task: (returnval){ [ 841.796380] env[61957]: value = "task-1277490" [ 841.796380] env[61957]: _type = "Task" [ 841.796380] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.805587] env[61957]: DEBUG oslo_vmware.api [None req-be584a65-2286-47ec-9ff6-29a7835a7d50 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': task-1277490, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.818964] env[61957]: INFO nova.compute.manager [None req-3c3df299-fd1f-460a-8cc2-979af1db30d1 tempest-ServersTestJSON-1093633677 tempest-ServersTestJSON-1093633677-project-member] [instance: ead6aae1-36b5-4f57-9129-3bb02cf103ce] Took 29.87 seconds to build instance. [ 841.999258] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Instance a8ca1ad6-d636-4fa2-b0ac-53b020e392ce has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61957) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 842.023019] env[61957]: DEBUG oslo_vmware.api [None req-b1306d4e-743c-4d53-8fec-4c8136571e40 tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Task: {'id': task-1277487, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.067216] env[61957]: DEBUG oslo_vmware.api [None req-227f29c0-ef6f-447c-b621-3f65bc420b4e tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] Task: {'id': task-1277488, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.062992} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.067216] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-227f29c0-ef6f-447c-b621-3f65bc420b4e tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] [instance: 37a9e1c1-9f6c-4047-9ff5-e141d0ca383e] Extended root virtual disk {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 842.068029] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-445fa1d2-cd6d-483b-b938-9bf642c9eb23 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.107291] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-227f29c0-ef6f-447c-b621-3f65bc420b4e tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] [instance: 37a9e1c1-9f6c-4047-9ff5-e141d0ca383e] Reconfiguring VM instance instance-0000003b to attach disk [datastore2] 37a9e1c1-9f6c-4047-9ff5-e141d0ca383e/37a9e1c1-9f6c-4047-9ff5-e141d0ca383e.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 842.107878] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-eaabdcad-5963-4276-becd-8818ab53c552 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.137699] env[61957]: DEBUG oslo_vmware.api [None req-227f29c0-ef6f-447c-b621-3f65bc420b4e tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] Waiting for the task: (returnval){ [ 842.137699] env[61957]: value = "task-1277491" [ 842.137699] env[61957]: _type = "Task" [ 842.137699] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.147583] env[61957]: DEBUG oslo_vmware.api [None req-227f29c0-ef6f-447c-b621-3f65bc420b4e tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] Task: {'id': task-1277491, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.236828] env[61957]: DEBUG oslo_vmware.api [None req-82a5645a-fb19-4ad1-aa47-e12b7f048b77 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Task: {'id': task-1277489, 'name': Rename_Task, 'duration_secs': 0.174742} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.237242] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-82a5645a-fb19-4ad1-aa47-e12b7f048b77 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] [instance: 17f00b20-9d3b-45e6-919d-6fab9999ec77] Powering on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 842.237884] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b59e54d0-1047-4026-8800-470aeab7688e {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.245446] env[61957]: DEBUG oslo_vmware.api [None req-82a5645a-fb19-4ad1-aa47-e12b7f048b77 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Waiting for the task: (returnval){ [ 842.245446] env[61957]: value = "task-1277492" [ 842.245446] env[61957]: _type = "Task" [ 842.245446] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.255906] env[61957]: DEBUG oslo_vmware.api [None req-82a5645a-fb19-4ad1-aa47-e12b7f048b77 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Task: {'id': task-1277492, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.306411] env[61957]: DEBUG oslo_vmware.api [None req-be584a65-2286-47ec-9ff6-29a7835a7d50 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': task-1277490, 'name': PowerOnVM_Task, 'duration_secs': 0.494876} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.307017] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-be584a65-2286-47ec-9ff6-29a7835a7d50 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: 37ce46a1-8f9a-4d15-bd81-e40845a0e48a] Powered on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 842.307017] env[61957]: INFO nova.compute.manager [None req-be584a65-2286-47ec-9ff6-29a7835a7d50 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: 37ce46a1-8f9a-4d15-bd81-e40845a0e48a] Took 8.94 seconds to spawn the instance on the hypervisor. [ 842.307173] env[61957]: DEBUG nova.compute.manager [None req-be584a65-2286-47ec-9ff6-29a7835a7d50 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: 37ce46a1-8f9a-4d15-bd81-e40845a0e48a] Checking state {{(pid=61957) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 842.308230] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc222d80-58bb-401c-b082-26309d52784e {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.324894] env[61957]: DEBUG oslo_concurrency.lockutils [None req-3c3df299-fd1f-460a-8cc2-979af1db30d1 tempest-ServersTestJSON-1093633677 tempest-ServersTestJSON-1093633677-project-member] Lock "ead6aae1-36b5-4f57-9129-3bb02cf103ce" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 106.657s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 842.495847] env[61957]: DEBUG nova.compute.manager [req-b64e60a4-22e1-4ffe-86e9-0deaa12ef991 req-1838a23c-2717-486b-824f-fde2d8e8c9d9 service nova] [instance: ead6aae1-36b5-4f57-9129-3bb02cf103ce] Received event network-changed-606b786e-f948-4d1c-a109-6390d9349ea8 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 842.496204] env[61957]: DEBUG nova.compute.manager [req-b64e60a4-22e1-4ffe-86e9-0deaa12ef991 req-1838a23c-2717-486b-824f-fde2d8e8c9d9 service nova] [instance: ead6aae1-36b5-4f57-9129-3bb02cf103ce] Refreshing instance network info cache due to event network-changed-606b786e-f948-4d1c-a109-6390d9349ea8. {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 842.496563] env[61957]: DEBUG oslo_concurrency.lockutils [req-b64e60a4-22e1-4ffe-86e9-0deaa12ef991 req-1838a23c-2717-486b-824f-fde2d8e8c9d9 service nova] Acquiring lock "refresh_cache-ead6aae1-36b5-4f57-9129-3bb02cf103ce" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 842.496822] env[61957]: DEBUG oslo_concurrency.lockutils [req-b64e60a4-22e1-4ffe-86e9-0deaa12ef991 req-1838a23c-2717-486b-824f-fde2d8e8c9d9 service nova] Acquired lock "refresh_cache-ead6aae1-36b5-4f57-9129-3bb02cf103ce" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 842.497117] env[61957]: DEBUG nova.network.neutron [req-b64e60a4-22e1-4ffe-86e9-0deaa12ef991 req-1838a23c-2717-486b-824f-fde2d8e8c9d9 service nova] [instance: ead6aae1-36b5-4f57-9129-3bb02cf103ce] Refreshing network info cache for port 606b786e-f948-4d1c-a109-6390d9349ea8 {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 842.503646] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Instance 736eabcf-70f8-4e1e-9830-270d9971fc9a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61957) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 842.503646] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Total usable vcpus: 48, total allocated vcpus: 7 {{(pid=61957) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 842.503646] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1856MB phys_disk=200GB used_disk=7GB total_vcpus=48 used_vcpus=7 pci_stats=[] {{(pid=61957) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 842.523550] env[61957]: DEBUG oslo_vmware.api [None req-b1306d4e-743c-4d53-8fec-4c8136571e40 tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Task: {'id': task-1277487, 'name': CreateSnapshot_Task, 'duration_secs': 0.645376} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.523963] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-b1306d4e-743c-4d53-8fec-4c8136571e40 tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] [instance: e0249936-d616-4ffb-8f77-d8107633c42a] Created Snapshot of the VM instance {{(pid=61957) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 842.524910] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c709660-8f04-4a8c-8f3b-e56e8776b1b2 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.649724] env[61957]: DEBUG oslo_vmware.api [None req-227f29c0-ef6f-447c-b621-3f65bc420b4e tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] Task: {'id': task-1277491, 'name': ReconfigVM_Task, 'duration_secs': 0.454661} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.650053] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-227f29c0-ef6f-447c-b621-3f65bc420b4e tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] [instance: 37a9e1c1-9f6c-4047-9ff5-e141d0ca383e] Reconfigured VM instance instance-0000003b to attach disk [datastore2] 37a9e1c1-9f6c-4047-9ff5-e141d0ca383e/37a9e1c1-9f6c-4047-9ff5-e141d0ca383e.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 842.651015] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-128ec576-a21a-4195-bcd9-0f0590f60c62 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.657189] env[61957]: DEBUG oslo_vmware.api [None req-227f29c0-ef6f-447c-b621-3f65bc420b4e tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] Waiting for the task: (returnval){ [ 842.657189] env[61957]: value = "task-1277493" [ 842.657189] env[61957]: _type = "Task" [ 842.657189] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.666324] env[61957]: DEBUG oslo_vmware.api [None req-227f29c0-ef6f-447c-b621-3f65bc420b4e tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] Task: {'id': task-1277493, 'name': Rename_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.758419] env[61957]: DEBUG oslo_vmware.api [None req-82a5645a-fb19-4ad1-aa47-e12b7f048b77 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Task: {'id': task-1277492, 'name': PowerOnVM_Task} progress is 64%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.829150] env[61957]: DEBUG nova.compute.manager [None req-85c79741-463e-4bc0-9ed2-4682b88fcc00 tempest-ServersV294TestFqdnHostnames-1426486100 tempest-ServersV294TestFqdnHostnames-1426486100-project-member] [instance: 2e4a5344-600f-4b61-826e-c15f96b50af2] Starting instance... {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 842.833762] env[61957]: INFO nova.compute.manager [None req-be584a65-2286-47ec-9ff6-29a7835a7d50 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: 37ce46a1-8f9a-4d15-bd81-e40845a0e48a] Took 28.95 seconds to build instance. [ 842.873320] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60a9e79e-e72e-4d55-bf9c-eaa9d1bc3d20 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.881151] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a9ca832-2031-494b-abd5-aeb95a0cf217 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.911345] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dba611e1-5ef5-477b-b8fa-88ab25c458cc {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.918862] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7eb55b4-53aa-4b5c-bf62-18d746c882cc {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.933386] env[61957]: DEBUG nova.compute.provider_tree [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Updating inventory in ProviderTree for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 843.047010] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-b1306d4e-743c-4d53-8fec-4c8136571e40 tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] [instance: e0249936-d616-4ffb-8f77-d8107633c42a] Creating linked-clone VM from snapshot {{(pid=61957) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 843.047369] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-3d542fcd-343e-4b90-b3e9-0c16f1ee5613 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.055585] env[61957]: DEBUG oslo_vmware.api [None req-b1306d4e-743c-4d53-8fec-4c8136571e40 tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Waiting for the task: (returnval){ [ 843.055585] env[61957]: value = "task-1277494" [ 843.055585] env[61957]: _type = "Task" [ 843.055585] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.064094] env[61957]: DEBUG oslo_vmware.api [None req-b1306d4e-743c-4d53-8fec-4c8136571e40 tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Task: {'id': task-1277494, 'name': CloneVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.168946] env[61957]: DEBUG oslo_vmware.api [None req-227f29c0-ef6f-447c-b621-3f65bc420b4e tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] Task: {'id': task-1277493, 'name': Rename_Task, 'duration_secs': 0.197178} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.169471] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-227f29c0-ef6f-447c-b621-3f65bc420b4e tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] [instance: 37a9e1c1-9f6c-4047-9ff5-e141d0ca383e] Powering on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 843.169972] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8256dfd0-7659-43b3-841c-74fec1a9d763 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.180194] env[61957]: DEBUG oslo_vmware.api [None req-227f29c0-ef6f-447c-b621-3f65bc420b4e tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] Waiting for the task: (returnval){ [ 843.180194] env[61957]: value = "task-1277495" [ 843.180194] env[61957]: _type = "Task" [ 843.180194] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.187682] env[61957]: DEBUG oslo_vmware.api [None req-227f29c0-ef6f-447c-b621-3f65bc420b4e tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] Task: {'id': task-1277495, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.229688] env[61957]: DEBUG nova.network.neutron [req-b64e60a4-22e1-4ffe-86e9-0deaa12ef991 req-1838a23c-2717-486b-824f-fde2d8e8c9d9 service nova] [instance: ead6aae1-36b5-4f57-9129-3bb02cf103ce] Updated VIF entry in instance network info cache for port 606b786e-f948-4d1c-a109-6390d9349ea8. {{(pid=61957) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 843.230884] env[61957]: DEBUG nova.network.neutron [req-b64e60a4-22e1-4ffe-86e9-0deaa12ef991 req-1838a23c-2717-486b-824f-fde2d8e8c9d9 service nova] [instance: ead6aae1-36b5-4f57-9129-3bb02cf103ce] Updating instance_info_cache with network_info: [{"id": "606b786e-f948-4d1c-a109-6390d9349ea8", "address": "fa:16:3e:fb:fd:15", "network": {"id": "200da3c5-aacb-42e6-b638-65442010d6c0", "bridge": "br-int", "label": "tempest-ServersTestJSON-1424742350-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.253", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1a127f9f057c43d2bddf261d2d875b63", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dacd109c-2442-41b8-b612-7ed3efbdaa94", "external-id": "nsx-vlan-transportzone-940", "segmentation_id": 940, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap606b786e-f9", "ovs_interfaceid": "606b786e-f948-4d1c-a109-6390d9349ea8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 843.257224] env[61957]: DEBUG oslo_vmware.api [None req-82a5645a-fb19-4ad1-aa47-e12b7f048b77 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Task: {'id': task-1277492, 'name': PowerOnVM_Task} progress is 68%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.343024] env[61957]: DEBUG oslo_concurrency.lockutils [None req-be584a65-2286-47ec-9ff6-29a7835a7d50 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Lock "37ce46a1-8f9a-4d15-bd81-e40845a0e48a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 103.246s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 843.358510] env[61957]: DEBUG oslo_concurrency.lockutils [None req-85c79741-463e-4bc0-9ed2-4682b88fcc00 tempest-ServersV294TestFqdnHostnames-1426486100 tempest-ServersV294TestFqdnHostnames-1426486100-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 843.455143] env[61957]: ERROR nova.scheduler.client.report [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] [req-e33c351b-b283-41d8-ad92-268a64f8fce7] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 0ceb6c9e-61c6-496d-8579-9d32627e96da. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-e33c351b-b283-41d8-ad92-268a64f8fce7"}]} [ 843.472521] env[61957]: DEBUG nova.scheduler.client.report [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Refreshing inventories for resource provider 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 843.488298] env[61957]: DEBUG nova.scheduler.client.report [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Updating ProviderTree inventory for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 843.488532] env[61957]: DEBUG nova.compute.provider_tree [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Updating inventory in ProviderTree for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 843.503845] env[61957]: DEBUG nova.scheduler.client.report [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Refreshing aggregate associations for resource provider 0ceb6c9e-61c6-496d-8579-9d32627e96da, aggregates: None {{(pid=61957) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 843.525273] env[61957]: DEBUG nova.scheduler.client.report [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Refreshing trait associations for resource provider 0ceb6c9e-61c6-496d-8579-9d32627e96da, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO,HW_ARCH_X86_64,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=61957) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 843.568612] env[61957]: DEBUG oslo_vmware.api [None req-b1306d4e-743c-4d53-8fec-4c8136571e40 tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Task: {'id': task-1277494, 'name': CloneVM_Task} progress is 94%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.688370] env[61957]: DEBUG oslo_vmware.api [None req-227f29c0-ef6f-447c-b621-3f65bc420b4e tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] Task: {'id': task-1277495, 'name': PowerOnVM_Task} progress is 1%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.733745] env[61957]: DEBUG oslo_concurrency.lockutils [req-b64e60a4-22e1-4ffe-86e9-0deaa12ef991 req-1838a23c-2717-486b-824f-fde2d8e8c9d9 service nova] Releasing lock "refresh_cache-ead6aae1-36b5-4f57-9129-3bb02cf103ce" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 843.756272] env[61957]: DEBUG oslo_vmware.api [None req-82a5645a-fb19-4ad1-aa47-e12b7f048b77 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Task: {'id': task-1277492, 'name': PowerOnVM_Task, 'duration_secs': 1.165958} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.759412] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-82a5645a-fb19-4ad1-aa47-e12b7f048b77 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] [instance: 17f00b20-9d3b-45e6-919d-6fab9999ec77] Powered on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 843.759667] env[61957]: DEBUG nova.compute.manager [None req-82a5645a-fb19-4ad1-aa47-e12b7f048b77 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] [instance: 17f00b20-9d3b-45e6-919d-6fab9999ec77] Checking state {{(pid=61957) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 843.761027] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e88cc9a0-a2c8-4c6f-a744-64821daaccd5 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.847069] env[61957]: DEBUG nova.compute.manager [None req-582b6bd7-3488-48cf-a2f8-f44952b10345 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] [instance: e40fa112-4648-428f-a403-b3e3b8319ea8] Starting instance... {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 843.891387] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc90411b-21e6-47f5-aa6d-bd84aba6fa8b {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.901366] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95767735-b262-4312-8eca-95f07767ab9a {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.936798] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4822369b-bed6-40eb-8b46-b60b76193b14 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.946138] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54c20f72-c938-4f09-9640-a9b2f439833e {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.961195] env[61957]: DEBUG nova.compute.provider_tree [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Updating inventory in ProviderTree for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 844.066995] env[61957]: DEBUG oslo_vmware.api [None req-b1306d4e-743c-4d53-8fec-4c8136571e40 tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Task: {'id': task-1277494, 'name': CloneVM_Task} progress is 95%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.189941] env[61957]: DEBUG oslo_vmware.api [None req-227f29c0-ef6f-447c-b621-3f65bc420b4e tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] Task: {'id': task-1277495, 'name': PowerOnVM_Task} progress is 64%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.282148] env[61957]: DEBUG oslo_concurrency.lockutils [None req-82a5645a-fb19-4ad1-aa47-e12b7f048b77 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 844.370718] env[61957]: DEBUG oslo_concurrency.lockutils [None req-582b6bd7-3488-48cf-a2f8-f44952b10345 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 844.481279] env[61957]: INFO nova.compute.manager [None req-d096c609-3f99-4aff-8b69-219b9552f608 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: 37ce46a1-8f9a-4d15-bd81-e40845a0e48a] Rebuilding instance [ 844.497371] env[61957]: DEBUG nova.scheduler.client.report [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Updated inventory for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da with generation 80 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 844.497602] env[61957]: DEBUG nova.compute.provider_tree [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Updating resource provider 0ceb6c9e-61c6-496d-8579-9d32627e96da generation from 80 to 81 during operation: update_inventory {{(pid=61957) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 844.497754] env[61957]: DEBUG nova.compute.provider_tree [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Updating inventory in ProviderTree for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 844.525940] env[61957]: DEBUG nova.compute.manager [None req-d096c609-3f99-4aff-8b69-219b9552f608 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: 37ce46a1-8f9a-4d15-bd81-e40845a0e48a] Checking state {{(pid=61957) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 844.527041] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da1f2db5-943d-4d72-8bdb-6967b28b98ff {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.567551] env[61957]: DEBUG oslo_vmware.api [None req-b1306d4e-743c-4d53-8fec-4c8136571e40 tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Task: {'id': task-1277494, 'name': CloneVM_Task, 'duration_secs': 1.247939} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.567912] env[61957]: INFO nova.virt.vmwareapi.vmops [None req-b1306d4e-743c-4d53-8fec-4c8136571e40 tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] [instance: e0249936-d616-4ffb-8f77-d8107633c42a] Created linked-clone VM from snapshot [ 844.568803] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7d17113-dd52-4442-8fb6-68823b520a6b {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.577654] env[61957]: DEBUG nova.virt.vmwareapi.images [None req-b1306d4e-743c-4d53-8fec-4c8136571e40 tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] [instance: e0249936-d616-4ffb-8f77-d8107633c42a] Uploading image 3e563c87-a879-4f60-8e71-1dde6db914a4 {{(pid=61957) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 844.597775] env[61957]: DEBUG oslo_vmware.rw_handles [None req-b1306d4e-743c-4d53-8fec-4c8136571e40 tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 844.597775] env[61957]: value = "vm-274503" [ 844.597775] env[61957]: _type = "VirtualMachine" [ 844.597775] env[61957]: }. {{(pid=61957) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 844.598103] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-de1d512d-a029-416b-956b-47227ddaafdc {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.605103] env[61957]: DEBUG oslo_vmware.rw_handles [None req-b1306d4e-743c-4d53-8fec-4c8136571e40 tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Lease: (returnval){ [ 844.605103] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]529ebb78-8767-8596-08f6-066e86a0a9e6" [ 844.605103] env[61957]: _type = "HttpNfcLease" [ 844.605103] env[61957]: } obtained for exporting VM: (result){ [ 844.605103] env[61957]: value = "vm-274503" [ 844.605103] env[61957]: _type = "VirtualMachine" [ 844.605103] env[61957]: }. {{(pid=61957) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 844.605353] env[61957]: DEBUG oslo_vmware.api [None req-b1306d4e-743c-4d53-8fec-4c8136571e40 tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Waiting for the lease: (returnval){ [ 844.605353] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]529ebb78-8767-8596-08f6-066e86a0a9e6" [ 844.605353] env[61957]: _type = "HttpNfcLease" [ 844.605353] env[61957]: } to be ready. {{(pid=61957) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 844.612224] env[61957]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 844.612224] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]529ebb78-8767-8596-08f6-066e86a0a9e6" [ 844.612224] env[61957]: _type = "HttpNfcLease" [ 844.612224] env[61957]: } is initializing. {{(pid=61957) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 844.671048] env[61957]: DEBUG oslo_concurrency.lockutils [None req-afaaf689-a0df-401a-b5c9-bce0445ce0ad tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Acquiring lock "17f00b20-9d3b-45e6-919d-6fab9999ec77" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 844.671048] env[61957]: DEBUG oslo_concurrency.lockutils [None req-afaaf689-a0df-401a-b5c9-bce0445ce0ad tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Lock "17f00b20-9d3b-45e6-919d-6fab9999ec77" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 844.671048] env[61957]: DEBUG oslo_concurrency.lockutils [None req-afaaf689-a0df-401a-b5c9-bce0445ce0ad tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Acquiring lock "17f00b20-9d3b-45e6-919d-6fab9999ec77-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 844.671284] env[61957]: DEBUG oslo_concurrency.lockutils [None req-afaaf689-a0df-401a-b5c9-bce0445ce0ad tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Lock "17f00b20-9d3b-45e6-919d-6fab9999ec77-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 844.671391] env[61957]: DEBUG oslo_concurrency.lockutils [None req-afaaf689-a0df-401a-b5c9-bce0445ce0ad tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Lock "17f00b20-9d3b-45e6-919d-6fab9999ec77-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 844.673638] env[61957]: INFO nova.compute.manager [None req-afaaf689-a0df-401a-b5c9-bce0445ce0ad tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] [instance: 17f00b20-9d3b-45e6-919d-6fab9999ec77] Terminating instance [ 844.675303] env[61957]: DEBUG oslo_concurrency.lockutils [None req-afaaf689-a0df-401a-b5c9-bce0445ce0ad tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Acquiring lock "refresh_cache-17f00b20-9d3b-45e6-919d-6fab9999ec77" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 844.675465] env[61957]: DEBUG oslo_concurrency.lockutils [None req-afaaf689-a0df-401a-b5c9-bce0445ce0ad tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Acquired lock "refresh_cache-17f00b20-9d3b-45e6-919d-6fab9999ec77" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 844.675633] env[61957]: DEBUG nova.network.neutron [None req-afaaf689-a0df-401a-b5c9-bce0445ce0ad tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] [instance: 17f00b20-9d3b-45e6-919d-6fab9999ec77] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 844.690053] env[61957]: DEBUG oslo_vmware.api [None req-227f29c0-ef6f-447c-b621-3f65bc420b4e tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] Task: {'id': task-1277495, 'name': PowerOnVM_Task, 'duration_secs': 1.246444} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.690335] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-227f29c0-ef6f-447c-b621-3f65bc420b4e tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] [instance: 37a9e1c1-9f6c-4047-9ff5-e141d0ca383e] Powered on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 844.690543] env[61957]: INFO nova.compute.manager [None req-227f29c0-ef6f-447c-b621-3f65bc420b4e tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] [instance: 37a9e1c1-9f6c-4047-9ff5-e141d0ca383e] Took 8.78 seconds to spawn the instance on the hypervisor. [ 844.690725] env[61957]: DEBUG nova.compute.manager [None req-227f29c0-ef6f-447c-b621-3f65bc420b4e tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] [instance: 37a9e1c1-9f6c-4047-9ff5-e141d0ca383e] Checking state {{(pid=61957) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 844.691529] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4801361-d00d-4552-9947-8474a53f032b {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.004657] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61957) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 845.004951] env[61957]: DEBUG oslo_concurrency.lockutils [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 11.122s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 845.005279] env[61957]: DEBUG oslo_concurrency.lockutils [None req-a20e46ee-0368-4600-9ced-5e54f65fe0cc tempest-InstanceActionsV221TestJSON-1768084888 tempest-InstanceActionsV221TestJSON-1768084888-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 27.215s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 845.005521] env[61957]: DEBUG oslo_concurrency.lockutils [None req-a20e46ee-0368-4600-9ced-5e54f65fe0cc tempest-InstanceActionsV221TestJSON-1768084888 tempest-InstanceActionsV221TestJSON-1768084888-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 845.007465] env[61957]: DEBUG oslo_concurrency.lockutils [None req-e4109c08-00a8-4666-8e84-183191cbc2fb tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.169s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 845.008958] env[61957]: INFO nova.compute.claims [None req-e4109c08-00a8-4666-8e84-183191cbc2fb tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: 0adae8e1-8c2f-4110-805b-1f286debc833] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 845.011831] env[61957]: DEBUG oslo_service.periodic_task [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=61957) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 845.012037] env[61957]: DEBUG nova.compute.manager [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Cleaning up deleted instances {{(pid=61957) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11226}} [ 845.030621] env[61957]: INFO nova.scheduler.client.report [None req-a20e46ee-0368-4600-9ced-5e54f65fe0cc tempest-InstanceActionsV221TestJSON-1768084888 tempest-InstanceActionsV221TestJSON-1768084888-project-member] Deleted allocations for instance 6e7bd89c-2c2a-450a-9858-3526d96c28ab [ 845.040714] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-d096c609-3f99-4aff-8b69-219b9552f608 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: 37ce46a1-8f9a-4d15-bd81-e40845a0e48a] Powering off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 845.041318] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ff05b0da-61a8-4050-8c37-1c3fcd29682c {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.050132] env[61957]: DEBUG oslo_vmware.api [None req-d096c609-3f99-4aff-8b69-219b9552f608 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Waiting for the task: (returnval){ [ 845.050132] env[61957]: value = "task-1277497" [ 845.050132] env[61957]: _type = "Task" [ 845.050132] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.058754] env[61957]: DEBUG oslo_vmware.api [None req-d096c609-3f99-4aff-8b69-219b9552f608 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': task-1277497, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.113855] env[61957]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 845.113855] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]529ebb78-8767-8596-08f6-066e86a0a9e6" [ 845.113855] env[61957]: _type = "HttpNfcLease" [ 845.113855] env[61957]: } is ready. {{(pid=61957) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 845.114363] env[61957]: DEBUG oslo_vmware.rw_handles [None req-b1306d4e-743c-4d53-8fec-4c8136571e40 tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 845.114363] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]529ebb78-8767-8596-08f6-066e86a0a9e6" [ 845.114363] env[61957]: _type = "HttpNfcLease" [ 845.114363] env[61957]: }. {{(pid=61957) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 845.115165] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b78c6a01-c4a5-4862-b8f0-325cefac1456 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.123156] env[61957]: DEBUG oslo_vmware.rw_handles [None req-b1306d4e-743c-4d53-8fec-4c8136571e40 tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52382179-d7d4-71d0-9e9f-1f0dce243f6a/disk-0.vmdk from lease info. {{(pid=61957) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 845.123156] env[61957]: DEBUG oslo_vmware.rw_handles [None req-b1306d4e-743c-4d53-8fec-4c8136571e40 tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52382179-d7d4-71d0-9e9f-1f0dce243f6a/disk-0.vmdk for reading. {{(pid=61957) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 845.195892] env[61957]: DEBUG nova.network.neutron [None req-afaaf689-a0df-401a-b5c9-bce0445ce0ad tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] [instance: 17f00b20-9d3b-45e6-919d-6fab9999ec77] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 845.208156] env[61957]: INFO nova.compute.manager [None req-227f29c0-ef6f-447c-b621-3f65bc420b4e tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] [instance: 37a9e1c1-9f6c-4047-9ff5-e141d0ca383e] Took 29.88 seconds to build instance. [ 845.243557] env[61957]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-97e6feeb-821c-4097-abe1-54af9a954e62 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.264273] env[61957]: DEBUG nova.network.neutron [None req-afaaf689-a0df-401a-b5c9-bce0445ce0ad tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] [instance: 17f00b20-9d3b-45e6-919d-6fab9999ec77] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 845.517612] env[61957]: DEBUG nova.compute.manager [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] There are 4 instances to clean {{(pid=61957) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11235}} [ 845.517992] env[61957]: DEBUG nova.compute.manager [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] [instance: 7f3850c0-a455-4e5b-a329-fb3b02f9725f] Instance has had 0 of 5 cleanup attempts {{(pid=61957) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 845.539440] env[61957]: DEBUG oslo_concurrency.lockutils [None req-a20e46ee-0368-4600-9ced-5e54f65fe0cc tempest-InstanceActionsV221TestJSON-1768084888 tempest-InstanceActionsV221TestJSON-1768084888-project-member] Lock "6e7bd89c-2c2a-450a-9858-3526d96c28ab" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 30.830s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 845.561828] env[61957]: DEBUG oslo_vmware.api [None req-d096c609-3f99-4aff-8b69-219b9552f608 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': task-1277497, 'name': PowerOffVM_Task, 'duration_secs': 0.325781} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.562703] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-d096c609-3f99-4aff-8b69-219b9552f608 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: 37ce46a1-8f9a-4d15-bd81-e40845a0e48a] Powered off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 845.562922] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-d096c609-3f99-4aff-8b69-219b9552f608 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: 37ce46a1-8f9a-4d15-bd81-e40845a0e48a] Destroying instance {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 845.564759] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76393d19-7886-4ca4-bd82-72897441481b {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.572436] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-d096c609-3f99-4aff-8b69-219b9552f608 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: 37ce46a1-8f9a-4d15-bd81-e40845a0e48a] Unregistering the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 845.572750] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-59d79cda-58ed-4785-bf7f-16a44bf51f75 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.710279] env[61957]: DEBUG oslo_concurrency.lockutils [None req-227f29c0-ef6f-447c-b621-3f65bc420b4e tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] Lock "37a9e1c1-9f6c-4047-9ff5-e141d0ca383e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 101.680s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 845.768524] env[61957]: DEBUG oslo_concurrency.lockutils [None req-afaaf689-a0df-401a-b5c9-bce0445ce0ad tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Releasing lock "refresh_cache-17f00b20-9d3b-45e6-919d-6fab9999ec77" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 845.769201] env[61957]: DEBUG nova.compute.manager [None req-afaaf689-a0df-401a-b5c9-bce0445ce0ad tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] [instance: 17f00b20-9d3b-45e6-919d-6fab9999ec77] Start destroying the instance on the hypervisor. {{(pid=61957) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 845.769573] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-afaaf689-a0df-401a-b5c9-bce0445ce0ad tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] [instance: 17f00b20-9d3b-45e6-919d-6fab9999ec77] Destroying instance {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 845.770962] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1dcc5d83-9632-40a8-bcf8-3c2ceb14112a {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.781258] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-afaaf689-a0df-401a-b5c9-bce0445ce0ad tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] [instance: 17f00b20-9d3b-45e6-919d-6fab9999ec77] Powering off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 845.781680] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-47be6a84-ea94-46d4-ae9a-21cd98d2047e {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.792042] env[61957]: DEBUG oslo_vmware.api [None req-afaaf689-a0df-401a-b5c9-bce0445ce0ad tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Waiting for the task: (returnval){ [ 845.792042] env[61957]: value = "task-1277499" [ 845.792042] env[61957]: _type = "Task" [ 845.792042] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.800035] env[61957]: DEBUG oslo_vmware.api [None req-afaaf689-a0df-401a-b5c9-bce0445ce0ad tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Task: {'id': task-1277499, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.025389] env[61957]: DEBUG nova.compute.manager [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] [instance: 188b0fd3-5d71-4feb-aca5-75a2bd28895a] Instance has had 0 of 5 cleanup attempts {{(pid=61957) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 846.212955] env[61957]: DEBUG nova.compute.manager [None req-06bfc4e2-99e2-4510-98f7-f5b198d6231a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 1a0327c2-5671-4970-9db7-c7cc912d8678] Starting instance... {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 846.301319] env[61957]: DEBUG oslo_vmware.api [None req-afaaf689-a0df-401a-b5c9-bce0445ce0ad tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Task: {'id': task-1277499, 'name': PowerOffVM_Task, 'duration_secs': 0.242212} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.304340] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-afaaf689-a0df-401a-b5c9-bce0445ce0ad tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] [instance: 17f00b20-9d3b-45e6-919d-6fab9999ec77] Powered off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 846.304605] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-afaaf689-a0df-401a-b5c9-bce0445ce0ad tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] [instance: 17f00b20-9d3b-45e6-919d-6fab9999ec77] Unregistering the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 846.305524] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-175d0c82-405c-4ad4-81ea-825eb6ece739 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.331878] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-afaaf689-a0df-401a-b5c9-bce0445ce0ad tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] [instance: 17f00b20-9d3b-45e6-919d-6fab9999ec77] Unregistered the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 846.332273] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-afaaf689-a0df-401a-b5c9-bce0445ce0ad tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] [instance: 17f00b20-9d3b-45e6-919d-6fab9999ec77] Deleting contents of the VM from datastore datastore2 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 846.332751] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-afaaf689-a0df-401a-b5c9-bce0445ce0ad tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Deleting the datastore file [datastore2] 17f00b20-9d3b-45e6-919d-6fab9999ec77 {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 846.332811] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-79ce8c11-4813-423d-97fa-f0a14f98ef04 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.341285] env[61957]: DEBUG oslo_vmware.api [None req-afaaf689-a0df-401a-b5c9-bce0445ce0ad tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Waiting for the task: (returnval){ [ 846.341285] env[61957]: value = "task-1277501" [ 846.341285] env[61957]: _type = "Task" [ 846.341285] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.351661] env[61957]: DEBUG oslo_vmware.api [None req-afaaf689-a0df-401a-b5c9-bce0445ce0ad tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Task: {'id': task-1277501, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.533528] env[61957]: DEBUG nova.compute.manager [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] [instance: 4ad746d0-4076-4c97-8ea9-20e49fda59ca] Instance has had 0 of 5 cleanup attempts {{(pid=61957) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 846.577333] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fd8e1e6-12cb-4ccc-9feb-23aed6e54e6e {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.586387] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f81efda-37cd-4e85-ada7-f6248e7cd698 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.623456] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09bcb33f-1eed-499d-af40-fdff7456757d {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.632832] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa489ed2-1a92-48fd-ad89-f73b077f7dab {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.648551] env[61957]: DEBUG nova.compute.provider_tree [None req-e4109c08-00a8-4666-8e84-183191cbc2fb tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Updating inventory in ProviderTree for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 846.740035] env[61957]: DEBUG oslo_concurrency.lockutils [None req-06bfc4e2-99e2-4510-98f7-f5b198d6231a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 846.853477] env[61957]: DEBUG oslo_vmware.api [None req-afaaf689-a0df-401a-b5c9-bce0445ce0ad tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Task: {'id': task-1277501, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.197951} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.853632] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-afaaf689-a0df-401a-b5c9-bce0445ce0ad tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Deleted the datastore file {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 846.853760] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-afaaf689-a0df-401a-b5c9-bce0445ce0ad tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] [instance: 17f00b20-9d3b-45e6-919d-6fab9999ec77] Deleted contents of the VM from datastore datastore2 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 846.853890] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-afaaf689-a0df-401a-b5c9-bce0445ce0ad tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] [instance: 17f00b20-9d3b-45e6-919d-6fab9999ec77] Instance destroyed {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 846.854110] env[61957]: INFO nova.compute.manager [None req-afaaf689-a0df-401a-b5c9-bce0445ce0ad tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] [instance: 17f00b20-9d3b-45e6-919d-6fab9999ec77] Took 1.08 seconds to destroy the instance on the hypervisor. [ 846.854432] env[61957]: DEBUG oslo.service.loopingcall [None req-afaaf689-a0df-401a-b5c9-bce0445ce0ad tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 846.854694] env[61957]: DEBUG nova.compute.manager [-] [instance: 17f00b20-9d3b-45e6-919d-6fab9999ec77] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 846.854805] env[61957]: DEBUG nova.network.neutron [-] [instance: 17f00b20-9d3b-45e6-919d-6fab9999ec77] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 846.878694] env[61957]: DEBUG nova.network.neutron [-] [instance: 17f00b20-9d3b-45e6-919d-6fab9999ec77] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 847.037998] env[61957]: DEBUG nova.compute.manager [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] [instance: c5af147e-3526-4014-98bc-7ad163ae89d9] Instance has had 0 of 5 cleanup attempts {{(pid=61957) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 847.210604] env[61957]: DEBUG nova.scheduler.client.report [None req-e4109c08-00a8-4666-8e84-183191cbc2fb tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Updated inventory for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da with generation 81 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 847.210927] env[61957]: DEBUG nova.compute.provider_tree [None req-e4109c08-00a8-4666-8e84-183191cbc2fb tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Updating resource provider 0ceb6c9e-61c6-496d-8579-9d32627e96da generation from 81 to 82 during operation: update_inventory {{(pid=61957) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 847.211144] env[61957]: DEBUG nova.compute.provider_tree [None req-e4109c08-00a8-4666-8e84-183191cbc2fb tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Updating inventory in ProviderTree for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 847.381634] env[61957]: DEBUG nova.network.neutron [-] [instance: 17f00b20-9d3b-45e6-919d-6fab9999ec77] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 847.542056] env[61957]: DEBUG oslo_service.periodic_task [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=61957) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 847.542255] env[61957]: DEBUG nova.compute.manager [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Cleaning up deleted instances with incomplete migration {{(pid=61957) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11264}} [ 847.717752] env[61957]: DEBUG oslo_concurrency.lockutils [None req-e4109c08-00a8-4666-8e84-183191cbc2fb tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.710s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 847.718381] env[61957]: DEBUG nova.compute.manager [None req-e4109c08-00a8-4666-8e84-183191cbc2fb tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: 0adae8e1-8c2f-4110-805b-1f286debc833] Start building networks asynchronously for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 847.721499] env[61957]: DEBUG oslo_concurrency.lockutils [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.145s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 847.725030] env[61957]: INFO nova.compute.claims [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: 6438fe40-046c-45d5-9986-8f182ecde49f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 847.769421] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-d096c609-3f99-4aff-8b69-219b9552f608 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: 37ce46a1-8f9a-4d15-bd81-e40845a0e48a] Unregistered the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 847.769756] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-d096c609-3f99-4aff-8b69-219b9552f608 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: 37ce46a1-8f9a-4d15-bd81-e40845a0e48a] Deleting contents of the VM from datastore datastore1 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 847.769939] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-d096c609-3f99-4aff-8b69-219b9552f608 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Deleting the datastore file [datastore1] 37ce46a1-8f9a-4d15-bd81-e40845a0e48a {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 847.770255] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b2adf38d-cf5d-4abd-bdf2-1ba0044f4f62 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.776998] env[61957]: DEBUG oslo_vmware.api [None req-d096c609-3f99-4aff-8b69-219b9552f608 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Waiting for the task: (returnval){ [ 847.776998] env[61957]: value = "task-1277502" [ 847.776998] env[61957]: _type = "Task" [ 847.776998] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.785105] env[61957]: DEBUG oslo_vmware.api [None req-d096c609-3f99-4aff-8b69-219b9552f608 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': task-1277502, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.885318] env[61957]: INFO nova.compute.manager [-] [instance: 17f00b20-9d3b-45e6-919d-6fab9999ec77] Took 1.03 seconds to deallocate network for instance. [ 848.045466] env[61957]: DEBUG oslo_service.periodic_task [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=61957) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 848.227714] env[61957]: DEBUG nova.compute.utils [None req-e4109c08-00a8-4666-8e84-183191cbc2fb tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Using /dev/sd instead of None {{(pid=61957) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 848.231667] env[61957]: DEBUG nova.compute.manager [None req-e4109c08-00a8-4666-8e84-183191cbc2fb tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: 0adae8e1-8c2f-4110-805b-1f286debc833] Allocating IP information in the background. {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 848.231667] env[61957]: DEBUG nova.network.neutron [None req-e4109c08-00a8-4666-8e84-183191cbc2fb tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: 0adae8e1-8c2f-4110-805b-1f286debc833] allocate_for_instance() {{(pid=61957) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 848.287793] env[61957]: DEBUG oslo_vmware.api [None req-d096c609-3f99-4aff-8b69-219b9552f608 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': task-1277502, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.208562} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.288046] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-d096c609-3f99-4aff-8b69-219b9552f608 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Deleted the datastore file {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 848.288245] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-d096c609-3f99-4aff-8b69-219b9552f608 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: 37ce46a1-8f9a-4d15-bd81-e40845a0e48a] Deleted contents of the VM from datastore datastore1 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 848.288452] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-d096c609-3f99-4aff-8b69-219b9552f608 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: 37ce46a1-8f9a-4d15-bd81-e40845a0e48a] Instance destroyed {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 848.296722] env[61957]: DEBUG nova.policy [None req-e4109c08-00a8-4666-8e84-183191cbc2fb tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '76e1754f96ed438ea0ed91ad337419f1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '975189ebb3cc4cdb9391880f0c9ba6ba', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61957) authorize /opt/stack/nova/nova/policy.py:203}} [ 848.393716] env[61957]: DEBUG oslo_concurrency.lockutils [None req-afaaf689-a0df-401a-b5c9-bce0445ce0ad tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 848.576142] env[61957]: DEBUG nova.network.neutron [None req-e4109c08-00a8-4666-8e84-183191cbc2fb tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: 0adae8e1-8c2f-4110-805b-1f286debc833] Successfully created port: 06b59d3b-7dab-4254-9ccd-cbfcce2fce87 {{(pid=61957) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 848.732040] env[61957]: DEBUG nova.compute.manager [None req-e4109c08-00a8-4666-8e84-183191cbc2fb tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: 0adae8e1-8c2f-4110-805b-1f286debc833] Start building block device mappings for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 848.993346] env[61957]: DEBUG oslo_concurrency.lockutils [None req-0d2ed604-6dce-44bf-b238-71a34c40662a tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] Acquiring lock "37a9e1c1-9f6c-4047-9ff5-e141d0ca383e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 848.993638] env[61957]: DEBUG oslo_concurrency.lockutils [None req-0d2ed604-6dce-44bf-b238-71a34c40662a tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] Lock "37a9e1c1-9f6c-4047-9ff5-e141d0ca383e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 848.993854] env[61957]: DEBUG oslo_concurrency.lockutils [None req-0d2ed604-6dce-44bf-b238-71a34c40662a tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] Acquiring lock "37a9e1c1-9f6c-4047-9ff5-e141d0ca383e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 848.994060] env[61957]: DEBUG oslo_concurrency.lockutils [None req-0d2ed604-6dce-44bf-b238-71a34c40662a tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] Lock "37a9e1c1-9f6c-4047-9ff5-e141d0ca383e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 848.994255] env[61957]: DEBUG oslo_concurrency.lockutils [None req-0d2ed604-6dce-44bf-b238-71a34c40662a tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] Lock "37a9e1c1-9f6c-4047-9ff5-e141d0ca383e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 848.996976] env[61957]: INFO nova.compute.manager [None req-0d2ed604-6dce-44bf-b238-71a34c40662a tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] [instance: 37a9e1c1-9f6c-4047-9ff5-e141d0ca383e] Terminating instance [ 848.999110] env[61957]: DEBUG nova.compute.manager [None req-0d2ed604-6dce-44bf-b238-71a34c40662a tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] [instance: 37a9e1c1-9f6c-4047-9ff5-e141d0ca383e] Start destroying the instance on the hypervisor. {{(pid=61957) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 848.999368] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-0d2ed604-6dce-44bf-b238-71a34c40662a tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] [instance: 37a9e1c1-9f6c-4047-9ff5-e141d0ca383e] Destroying instance {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 849.000386] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4be0ba8-3254-4c41-8269-153ad0a73fe4 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.011476] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-0d2ed604-6dce-44bf-b238-71a34c40662a tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] [instance: 37a9e1c1-9f6c-4047-9ff5-e141d0ca383e] Powering off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 849.011758] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-84d1c6db-0458-4571-bc8b-94e876caae93 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.020119] env[61957]: DEBUG oslo_vmware.api [None req-0d2ed604-6dce-44bf-b238-71a34c40662a tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] Waiting for the task: (returnval){ [ 849.020119] env[61957]: value = "task-1277503" [ 849.020119] env[61957]: _type = "Task" [ 849.020119] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.028185] env[61957]: DEBUG oslo_vmware.api [None req-0d2ed604-6dce-44bf-b238-71a34c40662a tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] Task: {'id': task-1277503, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.079877] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-290c47b8-61df-43c2-9a3f-de1cbb210c00 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.088134] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e1fdabc-450d-4765-b990-d874f3319f7c {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.121506] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48cd8d19-4093-42aa-8ca8-a9ecd6f28a97 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.130168] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca490b7b-8807-43e7-8bae-1238c8305363 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.144574] env[61957]: DEBUG nova.compute.provider_tree [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 849.324778] env[61957]: DEBUG nova.virt.hardware [None req-d096c609-3f99-4aff-8b69-219b9552f608 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T17:22:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T17:21:50Z,direct_url=,disk_format='vmdk',id=11c76a2c-f705-470a-ba9d-4657858bab38,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='81d3096d0f094373913fc3dc8f5c3c6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T17:21:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 849.325116] env[61957]: DEBUG nova.virt.hardware [None req-d096c609-3f99-4aff-8b69-219b9552f608 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Flavor limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 849.325294] env[61957]: DEBUG nova.virt.hardware [None req-d096c609-3f99-4aff-8b69-219b9552f608 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Image limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 849.325477] env[61957]: DEBUG nova.virt.hardware [None req-d096c609-3f99-4aff-8b69-219b9552f608 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Flavor pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 849.325684] env[61957]: DEBUG nova.virt.hardware [None req-d096c609-3f99-4aff-8b69-219b9552f608 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Image pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 849.325768] env[61957]: DEBUG nova.virt.hardware [None req-d096c609-3f99-4aff-8b69-219b9552f608 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 849.325973] env[61957]: DEBUG nova.virt.hardware [None req-d096c609-3f99-4aff-8b69-219b9552f608 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 849.326147] env[61957]: DEBUG nova.virt.hardware [None req-d096c609-3f99-4aff-8b69-219b9552f608 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 849.326313] env[61957]: DEBUG nova.virt.hardware [None req-d096c609-3f99-4aff-8b69-219b9552f608 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Got 1 possible topologies {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 849.326472] env[61957]: DEBUG nova.virt.hardware [None req-d096c609-3f99-4aff-8b69-219b9552f608 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 849.327350] env[61957]: DEBUG nova.virt.hardware [None req-d096c609-3f99-4aff-8b69-219b9552f608 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 849.327660] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32f31c8a-c025-4dc9-8cd4-8002c6fe76b3 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.337068] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-668bcad7-9da3-4dfa-bb06-34e864b96b99 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.351458] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-d096c609-3f99-4aff-8b69-219b9552f608 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: 37ce46a1-8f9a-4d15-bd81-e40845a0e48a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:26:32:6a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '20e3f794-c7a3-4696-9488-ecf34c570ef9', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7234f796-0abc-4e81-ba04-29607283618c', 'vif_model': 'vmxnet3'}] {{(pid=61957) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 849.359695] env[61957]: DEBUG oslo.service.loopingcall [None req-d096c609-3f99-4aff-8b69-219b9552f608 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 849.360022] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 37ce46a1-8f9a-4d15-bd81-e40845a0e48a] Creating VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 849.360354] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8b80861e-0139-42fe-95bc-12e87ae1c6b6 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.379971] env[61957]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 849.379971] env[61957]: value = "task-1277504" [ 849.379971] env[61957]: _type = "Task" [ 849.379971] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.388226] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277504, 'name': CreateVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.531432] env[61957]: DEBUG oslo_vmware.api [None req-0d2ed604-6dce-44bf-b238-71a34c40662a tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] Task: {'id': task-1277503, 'name': PowerOffVM_Task, 'duration_secs': 0.234548} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.531729] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-0d2ed604-6dce-44bf-b238-71a34c40662a tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] [instance: 37a9e1c1-9f6c-4047-9ff5-e141d0ca383e] Powered off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 849.531902] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-0d2ed604-6dce-44bf-b238-71a34c40662a tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] [instance: 37a9e1c1-9f6c-4047-9ff5-e141d0ca383e] Unregistering the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 849.532185] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-682cc29b-c566-4f81-9aa6-bc96bd00ea13 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.647658] env[61957]: DEBUG nova.scheduler.client.report [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 849.683960] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-0d2ed604-6dce-44bf-b238-71a34c40662a tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] [instance: 37a9e1c1-9f6c-4047-9ff5-e141d0ca383e] Unregistered the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 849.684443] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-0d2ed604-6dce-44bf-b238-71a34c40662a tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] [instance: 37a9e1c1-9f6c-4047-9ff5-e141d0ca383e] Deleting contents of the VM from datastore datastore2 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 849.684694] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-0d2ed604-6dce-44bf-b238-71a34c40662a tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] Deleting the datastore file [datastore2] 37a9e1c1-9f6c-4047-9ff5-e141d0ca383e {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 849.685399] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8ed113c9-57c0-4bde-84a9-523e1a150f3f {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.692447] env[61957]: DEBUG oslo_vmware.api [None req-0d2ed604-6dce-44bf-b238-71a34c40662a tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] Waiting for the task: (returnval){ [ 849.692447] env[61957]: value = "task-1277506" [ 849.692447] env[61957]: _type = "Task" [ 849.692447] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.700854] env[61957]: DEBUG oslo_vmware.api [None req-0d2ed604-6dce-44bf-b238-71a34c40662a tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] Task: {'id': task-1277506, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.747300] env[61957]: DEBUG nova.compute.manager [None req-e4109c08-00a8-4666-8e84-183191cbc2fb tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: 0adae8e1-8c2f-4110-805b-1f286debc833] Start spawning the instance on the hypervisor. {{(pid=61957) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 849.772704] env[61957]: DEBUG nova.virt.hardware [None req-e4109c08-00a8-4666-8e84-183191cbc2fb tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T17:22:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T17:21:50Z,direct_url=,disk_format='vmdk',id=11c76a2c-f705-470a-ba9d-4657858bab38,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='81d3096d0f094373913fc3dc8f5c3c6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T17:21:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 849.773026] env[61957]: DEBUG nova.virt.hardware [None req-e4109c08-00a8-4666-8e84-183191cbc2fb tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Flavor limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 849.773278] env[61957]: DEBUG nova.virt.hardware [None req-e4109c08-00a8-4666-8e84-183191cbc2fb tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Image limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 849.773516] env[61957]: DEBUG nova.virt.hardware [None req-e4109c08-00a8-4666-8e84-183191cbc2fb tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Flavor pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 849.773709] env[61957]: DEBUG nova.virt.hardware [None req-e4109c08-00a8-4666-8e84-183191cbc2fb tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Image pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 849.773897] env[61957]: DEBUG nova.virt.hardware [None req-e4109c08-00a8-4666-8e84-183191cbc2fb tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 849.774160] env[61957]: DEBUG nova.virt.hardware [None req-e4109c08-00a8-4666-8e84-183191cbc2fb tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 849.774359] env[61957]: DEBUG nova.virt.hardware [None req-e4109c08-00a8-4666-8e84-183191cbc2fb tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 849.774542] env[61957]: DEBUG nova.virt.hardware [None req-e4109c08-00a8-4666-8e84-183191cbc2fb tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Got 1 possible topologies {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 849.774758] env[61957]: DEBUG nova.virt.hardware [None req-e4109c08-00a8-4666-8e84-183191cbc2fb tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 849.774956] env[61957]: DEBUG nova.virt.hardware [None req-e4109c08-00a8-4666-8e84-183191cbc2fb tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 849.775886] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01f40631-e44f-4590-a8d7-38bf80cba329 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.785741] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-520a4a64-f37a-4e0a-ba4a-74f74a62ee7b {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.889197] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277504, 'name': CreateVM_Task, 'duration_secs': 0.426147} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.889387] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 37ce46a1-8f9a-4d15-bd81-e40845a0e48a] Created VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 849.890127] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d096c609-3f99-4aff-8b69-219b9552f608 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 849.890304] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d096c609-3f99-4aff-8b69-219b9552f608 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 849.890630] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d096c609-3f99-4aff-8b69-219b9552f608 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 849.890910] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-97deed76-40ac-4919-b9e1-235eb704b762 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.895649] env[61957]: DEBUG oslo_vmware.api [None req-d096c609-3f99-4aff-8b69-219b9552f608 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Waiting for the task: (returnval){ [ 849.895649] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52a50a38-a4e7-0a4d-b99d-8ffe9e8a6644" [ 849.895649] env[61957]: _type = "Task" [ 849.895649] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.903558] env[61957]: DEBUG oslo_vmware.api [None req-d096c609-3f99-4aff-8b69-219b9552f608 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52a50a38-a4e7-0a4d-b99d-8ffe9e8a6644, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.004106] env[61957]: DEBUG nova.compute.manager [req-9e9b7b65-7b0b-4319-bb8a-f76bae976fd0 req-d966ceb8-fedb-4373-8f94-cee49f9b8ad0 service nova] [instance: 0adae8e1-8c2f-4110-805b-1f286debc833] Received event network-vif-plugged-06b59d3b-7dab-4254-9ccd-cbfcce2fce87 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 850.005994] env[61957]: DEBUG oslo_concurrency.lockutils [req-9e9b7b65-7b0b-4319-bb8a-f76bae976fd0 req-d966ceb8-fedb-4373-8f94-cee49f9b8ad0 service nova] Acquiring lock "0adae8e1-8c2f-4110-805b-1f286debc833-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 850.005994] env[61957]: DEBUG oslo_concurrency.lockutils [req-9e9b7b65-7b0b-4319-bb8a-f76bae976fd0 req-d966ceb8-fedb-4373-8f94-cee49f9b8ad0 service nova] Lock "0adae8e1-8c2f-4110-805b-1f286debc833-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 850.005994] env[61957]: DEBUG oslo_concurrency.lockutils [req-9e9b7b65-7b0b-4319-bb8a-f76bae976fd0 req-d966ceb8-fedb-4373-8f94-cee49f9b8ad0 service nova] Lock "0adae8e1-8c2f-4110-805b-1f286debc833-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 850.005994] env[61957]: DEBUG nova.compute.manager [req-9e9b7b65-7b0b-4319-bb8a-f76bae976fd0 req-d966ceb8-fedb-4373-8f94-cee49f9b8ad0 service nova] [instance: 0adae8e1-8c2f-4110-805b-1f286debc833] No waiting events found dispatching network-vif-plugged-06b59d3b-7dab-4254-9ccd-cbfcce2fce87 {{(pid=61957) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 850.005994] env[61957]: WARNING nova.compute.manager [req-9e9b7b65-7b0b-4319-bb8a-f76bae976fd0 req-d966ceb8-fedb-4373-8f94-cee49f9b8ad0 service nova] [instance: 0adae8e1-8c2f-4110-805b-1f286debc833] Received unexpected event network-vif-plugged-06b59d3b-7dab-4254-9ccd-cbfcce2fce87 for instance with vm_state building and task_state spawning. [ 850.096819] env[61957]: DEBUG nova.network.neutron [None req-e4109c08-00a8-4666-8e84-183191cbc2fb tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: 0adae8e1-8c2f-4110-805b-1f286debc833] Successfully updated port: 06b59d3b-7dab-4254-9ccd-cbfcce2fce87 {{(pid=61957) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 850.153197] env[61957]: DEBUG oslo_concurrency.lockutils [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.432s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 850.153781] env[61957]: DEBUG nova.compute.manager [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: 6438fe40-046c-45d5-9986-8f182ecde49f] Start building networks asynchronously for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 850.156952] env[61957]: DEBUG oslo_concurrency.lockutils [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.978s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 850.158420] env[61957]: INFO nova.compute.claims [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: f66db265-887e-4d61-b848-c609e5c884cb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 850.203642] env[61957]: DEBUG oslo_vmware.api [None req-0d2ed604-6dce-44bf-b238-71a34c40662a tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] Task: {'id': task-1277506, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.247902} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.203977] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-0d2ed604-6dce-44bf-b238-71a34c40662a tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] Deleted the datastore file {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 850.204207] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-0d2ed604-6dce-44bf-b238-71a34c40662a tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] [instance: 37a9e1c1-9f6c-4047-9ff5-e141d0ca383e] Deleted contents of the VM from datastore datastore2 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 850.204400] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-0d2ed604-6dce-44bf-b238-71a34c40662a tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] [instance: 37a9e1c1-9f6c-4047-9ff5-e141d0ca383e] Instance destroyed {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 850.204573] env[61957]: INFO nova.compute.manager [None req-0d2ed604-6dce-44bf-b238-71a34c40662a tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] [instance: 37a9e1c1-9f6c-4047-9ff5-e141d0ca383e] Took 1.21 seconds to destroy the instance on the hypervisor. [ 850.204816] env[61957]: DEBUG oslo.service.loopingcall [None req-0d2ed604-6dce-44bf-b238-71a34c40662a tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 850.205023] env[61957]: DEBUG nova.compute.manager [-] [instance: 37a9e1c1-9f6c-4047-9ff5-e141d0ca383e] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 850.205118] env[61957]: DEBUG nova.network.neutron [-] [instance: 37a9e1c1-9f6c-4047-9ff5-e141d0ca383e] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 850.406887] env[61957]: DEBUG oslo_vmware.api [None req-d096c609-3f99-4aff-8b69-219b9552f608 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52a50a38-a4e7-0a4d-b99d-8ffe9e8a6644, 'name': SearchDatastore_Task, 'duration_secs': 0.012947} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.407482] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d096c609-3f99-4aff-8b69-219b9552f608 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 850.407482] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-d096c609-3f99-4aff-8b69-219b9552f608 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: 37ce46a1-8f9a-4d15-bd81-e40845a0e48a] Processing image 11c76a2c-f705-470a-ba9d-4657858bab38 {{(pid=61957) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 850.407712] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d096c609-3f99-4aff-8b69-219b9552f608 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 850.407904] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d096c609-3f99-4aff-8b69-219b9552f608 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 850.408127] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-d096c609-3f99-4aff-8b69-219b9552f608 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 850.408463] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b354cecf-277e-458d-b7eb-6d898a1c60fa {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.417305] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-d096c609-3f99-4aff-8b69-219b9552f608 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 850.417411] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-d096c609-3f99-4aff-8b69-219b9552f608 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61957) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 850.418199] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dd18fa19-4484-4e02-9e90-651770a563bd {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.423868] env[61957]: DEBUG oslo_vmware.api [None req-d096c609-3f99-4aff-8b69-219b9552f608 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Waiting for the task: (returnval){ [ 850.423868] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52896c07-e7a8-be5f-6e0d-f442613a9518" [ 850.423868] env[61957]: _type = "Task" [ 850.423868] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.432337] env[61957]: DEBUG oslo_vmware.api [None req-d096c609-3f99-4aff-8b69-219b9552f608 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52896c07-e7a8-be5f-6e0d-f442613a9518, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.599712] env[61957]: DEBUG oslo_concurrency.lockutils [None req-e4109c08-00a8-4666-8e84-183191cbc2fb tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Acquiring lock "refresh_cache-0adae8e1-8c2f-4110-805b-1f286debc833" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 850.600164] env[61957]: DEBUG oslo_concurrency.lockutils [None req-e4109c08-00a8-4666-8e84-183191cbc2fb tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Acquired lock "refresh_cache-0adae8e1-8c2f-4110-805b-1f286debc833" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 850.600164] env[61957]: DEBUG nova.network.neutron [None req-e4109c08-00a8-4666-8e84-183191cbc2fb tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: 0adae8e1-8c2f-4110-805b-1f286debc833] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 850.658868] env[61957]: DEBUG nova.compute.utils [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Using /dev/sd instead of None {{(pid=61957) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 850.663178] env[61957]: DEBUG nova.compute.manager [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: 6438fe40-046c-45d5-9986-8f182ecde49f] Allocating IP information in the background. {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 850.663178] env[61957]: DEBUG nova.network.neutron [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: 6438fe40-046c-45d5-9986-8f182ecde49f] allocate_for_instance() {{(pid=61957) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 850.711695] env[61957]: DEBUG nova.policy [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9b453c031e7a4294b14dd48e62e1fd30', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '55dd45b4291b456f83a446dfa87f550a', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61957) authorize /opt/stack/nova/nova/policy.py:203}} [ 850.935398] env[61957]: DEBUG oslo_vmware.api [None req-d096c609-3f99-4aff-8b69-219b9552f608 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52896c07-e7a8-be5f-6e0d-f442613a9518, 'name': SearchDatastore_Task, 'duration_secs': 0.012715} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.936318] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-37ca58ec-df39-42e8-9fe9-9440160e86ab {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.943266] env[61957]: DEBUG oslo_vmware.api [None req-d096c609-3f99-4aff-8b69-219b9552f608 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Waiting for the task: (returnval){ [ 850.943266] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]5200c187-89f7-8793-8402-8642bf5097f7" [ 850.943266] env[61957]: _type = "Task" [ 850.943266] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.953826] env[61957]: DEBUG oslo_vmware.api [None req-d096c609-3f99-4aff-8b69-219b9552f608 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]5200c187-89f7-8793-8402-8642bf5097f7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.990578] env[61957]: DEBUG nova.network.neutron [-] [instance: 37a9e1c1-9f6c-4047-9ff5-e141d0ca383e] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 850.998735] env[61957]: DEBUG nova.network.neutron [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: 6438fe40-046c-45d5-9986-8f182ecde49f] Successfully created port: e247dac1-1a7a-4b92-a2c0-66cfbe089af3 {{(pid=61957) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 851.145144] env[61957]: DEBUG nova.network.neutron [None req-e4109c08-00a8-4666-8e84-183191cbc2fb tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: 0adae8e1-8c2f-4110-805b-1f286debc833] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 851.164060] env[61957]: DEBUG nova.compute.manager [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: 6438fe40-046c-45d5-9986-8f182ecde49f] Start building block device mappings for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 851.404260] env[61957]: DEBUG nova.network.neutron [None req-e4109c08-00a8-4666-8e84-183191cbc2fb tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: 0adae8e1-8c2f-4110-805b-1f286debc833] Updating instance_info_cache with network_info: [{"id": "06b59d3b-7dab-4254-9ccd-cbfcce2fce87", "address": "fa:16:3e:88:c8:e7", "network": {"id": "dc7d4c3b-1f91-4937-a2b0-1084e5c478f0", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1637370260-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "975189ebb3cc4cdb9391880f0c9ba6ba", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b71230ae-e879-4384-88ce-fe64c86fce22", "external-id": "nsx-vlan-transportzone-473", "segmentation_id": 473, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap06b59d3b-7d", "ovs_interfaceid": "06b59d3b-7dab-4254-9ccd-cbfcce2fce87", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 851.456647] env[61957]: DEBUG oslo_vmware.api [None req-d096c609-3f99-4aff-8b69-219b9552f608 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]5200c187-89f7-8793-8402-8642bf5097f7, 'name': SearchDatastore_Task, 'duration_secs': 0.012895} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.459247] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d096c609-3f99-4aff-8b69-219b9552f608 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 851.459515] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-d096c609-3f99-4aff-8b69-219b9552f608 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore2] 37ce46a1-8f9a-4d15-bd81-e40845a0e48a/37ce46a1-8f9a-4d15-bd81-e40845a0e48a.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 851.460024] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-cb64d4d5-fb67-4fd8-ad5f-9ebfc39bbdb0 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.468155] env[61957]: DEBUG oslo_vmware.api [None req-d096c609-3f99-4aff-8b69-219b9552f608 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Waiting for the task: (returnval){ [ 851.468155] env[61957]: value = "task-1277507" [ 851.468155] env[61957]: _type = "Task" [ 851.468155] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.479085] env[61957]: DEBUG oslo_vmware.api [None req-d096c609-3f99-4aff-8b69-219b9552f608 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': task-1277507, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.493244] env[61957]: INFO nova.compute.manager [-] [instance: 37a9e1c1-9f6c-4047-9ff5-e141d0ca383e] Took 1.29 seconds to deallocate network for instance. [ 851.547882] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b90aff36-f14b-4284-8789-2e988e9115b8 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.556745] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7b90bcd-0b4a-48ea-91e8-cf75937e662c {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.589392] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb418ecf-65ce-4327-946a-4d700281442e {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.597299] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7461f1fb-82b9-4472-9eef-0210b285341d {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.615274] env[61957]: DEBUG nova.compute.provider_tree [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 851.907688] env[61957]: DEBUG oslo_concurrency.lockutils [None req-e4109c08-00a8-4666-8e84-183191cbc2fb tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Releasing lock "refresh_cache-0adae8e1-8c2f-4110-805b-1f286debc833" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 851.908099] env[61957]: DEBUG nova.compute.manager [None req-e4109c08-00a8-4666-8e84-183191cbc2fb tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: 0adae8e1-8c2f-4110-805b-1f286debc833] Instance network_info: |[{"id": "06b59d3b-7dab-4254-9ccd-cbfcce2fce87", "address": "fa:16:3e:88:c8:e7", "network": {"id": "dc7d4c3b-1f91-4937-a2b0-1084e5c478f0", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1637370260-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "975189ebb3cc4cdb9391880f0c9ba6ba", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b71230ae-e879-4384-88ce-fe64c86fce22", "external-id": "nsx-vlan-transportzone-473", "segmentation_id": 473, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap06b59d3b-7d", "ovs_interfaceid": "06b59d3b-7dab-4254-9ccd-cbfcce2fce87", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 851.910033] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-e4109c08-00a8-4666-8e84-183191cbc2fb tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: 0adae8e1-8c2f-4110-805b-1f286debc833] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:88:c8:e7', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'b71230ae-e879-4384-88ce-fe64c86fce22', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '06b59d3b-7dab-4254-9ccd-cbfcce2fce87', 'vif_model': 'vmxnet3'}] {{(pid=61957) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 851.917175] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-e4109c08-00a8-4666-8e84-183191cbc2fb tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Creating folder: Project (975189ebb3cc4cdb9391880f0c9ba6ba). Parent ref: group-v274445. {{(pid=61957) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 851.917566] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9243542e-ac1f-491f-8c2e-e82bcfee65dc {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.929824] env[61957]: INFO nova.virt.vmwareapi.vm_util [None req-e4109c08-00a8-4666-8e84-183191cbc2fb tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Created folder: Project (975189ebb3cc4cdb9391880f0c9ba6ba) in parent group-v274445. [ 851.930209] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-e4109c08-00a8-4666-8e84-183191cbc2fb tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Creating folder: Instances. Parent ref: group-v274505. {{(pid=61957) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 851.930433] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9b4ded8d-3bf1-4159-80ea-b0593149ddaf {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.942351] env[61957]: INFO nova.virt.vmwareapi.vm_util [None req-e4109c08-00a8-4666-8e84-183191cbc2fb tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Created folder: Instances in parent group-v274505. [ 851.942831] env[61957]: DEBUG oslo.service.loopingcall [None req-e4109c08-00a8-4666-8e84-183191cbc2fb tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 851.943086] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0adae8e1-8c2f-4110-805b-1f286debc833] Creating VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 851.943324] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a244fd9a-f3bb-40d9-a45b-44876039abb4 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.963684] env[61957]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 851.963684] env[61957]: value = "task-1277510" [ 851.963684] env[61957]: _type = "Task" [ 851.963684] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.977189] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277510, 'name': CreateVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.983698] env[61957]: DEBUG oslo_vmware.api [None req-d096c609-3f99-4aff-8b69-219b9552f608 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': task-1277507, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.002957] env[61957]: DEBUG oslo_concurrency.lockutils [None req-0d2ed604-6dce-44bf-b238-71a34c40662a tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 852.113104] env[61957]: DEBUG nova.compute.manager [req-a95f019e-a278-4e41-b6c4-81d810ebd89a req-167a57ef-4b0f-4460-b25f-3990a100030d service nova] [instance: 0adae8e1-8c2f-4110-805b-1f286debc833] Received event network-changed-06b59d3b-7dab-4254-9ccd-cbfcce2fce87 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 852.113338] env[61957]: DEBUG nova.compute.manager [req-a95f019e-a278-4e41-b6c4-81d810ebd89a req-167a57ef-4b0f-4460-b25f-3990a100030d service nova] [instance: 0adae8e1-8c2f-4110-805b-1f286debc833] Refreshing instance network info cache due to event network-changed-06b59d3b-7dab-4254-9ccd-cbfcce2fce87. {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 852.113567] env[61957]: DEBUG oslo_concurrency.lockutils [req-a95f019e-a278-4e41-b6c4-81d810ebd89a req-167a57ef-4b0f-4460-b25f-3990a100030d service nova] Acquiring lock "refresh_cache-0adae8e1-8c2f-4110-805b-1f286debc833" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 852.114256] env[61957]: DEBUG oslo_concurrency.lockutils [req-a95f019e-a278-4e41-b6c4-81d810ebd89a req-167a57ef-4b0f-4460-b25f-3990a100030d service nova] Acquired lock "refresh_cache-0adae8e1-8c2f-4110-805b-1f286debc833" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 852.114682] env[61957]: DEBUG nova.network.neutron [req-a95f019e-a278-4e41-b6c4-81d810ebd89a req-167a57ef-4b0f-4460-b25f-3990a100030d service nova] [instance: 0adae8e1-8c2f-4110-805b-1f286debc833] Refreshing network info cache for port 06b59d3b-7dab-4254-9ccd-cbfcce2fce87 {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 852.119612] env[61957]: DEBUG nova.scheduler.client.report [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 852.183560] env[61957]: DEBUG nova.compute.manager [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: 6438fe40-046c-45d5-9986-8f182ecde49f] Start spawning the instance on the hypervisor. {{(pid=61957) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 852.205562] env[61957]: DEBUG nova.virt.hardware [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T17:22:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T17:21:50Z,direct_url=,disk_format='vmdk',id=11c76a2c-f705-470a-ba9d-4657858bab38,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='81d3096d0f094373913fc3dc8f5c3c6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T17:21:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 852.205826] env[61957]: DEBUG nova.virt.hardware [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Flavor limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 852.205987] env[61957]: DEBUG nova.virt.hardware [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Image limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 852.206206] env[61957]: DEBUG nova.virt.hardware [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Flavor pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 852.206349] env[61957]: DEBUG nova.virt.hardware [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Image pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 852.206499] env[61957]: DEBUG nova.virt.hardware [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 852.206715] env[61957]: DEBUG nova.virt.hardware [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 852.206878] env[61957]: DEBUG nova.virt.hardware [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 852.207060] env[61957]: DEBUG nova.virt.hardware [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Got 1 possible topologies {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 852.207232] env[61957]: DEBUG nova.virt.hardware [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 852.207410] env[61957]: DEBUG nova.virt.hardware [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 852.208658] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20b3486f-4916-4a42-9dd3-2c95a70023a9 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.216711] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f8b8f94-75cd-49b0-aa82-989cb2a245ef {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.476240] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277510, 'name': CreateVM_Task, 'duration_secs': 0.403278} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.476832] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0adae8e1-8c2f-4110-805b-1f286debc833] Created VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 852.477567] env[61957]: DEBUG oslo_concurrency.lockutils [None req-e4109c08-00a8-4666-8e84-183191cbc2fb tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 852.477818] env[61957]: DEBUG oslo_concurrency.lockutils [None req-e4109c08-00a8-4666-8e84-183191cbc2fb tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 852.478291] env[61957]: DEBUG oslo_concurrency.lockutils [None req-e4109c08-00a8-4666-8e84-183191cbc2fb tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 852.478422] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-82444fc9-bbb6-4eff-ac4c-5bfc7f74fab2 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.483327] env[61957]: DEBUG oslo_vmware.api [None req-d096c609-3f99-4aff-8b69-219b9552f608 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': task-1277507, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.542403} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.483927] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-d096c609-3f99-4aff-8b69-219b9552f608 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore2] 37ce46a1-8f9a-4d15-bd81-e40845a0e48a/37ce46a1-8f9a-4d15-bd81-e40845a0e48a.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 852.484204] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-d096c609-3f99-4aff-8b69-219b9552f608 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: 37ce46a1-8f9a-4d15-bd81-e40845a0e48a] Extending root virtual disk to 1048576 {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 852.484461] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-958b5d6d-b460-494c-bd85-04e687d0b9f1 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.487285] env[61957]: DEBUG oslo_vmware.api [None req-e4109c08-00a8-4666-8e84-183191cbc2fb tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Waiting for the task: (returnval){ [ 852.487285] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]527d81f2-d883-b1ca-e89d-cf79db4498c6" [ 852.487285] env[61957]: _type = "Task" [ 852.487285] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.492680] env[61957]: DEBUG oslo_vmware.api [None req-d096c609-3f99-4aff-8b69-219b9552f608 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Waiting for the task: (returnval){ [ 852.492680] env[61957]: value = "task-1277511" [ 852.492680] env[61957]: _type = "Task" [ 852.492680] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.495787] env[61957]: DEBUG oslo_vmware.api [None req-e4109c08-00a8-4666-8e84-183191cbc2fb tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]527d81f2-d883-b1ca-e89d-cf79db4498c6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.503767] env[61957]: DEBUG oslo_vmware.api [None req-d096c609-3f99-4aff-8b69-219b9552f608 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': task-1277511, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.628308] env[61957]: DEBUG oslo_concurrency.lockutils [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.471s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 852.628592] env[61957]: DEBUG nova.compute.manager [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: f66db265-887e-4d61-b848-c609e5c884cb] Start building networks asynchronously for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 852.631641] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ac5c4b48-cfd7-420e-9aa1-225385350d11 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.777s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 852.636668] env[61957]: INFO nova.compute.claims [None req-ac5c4b48-cfd7-420e-9aa1-225385350d11 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] [instance: c23141ee-0cbb-4d1b-8390-c3073fe354f1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 852.725029] env[61957]: DEBUG nova.network.neutron [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: 6438fe40-046c-45d5-9986-8f182ecde49f] Successfully updated port: e247dac1-1a7a-4b92-a2c0-66cfbe089af3 {{(pid=61957) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 852.875421] env[61957]: DEBUG nova.network.neutron [req-a95f019e-a278-4e41-b6c4-81d810ebd89a req-167a57ef-4b0f-4460-b25f-3990a100030d service nova] [instance: 0adae8e1-8c2f-4110-805b-1f286debc833] Updated VIF entry in instance network info cache for port 06b59d3b-7dab-4254-9ccd-cbfcce2fce87. {{(pid=61957) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 852.875865] env[61957]: DEBUG nova.network.neutron [req-a95f019e-a278-4e41-b6c4-81d810ebd89a req-167a57ef-4b0f-4460-b25f-3990a100030d service nova] [instance: 0adae8e1-8c2f-4110-805b-1f286debc833] Updating instance_info_cache with network_info: [{"id": "06b59d3b-7dab-4254-9ccd-cbfcce2fce87", "address": "fa:16:3e:88:c8:e7", "network": {"id": "dc7d4c3b-1f91-4937-a2b0-1084e5c478f0", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1637370260-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "975189ebb3cc4cdb9391880f0c9ba6ba", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b71230ae-e879-4384-88ce-fe64c86fce22", "external-id": "nsx-vlan-transportzone-473", "segmentation_id": 473, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap06b59d3b-7d", "ovs_interfaceid": "06b59d3b-7dab-4254-9ccd-cbfcce2fce87", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 852.898263] env[61957]: DEBUG oslo_vmware.rw_handles [None req-b1306d4e-743c-4d53-8fec-4c8136571e40 tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52382179-d7d4-71d0-9e9f-1f0dce243f6a/disk-0.vmdk. {{(pid=61957) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 852.899171] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e2a293e-2d79-47b4-9ec4-be1300f75549 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.905440] env[61957]: DEBUG oslo_vmware.rw_handles [None req-b1306d4e-743c-4d53-8fec-4c8136571e40 tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52382179-d7d4-71d0-9e9f-1f0dce243f6a/disk-0.vmdk is in state: ready. {{(pid=61957) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 852.905615] env[61957]: ERROR oslo_vmware.rw_handles [None req-b1306d4e-743c-4d53-8fec-4c8136571e40 tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52382179-d7d4-71d0-9e9f-1f0dce243f6a/disk-0.vmdk due to incomplete transfer. [ 852.905830] env[61957]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-5fc1ff96-a36b-4310-855b-914f2165dfcf {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.913386] env[61957]: DEBUG oslo_vmware.rw_handles [None req-b1306d4e-743c-4d53-8fec-4c8136571e40 tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52382179-d7d4-71d0-9e9f-1f0dce243f6a/disk-0.vmdk. {{(pid=61957) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 852.913585] env[61957]: DEBUG nova.virt.vmwareapi.images [None req-b1306d4e-743c-4d53-8fec-4c8136571e40 tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] [instance: e0249936-d616-4ffb-8f77-d8107633c42a] Uploaded image 3e563c87-a879-4f60-8e71-1dde6db914a4 to the Glance image server {{(pid=61957) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 852.915541] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-b1306d4e-743c-4d53-8fec-4c8136571e40 tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] [instance: e0249936-d616-4ffb-8f77-d8107633c42a] Destroying the VM {{(pid=61957) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 852.915781] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-81f1eeef-78c1-49ad-beee-a692cc0aaa0e {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.921288] env[61957]: DEBUG oslo_vmware.api [None req-b1306d4e-743c-4d53-8fec-4c8136571e40 tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Waiting for the task: (returnval){ [ 852.921288] env[61957]: value = "task-1277512" [ 852.921288] env[61957]: _type = "Task" [ 852.921288] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.929200] env[61957]: DEBUG oslo_vmware.api [None req-b1306d4e-743c-4d53-8fec-4c8136571e40 tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Task: {'id': task-1277512, 'name': Destroy_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.997528] env[61957]: DEBUG oslo_vmware.api [None req-e4109c08-00a8-4666-8e84-183191cbc2fb tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]527d81f2-d883-b1ca-e89d-cf79db4498c6, 'name': SearchDatastore_Task, 'duration_secs': 0.019492} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.000707] env[61957]: DEBUG oslo_concurrency.lockutils [None req-e4109c08-00a8-4666-8e84-183191cbc2fb tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 853.000986] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-e4109c08-00a8-4666-8e84-183191cbc2fb tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: 0adae8e1-8c2f-4110-805b-1f286debc833] Processing image 11c76a2c-f705-470a-ba9d-4657858bab38 {{(pid=61957) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 853.001229] env[61957]: DEBUG oslo_concurrency.lockutils [None req-e4109c08-00a8-4666-8e84-183191cbc2fb tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 853.001381] env[61957]: DEBUG oslo_concurrency.lockutils [None req-e4109c08-00a8-4666-8e84-183191cbc2fb tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 853.001568] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-e4109c08-00a8-4666-8e84-183191cbc2fb tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 853.001831] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f876ed67-5404-49ae-aeb9-6b1fca4a51f9 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.008392] env[61957]: DEBUG oslo_vmware.api [None req-d096c609-3f99-4aff-8b69-219b9552f608 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': task-1277511, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065595} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.008638] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-d096c609-3f99-4aff-8b69-219b9552f608 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: 37ce46a1-8f9a-4d15-bd81-e40845a0e48a] Extended root virtual disk {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 853.009396] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4685fcf-7aef-482a-8127-baad61bd1284 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.012431] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-e4109c08-00a8-4666-8e84-183191cbc2fb tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 853.012602] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-e4109c08-00a8-4666-8e84-183191cbc2fb tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61957) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 853.013593] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-daacaa10-1f7b-4c03-b16c-3f1edaab2525 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.033049] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-d096c609-3f99-4aff-8b69-219b9552f608 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: 37ce46a1-8f9a-4d15-bd81-e40845a0e48a] Reconfiguring VM instance instance-0000003a to attach disk [datastore2] 37ce46a1-8f9a-4d15-bd81-e40845a0e48a/37ce46a1-8f9a-4d15-bd81-e40845a0e48a.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 853.033648] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b8a1a268-3c2f-414b-8e68-fc03bb5e9a4c {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.048765] env[61957]: DEBUG oslo_vmware.api [None req-e4109c08-00a8-4666-8e84-183191cbc2fb tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Waiting for the task: (returnval){ [ 853.048765] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52f4c9d4-b1f1-8ff8-8052-6539843c6aa1" [ 853.048765] env[61957]: _type = "Task" [ 853.048765] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.053280] env[61957]: DEBUG oslo_vmware.api [None req-d096c609-3f99-4aff-8b69-219b9552f608 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Waiting for the task: (returnval){ [ 853.053280] env[61957]: value = "task-1277513" [ 853.053280] env[61957]: _type = "Task" [ 853.053280] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.059547] env[61957]: DEBUG oslo_vmware.api [None req-e4109c08-00a8-4666-8e84-183191cbc2fb tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52f4c9d4-b1f1-8ff8-8052-6539843c6aa1, 'name': SearchDatastore_Task, 'duration_secs': 0.009536} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.060600] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-422d9ed4-b15b-47d3-834d-a4849828b081 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.065650] env[61957]: DEBUG oslo_vmware.api [None req-d096c609-3f99-4aff-8b69-219b9552f608 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': task-1277513, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.068448] env[61957]: DEBUG oslo_vmware.api [None req-e4109c08-00a8-4666-8e84-183191cbc2fb tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Waiting for the task: (returnval){ [ 853.068448] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]524d4658-3f7d-5d3b-05ef-732411faef51" [ 853.068448] env[61957]: _type = "Task" [ 853.068448] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.075522] env[61957]: DEBUG oslo_vmware.api [None req-e4109c08-00a8-4666-8e84-183191cbc2fb tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]524d4658-3f7d-5d3b-05ef-732411faef51, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.139053] env[61957]: DEBUG nova.compute.utils [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Using /dev/sd instead of None {{(pid=61957) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 853.139747] env[61957]: DEBUG nova.compute.manager [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: f66db265-887e-4d61-b848-c609e5c884cb] Allocating IP information in the background. {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 853.140092] env[61957]: DEBUG nova.network.neutron [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: f66db265-887e-4d61-b848-c609e5c884cb] allocate_for_instance() {{(pid=61957) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 853.181068] env[61957]: DEBUG nova.policy [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9b453c031e7a4294b14dd48e62e1fd30', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '55dd45b4291b456f83a446dfa87f550a', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61957) authorize /opt/stack/nova/nova/policy.py:203}} [ 853.228497] env[61957]: DEBUG oslo_concurrency.lockutils [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Acquiring lock "refresh_cache-6438fe40-046c-45d5-9986-8f182ecde49f" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 853.228598] env[61957]: DEBUG oslo_concurrency.lockutils [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Acquired lock "refresh_cache-6438fe40-046c-45d5-9986-8f182ecde49f" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 853.228750] env[61957]: DEBUG nova.network.neutron [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: 6438fe40-046c-45d5-9986-8f182ecde49f] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 853.380061] env[61957]: DEBUG oslo_concurrency.lockutils [req-a95f019e-a278-4e41-b6c4-81d810ebd89a req-167a57ef-4b0f-4460-b25f-3990a100030d service nova] Releasing lock "refresh_cache-0adae8e1-8c2f-4110-805b-1f286debc833" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 853.380391] env[61957]: DEBUG nova.compute.manager [req-a95f019e-a278-4e41-b6c4-81d810ebd89a req-167a57ef-4b0f-4460-b25f-3990a100030d service nova] [instance: 37a9e1c1-9f6c-4047-9ff5-e141d0ca383e] Received event network-vif-deleted-aee17432-3f8d-4d87-863f-20a333a4c92f {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 853.430866] env[61957]: DEBUG oslo_vmware.api [None req-b1306d4e-743c-4d53-8fec-4c8136571e40 tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Task: {'id': task-1277512, 'name': Destroy_Task} progress is 33%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.437533] env[61957]: DEBUG nova.network.neutron [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: f66db265-887e-4d61-b848-c609e5c884cb] Successfully created port: 0289d1f8-a41a-44e5-9b66-6f92ce51debf {{(pid=61957) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 853.564823] env[61957]: DEBUG oslo_vmware.api [None req-d096c609-3f99-4aff-8b69-219b9552f608 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': task-1277513, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.580043] env[61957]: DEBUG oslo_vmware.api [None req-e4109c08-00a8-4666-8e84-183191cbc2fb tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]524d4658-3f7d-5d3b-05ef-732411faef51, 'name': SearchDatastore_Task, 'duration_secs': 0.017385} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.580304] env[61957]: DEBUG oslo_concurrency.lockutils [None req-e4109c08-00a8-4666-8e84-183191cbc2fb tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 853.580556] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-e4109c08-00a8-4666-8e84-183191cbc2fb tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore2] 0adae8e1-8c2f-4110-805b-1f286debc833/0adae8e1-8c2f-4110-805b-1f286debc833.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 853.580841] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-cf04897d-3e36-4fe7-a184-dc939c0c5eee {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.586856] env[61957]: DEBUG oslo_vmware.api [None req-e4109c08-00a8-4666-8e84-183191cbc2fb tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Waiting for the task: (returnval){ [ 853.586856] env[61957]: value = "task-1277514" [ 853.586856] env[61957]: _type = "Task" [ 853.586856] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.594239] env[61957]: DEBUG oslo_vmware.api [None req-e4109c08-00a8-4666-8e84-183191cbc2fb tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': task-1277514, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.645102] env[61957]: DEBUG nova.compute.manager [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: f66db265-887e-4d61-b848-c609e5c884cb] Start building block device mappings for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 853.765358] env[61957]: DEBUG nova.network.neutron [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: 6438fe40-046c-45d5-9986-8f182ecde49f] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 853.940415] env[61957]: DEBUG oslo_vmware.api [None req-b1306d4e-743c-4d53-8fec-4c8136571e40 tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Task: {'id': task-1277512, 'name': Destroy_Task, 'duration_secs': 0.570763} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.940831] env[61957]: INFO nova.virt.vmwareapi.vm_util [None req-b1306d4e-743c-4d53-8fec-4c8136571e40 tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] [instance: e0249936-d616-4ffb-8f77-d8107633c42a] Destroyed the VM [ 853.941166] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-b1306d4e-743c-4d53-8fec-4c8136571e40 tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] [instance: e0249936-d616-4ffb-8f77-d8107633c42a] Deleting Snapshot of the VM instance {{(pid=61957) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 853.941752] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-22e84c28-1889-46b4-a2a3-eaf87a888bc2 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.948969] env[61957]: DEBUG oslo_vmware.api [None req-b1306d4e-743c-4d53-8fec-4c8136571e40 tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Waiting for the task: (returnval){ [ 853.948969] env[61957]: value = "task-1277515" [ 853.948969] env[61957]: _type = "Task" [ 853.948969] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.952510] env[61957]: DEBUG nova.network.neutron [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: 6438fe40-046c-45d5-9986-8f182ecde49f] Updating instance_info_cache with network_info: [{"id": "e247dac1-1a7a-4b92-a2c0-66cfbe089af3", "address": "fa:16:3e:8b:8f:72", "network": {"id": "2b5a7a9c-7143-4754-8d84-486c2ab13b21", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1508819166-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "55dd45b4291b456f83a446dfa87f550a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "572b7281-aad3-45fa-9cb2-fc1c70569948", "external-id": "nsx-vlan-transportzone-722", "segmentation_id": 722, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape247dac1-1a", "ovs_interfaceid": "e247dac1-1a7a-4b92-a2c0-66cfbe089af3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 853.964057] env[61957]: DEBUG oslo_vmware.api [None req-b1306d4e-743c-4d53-8fec-4c8136571e40 tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Task: {'id': task-1277515, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.018061] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e0b2bee-3ae1-4c99-a3c5-3c48e1970525 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.026206] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b5ce887-0ae2-43f7-ba95-c1e973cd798b {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.064652] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a69dc3a-d033-42e4-945c-52bbf685ebfc {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.073307] env[61957]: DEBUG oslo_vmware.api [None req-d096c609-3f99-4aff-8b69-219b9552f608 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': task-1277513, 'name': ReconfigVM_Task, 'duration_secs': 0.555694} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.075686] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-d096c609-3f99-4aff-8b69-219b9552f608 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: 37ce46a1-8f9a-4d15-bd81-e40845a0e48a] Reconfigured VM instance instance-0000003a to attach disk [datastore2] 37ce46a1-8f9a-4d15-bd81-e40845a0e48a/37ce46a1-8f9a-4d15-bd81-e40845a0e48a.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 854.076528] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-de1c1b69-ce21-4caf-b09f-369326489db4 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.079297] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-253ce139-4caf-4964-ae0d-69ee369f6795 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.100166] env[61957]: DEBUG nova.compute.provider_tree [None req-ac5c4b48-cfd7-420e-9aa1-225385350d11 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 854.101496] env[61957]: DEBUG oslo_vmware.api [None req-d096c609-3f99-4aff-8b69-219b9552f608 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Waiting for the task: (returnval){ [ 854.101496] env[61957]: value = "task-1277516" [ 854.101496] env[61957]: _type = "Task" [ 854.101496] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.112286] env[61957]: DEBUG oslo_vmware.api [None req-e4109c08-00a8-4666-8e84-183191cbc2fb tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': task-1277514, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.120129] env[61957]: DEBUG oslo_vmware.api [None req-d096c609-3f99-4aff-8b69-219b9552f608 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': task-1277516, 'name': Rename_Task} progress is 14%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.314128] env[61957]: DEBUG nova.compute.manager [req-1a9592b0-b842-48b3-8a68-0eb9ac7426c3 req-8f009fa1-8b46-4337-b2fe-8f303864bda8 service nova] [instance: 6438fe40-046c-45d5-9986-8f182ecde49f] Received event network-vif-plugged-e247dac1-1a7a-4b92-a2c0-66cfbe089af3 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 854.315207] env[61957]: DEBUG oslo_concurrency.lockutils [req-1a9592b0-b842-48b3-8a68-0eb9ac7426c3 req-8f009fa1-8b46-4337-b2fe-8f303864bda8 service nova] Acquiring lock "6438fe40-046c-45d5-9986-8f182ecde49f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 854.315207] env[61957]: DEBUG oslo_concurrency.lockutils [req-1a9592b0-b842-48b3-8a68-0eb9ac7426c3 req-8f009fa1-8b46-4337-b2fe-8f303864bda8 service nova] Lock "6438fe40-046c-45d5-9986-8f182ecde49f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 854.315207] env[61957]: DEBUG oslo_concurrency.lockutils [req-1a9592b0-b842-48b3-8a68-0eb9ac7426c3 req-8f009fa1-8b46-4337-b2fe-8f303864bda8 service nova] Lock "6438fe40-046c-45d5-9986-8f182ecde49f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 854.315883] env[61957]: DEBUG nova.compute.manager [req-1a9592b0-b842-48b3-8a68-0eb9ac7426c3 req-8f009fa1-8b46-4337-b2fe-8f303864bda8 service nova] [instance: 6438fe40-046c-45d5-9986-8f182ecde49f] No waiting events found dispatching network-vif-plugged-e247dac1-1a7a-4b92-a2c0-66cfbe089af3 {{(pid=61957) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 854.315883] env[61957]: WARNING nova.compute.manager [req-1a9592b0-b842-48b3-8a68-0eb9ac7426c3 req-8f009fa1-8b46-4337-b2fe-8f303864bda8 service nova] [instance: 6438fe40-046c-45d5-9986-8f182ecde49f] Received unexpected event network-vif-plugged-e247dac1-1a7a-4b92-a2c0-66cfbe089af3 for instance with vm_state building and task_state spawning. [ 854.315883] env[61957]: DEBUG nova.compute.manager [req-1a9592b0-b842-48b3-8a68-0eb9ac7426c3 req-8f009fa1-8b46-4337-b2fe-8f303864bda8 service nova] [instance: 6438fe40-046c-45d5-9986-8f182ecde49f] Received event network-changed-e247dac1-1a7a-4b92-a2c0-66cfbe089af3 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 854.315883] env[61957]: DEBUG nova.compute.manager [req-1a9592b0-b842-48b3-8a68-0eb9ac7426c3 req-8f009fa1-8b46-4337-b2fe-8f303864bda8 service nova] [instance: 6438fe40-046c-45d5-9986-8f182ecde49f] Refreshing instance network info cache due to event network-changed-e247dac1-1a7a-4b92-a2c0-66cfbe089af3. {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 854.316226] env[61957]: DEBUG oslo_concurrency.lockutils [req-1a9592b0-b842-48b3-8a68-0eb9ac7426c3 req-8f009fa1-8b46-4337-b2fe-8f303864bda8 service nova] Acquiring lock "refresh_cache-6438fe40-046c-45d5-9986-8f182ecde49f" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 854.458373] env[61957]: DEBUG oslo_concurrency.lockutils [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Releasing lock "refresh_cache-6438fe40-046c-45d5-9986-8f182ecde49f" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 854.458690] env[61957]: DEBUG nova.compute.manager [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: 6438fe40-046c-45d5-9986-8f182ecde49f] Instance network_info: |[{"id": "e247dac1-1a7a-4b92-a2c0-66cfbe089af3", "address": "fa:16:3e:8b:8f:72", "network": {"id": "2b5a7a9c-7143-4754-8d84-486c2ab13b21", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1508819166-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "55dd45b4291b456f83a446dfa87f550a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "572b7281-aad3-45fa-9cb2-fc1c70569948", "external-id": "nsx-vlan-transportzone-722", "segmentation_id": 722, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape247dac1-1a", "ovs_interfaceid": "e247dac1-1a7a-4b92-a2c0-66cfbe089af3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 854.459016] env[61957]: DEBUG oslo_concurrency.lockutils [req-1a9592b0-b842-48b3-8a68-0eb9ac7426c3 req-8f009fa1-8b46-4337-b2fe-8f303864bda8 service nova] Acquired lock "refresh_cache-6438fe40-046c-45d5-9986-8f182ecde49f" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 854.459214] env[61957]: DEBUG nova.network.neutron [req-1a9592b0-b842-48b3-8a68-0eb9ac7426c3 req-8f009fa1-8b46-4337-b2fe-8f303864bda8 service nova] [instance: 6438fe40-046c-45d5-9986-8f182ecde49f] Refreshing network info cache for port e247dac1-1a7a-4b92-a2c0-66cfbe089af3 {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 854.460506] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: 6438fe40-046c-45d5-9986-8f182ecde49f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:8b:8f:72', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '572b7281-aad3-45fa-9cb2-fc1c70569948', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e247dac1-1a7a-4b92-a2c0-66cfbe089af3', 'vif_model': 'vmxnet3'}] {{(pid=61957) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 854.467731] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Creating folder: Project (55dd45b4291b456f83a446dfa87f550a). Parent ref: group-v274445. {{(pid=61957) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 854.469341] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-cb1618f3-7a31-4a08-8dbc-40a9bc13e9d0 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.474826] env[61957]: DEBUG oslo_vmware.api [None req-b1306d4e-743c-4d53-8fec-4c8136571e40 tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Task: {'id': task-1277515, 'name': RemoveSnapshot_Task} progress is 15%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.484022] env[61957]: INFO nova.virt.vmwareapi.vm_util [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Created folder: Project (55dd45b4291b456f83a446dfa87f550a) in parent group-v274445. [ 854.484245] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Creating folder: Instances. Parent ref: group-v274508. {{(pid=61957) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 854.484692] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-72e0581b-0460-4bfa-9626-e1ad7cd8430a {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.494472] env[61957]: INFO nova.virt.vmwareapi.vm_util [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Created folder: Instances in parent group-v274508. [ 854.494764] env[61957]: DEBUG oslo.service.loopingcall [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 854.494995] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6438fe40-046c-45d5-9986-8f182ecde49f] Creating VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 854.495263] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-422280ca-5075-4153-98ef-865eed663ec5 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.514221] env[61957]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 854.514221] env[61957]: value = "task-1277519" [ 854.514221] env[61957]: _type = "Task" [ 854.514221] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.525738] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277519, 'name': CreateVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.598459] env[61957]: DEBUG oslo_vmware.api [None req-e4109c08-00a8-4666-8e84-183191cbc2fb tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': task-1277514, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.519675} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.598742] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-e4109c08-00a8-4666-8e84-183191cbc2fb tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore2] 0adae8e1-8c2f-4110-805b-1f286debc833/0adae8e1-8c2f-4110-805b-1f286debc833.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 854.599025] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-e4109c08-00a8-4666-8e84-183191cbc2fb tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: 0adae8e1-8c2f-4110-805b-1f286debc833] Extending root virtual disk to 1048576 {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 854.599376] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9112eb89-6cd5-4c80-bc62-b25fc4e5950e {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.607206] env[61957]: DEBUG oslo_vmware.api [None req-e4109c08-00a8-4666-8e84-183191cbc2fb tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Waiting for the task: (returnval){ [ 854.607206] env[61957]: value = "task-1277520" [ 854.607206] env[61957]: _type = "Task" [ 854.607206] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.614059] env[61957]: DEBUG nova.scheduler.client.report [None req-ac5c4b48-cfd7-420e-9aa1-225385350d11 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 854.617331] env[61957]: DEBUG oslo_vmware.api [None req-d096c609-3f99-4aff-8b69-219b9552f608 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': task-1277516, 'name': Rename_Task, 'duration_secs': 0.179594} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.618375] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-d096c609-3f99-4aff-8b69-219b9552f608 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: 37ce46a1-8f9a-4d15-bd81-e40845a0e48a] Powering on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 854.618674] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-fbd16a81-9a97-4f77-8c6b-cba4609cfe36 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.625099] env[61957]: DEBUG oslo_vmware.api [None req-e4109c08-00a8-4666-8e84-183191cbc2fb tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': task-1277520, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.626691] env[61957]: DEBUG oslo_vmware.api [None req-d096c609-3f99-4aff-8b69-219b9552f608 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Waiting for the task: (returnval){ [ 854.626691] env[61957]: value = "task-1277521" [ 854.626691] env[61957]: _type = "Task" [ 854.626691] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.636559] env[61957]: DEBUG oslo_vmware.api [None req-d096c609-3f99-4aff-8b69-219b9552f608 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': task-1277521, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.660626] env[61957]: DEBUG nova.compute.manager [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: f66db265-887e-4d61-b848-c609e5c884cb] Start spawning the instance on the hypervisor. {{(pid=61957) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 854.693799] env[61957]: DEBUG nova.virt.hardware [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T17:22:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T17:21:50Z,direct_url=,disk_format='vmdk',id=11c76a2c-f705-470a-ba9d-4657858bab38,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='81d3096d0f094373913fc3dc8f5c3c6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T17:21:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 854.694088] env[61957]: DEBUG nova.virt.hardware [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Flavor limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 854.694281] env[61957]: DEBUG nova.virt.hardware [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Image limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 854.694515] env[61957]: DEBUG nova.virt.hardware [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Flavor pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 854.694695] env[61957]: DEBUG nova.virt.hardware [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Image pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 854.694904] env[61957]: DEBUG nova.virt.hardware [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 854.695250] env[61957]: DEBUG nova.virt.hardware [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 854.695393] env[61957]: DEBUG nova.virt.hardware [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 854.695642] env[61957]: DEBUG nova.virt.hardware [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Got 1 possible topologies {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 854.695914] env[61957]: DEBUG nova.virt.hardware [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 854.696242] env[61957]: DEBUG nova.virt.hardware [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 854.698190] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53c0298c-4022-4f6a-9875-ff0bcd92f0f1 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.708540] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77391505-c909-4311-982a-49402a31295c {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.928139] env[61957]: DEBUG nova.network.neutron [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: f66db265-887e-4d61-b848-c609e5c884cb] Successfully updated port: 0289d1f8-a41a-44e5-9b66-6f92ce51debf {{(pid=61957) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 854.964413] env[61957]: DEBUG oslo_vmware.api [None req-b1306d4e-743c-4d53-8fec-4c8136571e40 tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Task: {'id': task-1277515, 'name': RemoveSnapshot_Task} progress is 84%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.024600] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277519, 'name': CreateVM_Task} progress is 99%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.117109] env[61957]: DEBUG oslo_vmware.api [None req-e4109c08-00a8-4666-8e84-183191cbc2fb tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': task-1277520, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.146164} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.117353] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-e4109c08-00a8-4666-8e84-183191cbc2fb tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: 0adae8e1-8c2f-4110-805b-1f286debc833] Extended root virtual disk {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 855.118217] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8364c8d9-288f-44a5-bb09-d79c301e50c3 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.121142] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ac5c4b48-cfd7-420e-9aa1-225385350d11 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.490s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 855.121620] env[61957]: DEBUG nova.compute.manager [None req-ac5c4b48-cfd7-420e-9aa1-225385350d11 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] [instance: c23141ee-0cbb-4d1b-8390-c3073fe354f1] Start building networks asynchronously for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 855.124267] env[61957]: DEBUG oslo_concurrency.lockutils [None req-09e8eb69-1030-4228-8cb0-915d5cee1a52 tempest-AttachInterfacesV270Test-2032541732 tempest-AttachInterfacesV270Test-2032541732-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.729s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 855.125614] env[61957]: INFO nova.compute.claims [None req-09e8eb69-1030-4228-8cb0-915d5cee1a52 tempest-AttachInterfacesV270Test-2032541732 tempest-AttachInterfacesV270Test-2032541732-project-member] [instance: 281b9644-0e7d-48b9-a7b6-45fd6102d558] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 855.152851] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-e4109c08-00a8-4666-8e84-183191cbc2fb tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: 0adae8e1-8c2f-4110-805b-1f286debc833] Reconfiguring VM instance instance-0000003c to attach disk [datastore2] 0adae8e1-8c2f-4110-805b-1f286debc833/0adae8e1-8c2f-4110-805b-1f286debc833.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 855.153208] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c8f1eb71-7514-45a9-9394-d4e65033fff5 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.174376] env[61957]: DEBUG oslo_vmware.api [None req-d096c609-3f99-4aff-8b69-219b9552f608 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': task-1277521, 'name': PowerOnVM_Task} progress is 71%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.175726] env[61957]: DEBUG oslo_vmware.api [None req-e4109c08-00a8-4666-8e84-183191cbc2fb tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Waiting for the task: (returnval){ [ 855.175726] env[61957]: value = "task-1277522" [ 855.175726] env[61957]: _type = "Task" [ 855.175726] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.180987] env[61957]: DEBUG nova.network.neutron [req-1a9592b0-b842-48b3-8a68-0eb9ac7426c3 req-8f009fa1-8b46-4337-b2fe-8f303864bda8 service nova] [instance: 6438fe40-046c-45d5-9986-8f182ecde49f] Updated VIF entry in instance network info cache for port e247dac1-1a7a-4b92-a2c0-66cfbe089af3. {{(pid=61957) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 855.181382] env[61957]: DEBUG nova.network.neutron [req-1a9592b0-b842-48b3-8a68-0eb9ac7426c3 req-8f009fa1-8b46-4337-b2fe-8f303864bda8 service nova] [instance: 6438fe40-046c-45d5-9986-8f182ecde49f] Updating instance_info_cache with network_info: [{"id": "e247dac1-1a7a-4b92-a2c0-66cfbe089af3", "address": "fa:16:3e:8b:8f:72", "network": {"id": "2b5a7a9c-7143-4754-8d84-486c2ab13b21", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1508819166-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "55dd45b4291b456f83a446dfa87f550a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "572b7281-aad3-45fa-9cb2-fc1c70569948", "external-id": "nsx-vlan-transportzone-722", "segmentation_id": 722, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape247dac1-1a", "ovs_interfaceid": "e247dac1-1a7a-4b92-a2c0-66cfbe089af3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 855.185813] env[61957]: DEBUG oslo_vmware.api [None req-e4109c08-00a8-4666-8e84-183191cbc2fb tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': task-1277522, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.432098] env[61957]: DEBUG oslo_concurrency.lockutils [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Acquiring lock "refresh_cache-f66db265-887e-4d61-b848-c609e5c884cb" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 855.432098] env[61957]: DEBUG oslo_concurrency.lockutils [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Acquired lock "refresh_cache-f66db265-887e-4d61-b848-c609e5c884cb" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 855.432098] env[61957]: DEBUG nova.network.neutron [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: f66db265-887e-4d61-b848-c609e5c884cb] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 855.465370] env[61957]: DEBUG oslo_vmware.api [None req-b1306d4e-743c-4d53-8fec-4c8136571e40 tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Task: {'id': task-1277515, 'name': RemoveSnapshot_Task, 'duration_secs': 1.452634} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.467145] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-b1306d4e-743c-4d53-8fec-4c8136571e40 tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] [instance: e0249936-d616-4ffb-8f77-d8107633c42a] Deleted Snapshot of the VM instance {{(pid=61957) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 855.467145] env[61957]: INFO nova.compute.manager [None req-b1306d4e-743c-4d53-8fec-4c8136571e40 tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] [instance: e0249936-d616-4ffb-8f77-d8107633c42a] Took 14.50 seconds to snapshot the instance on the hypervisor. [ 855.526659] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277519, 'name': CreateVM_Task, 'duration_secs': 0.909593} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.526837] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6438fe40-046c-45d5-9986-8f182ecde49f] Created VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 855.527536] env[61957]: DEBUG oslo_concurrency.lockutils [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 855.527710] env[61957]: DEBUG oslo_concurrency.lockutils [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 855.528070] env[61957]: DEBUG oslo_concurrency.lockutils [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 855.528384] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8cff2440-75ff-42e9-86ca-c8a6854f20c4 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.533121] env[61957]: DEBUG oslo_vmware.api [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Waiting for the task: (returnval){ [ 855.533121] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52d5eabf-4b73-7d0d-bcfd-519c0f59409b" [ 855.533121] env[61957]: _type = "Task" [ 855.533121] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.542180] env[61957]: DEBUG oslo_vmware.api [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52d5eabf-4b73-7d0d-bcfd-519c0f59409b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.635312] env[61957]: DEBUG nova.compute.utils [None req-ac5c4b48-cfd7-420e-9aa1-225385350d11 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Using /dev/sd instead of None {{(pid=61957) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 855.636795] env[61957]: DEBUG nova.compute.manager [None req-ac5c4b48-cfd7-420e-9aa1-225385350d11 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] [instance: c23141ee-0cbb-4d1b-8390-c3073fe354f1] Allocating IP information in the background. {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 855.636961] env[61957]: DEBUG nova.network.neutron [None req-ac5c4b48-cfd7-420e-9aa1-225385350d11 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] [instance: c23141ee-0cbb-4d1b-8390-c3073fe354f1] allocate_for_instance() {{(pid=61957) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 855.649620] env[61957]: DEBUG oslo_vmware.api [None req-d096c609-3f99-4aff-8b69-219b9552f608 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': task-1277521, 'name': PowerOnVM_Task, 'duration_secs': 0.887661} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.649620] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-d096c609-3f99-4aff-8b69-219b9552f608 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: 37ce46a1-8f9a-4d15-bd81-e40845a0e48a] Powered on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 855.649620] env[61957]: DEBUG nova.compute.manager [None req-d096c609-3f99-4aff-8b69-219b9552f608 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: 37ce46a1-8f9a-4d15-bd81-e40845a0e48a] Checking state {{(pid=61957) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 855.650316] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe798bee-852f-4fd4-8045-a2888b4f81ca {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.686104] env[61957]: DEBUG oslo_vmware.api [None req-e4109c08-00a8-4666-8e84-183191cbc2fb tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': task-1277522, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.686487] env[61957]: DEBUG oslo_concurrency.lockutils [req-1a9592b0-b842-48b3-8a68-0eb9ac7426c3 req-8f009fa1-8b46-4337-b2fe-8f303864bda8 service nova] Releasing lock "refresh_cache-6438fe40-046c-45d5-9986-8f182ecde49f" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 855.690842] env[61957]: DEBUG nova.policy [None req-ac5c4b48-cfd7-420e-9aa1-225385350d11 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd0405e9c38cb4012a0212284be7e9aed', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1673c854272f4b14a623a73b8f4b687b', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61957) authorize /opt/stack/nova/nova/policy.py:203}} [ 855.976176] env[61957]: DEBUG nova.network.neutron [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: f66db265-887e-4d61-b848-c609e5c884cb] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 856.018967] env[61957]: DEBUG nova.network.neutron [None req-ac5c4b48-cfd7-420e-9aa1-225385350d11 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] [instance: c23141ee-0cbb-4d1b-8390-c3073fe354f1] Successfully created port: a9884dda-2fb3-4e26-84b1-877dd0e0459c {{(pid=61957) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 856.043803] env[61957]: DEBUG oslo_vmware.api [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52d5eabf-4b73-7d0d-bcfd-519c0f59409b, 'name': SearchDatastore_Task, 'duration_secs': 0.031535} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.046592] env[61957]: DEBUG oslo_concurrency.lockutils [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 856.046826] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: 6438fe40-046c-45d5-9986-8f182ecde49f] Processing image 11c76a2c-f705-470a-ba9d-4657858bab38 {{(pid=61957) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 856.047158] env[61957]: DEBUG oslo_concurrency.lockutils [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 856.047345] env[61957]: DEBUG oslo_concurrency.lockutils [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 856.047550] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 856.047816] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-85ddc571-f9bf-4781-b77a-93f37d59d9a7 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.056397] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 856.056582] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61957) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 856.057515] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d49aab56-c4d9-4a96-861e-352151858841 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.064270] env[61957]: DEBUG oslo_vmware.api [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Waiting for the task: (returnval){ [ 856.064270] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]5257c764-03d1-7101-3703-b881afb22abf" [ 856.064270] env[61957]: _type = "Task" [ 856.064270] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.072404] env[61957]: DEBUG oslo_vmware.api [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]5257c764-03d1-7101-3703-b881afb22abf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.145708] env[61957]: DEBUG nova.compute.manager [None req-ac5c4b48-cfd7-420e-9aa1-225385350d11 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] [instance: c23141ee-0cbb-4d1b-8390-c3073fe354f1] Start building block device mappings for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 856.173766] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d096c609-3f99-4aff-8b69-219b9552f608 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 856.195513] env[61957]: DEBUG oslo_vmware.api [None req-e4109c08-00a8-4666-8e84-183191cbc2fb tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': task-1277522, 'name': ReconfigVM_Task, 'duration_secs': 0.550882} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.196492] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-e4109c08-00a8-4666-8e84-183191cbc2fb tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: 0adae8e1-8c2f-4110-805b-1f286debc833] Reconfigured VM instance instance-0000003c to attach disk [datastore2] 0adae8e1-8c2f-4110-805b-1f286debc833/0adae8e1-8c2f-4110-805b-1f286debc833.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 856.196492] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b518d5b1-1a2f-4fae-835c-f915139bc6e1 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.205726] env[61957]: DEBUG oslo_vmware.api [None req-e4109c08-00a8-4666-8e84-183191cbc2fb tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Waiting for the task: (returnval){ [ 856.205726] env[61957]: value = "task-1277523" [ 856.205726] env[61957]: _type = "Task" [ 856.205726] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.216057] env[61957]: DEBUG oslo_vmware.api [None req-e4109c08-00a8-4666-8e84-183191cbc2fb tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': task-1277523, 'name': Rename_Task} progress is 5%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.399420] env[61957]: DEBUG nova.network.neutron [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: f66db265-887e-4d61-b848-c609e5c884cb] Updating instance_info_cache with network_info: [{"id": "0289d1f8-a41a-44e5-9b66-6f92ce51debf", "address": "fa:16:3e:a0:ed:a1", "network": {"id": "2b5a7a9c-7143-4754-8d84-486c2ab13b21", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1508819166-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "55dd45b4291b456f83a446dfa87f550a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "572b7281-aad3-45fa-9cb2-fc1c70569948", "external-id": "nsx-vlan-transportzone-722", "segmentation_id": 722, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0289d1f8-a4", "ovs_interfaceid": "0289d1f8-a41a-44e5-9b66-6f92ce51debf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 856.490579] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dff21533-e6f2-43e6-81b1-6b60d5973893 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.498210] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9bcd004-beea-486b-8eb0-73bd8db0b0b4 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.533628] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8acca1cd-79d7-482f-8c7d-5b0920300d74 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.537253] env[61957]: DEBUG nova.compute.manager [req-353b6cf9-83b6-435f-8de2-a88dc3ed18c1 req-44c420f1-5db5-4352-b978-d531cda2b37e service nova] [instance: f66db265-887e-4d61-b848-c609e5c884cb] Received event network-vif-plugged-0289d1f8-a41a-44e5-9b66-6f92ce51debf {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 856.537497] env[61957]: DEBUG oslo_concurrency.lockutils [req-353b6cf9-83b6-435f-8de2-a88dc3ed18c1 req-44c420f1-5db5-4352-b978-d531cda2b37e service nova] Acquiring lock "f66db265-887e-4d61-b848-c609e5c884cb-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 856.537602] env[61957]: DEBUG oslo_concurrency.lockutils [req-353b6cf9-83b6-435f-8de2-a88dc3ed18c1 req-44c420f1-5db5-4352-b978-d531cda2b37e service nova] Lock "f66db265-887e-4d61-b848-c609e5c884cb-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 856.537783] env[61957]: DEBUG oslo_concurrency.lockutils [req-353b6cf9-83b6-435f-8de2-a88dc3ed18c1 req-44c420f1-5db5-4352-b978-d531cda2b37e service nova] Lock "f66db265-887e-4d61-b848-c609e5c884cb-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 856.537920] env[61957]: DEBUG nova.compute.manager [req-353b6cf9-83b6-435f-8de2-a88dc3ed18c1 req-44c420f1-5db5-4352-b978-d531cda2b37e service nova] [instance: f66db265-887e-4d61-b848-c609e5c884cb] No waiting events found dispatching network-vif-plugged-0289d1f8-a41a-44e5-9b66-6f92ce51debf {{(pid=61957) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 856.538121] env[61957]: WARNING nova.compute.manager [req-353b6cf9-83b6-435f-8de2-a88dc3ed18c1 req-44c420f1-5db5-4352-b978-d531cda2b37e service nova] [instance: f66db265-887e-4d61-b848-c609e5c884cb] Received unexpected event network-vif-plugged-0289d1f8-a41a-44e5-9b66-6f92ce51debf for instance with vm_state building and task_state spawning. [ 856.538292] env[61957]: DEBUG nova.compute.manager [req-353b6cf9-83b6-435f-8de2-a88dc3ed18c1 req-44c420f1-5db5-4352-b978-d531cda2b37e service nova] [instance: f66db265-887e-4d61-b848-c609e5c884cb] Received event network-changed-0289d1f8-a41a-44e5-9b66-6f92ce51debf {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 856.538460] env[61957]: DEBUG nova.compute.manager [req-353b6cf9-83b6-435f-8de2-a88dc3ed18c1 req-44c420f1-5db5-4352-b978-d531cda2b37e service nova] [instance: f66db265-887e-4d61-b848-c609e5c884cb] Refreshing instance network info cache due to event network-changed-0289d1f8-a41a-44e5-9b66-6f92ce51debf. {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 856.538634] env[61957]: DEBUG oslo_concurrency.lockutils [req-353b6cf9-83b6-435f-8de2-a88dc3ed18c1 req-44c420f1-5db5-4352-b978-d531cda2b37e service nova] Acquiring lock "refresh_cache-f66db265-887e-4d61-b848-c609e5c884cb" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 856.545794] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3452195-e03b-4f41-a5dc-d8a70cacb8f5 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.560084] env[61957]: DEBUG nova.compute.provider_tree [None req-09e8eb69-1030-4228-8cb0-915d5cee1a52 tempest-AttachInterfacesV270Test-2032541732 tempest-AttachInterfacesV270Test-2032541732-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 856.573094] env[61957]: DEBUG oslo_vmware.api [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]5257c764-03d1-7101-3703-b881afb22abf, 'name': SearchDatastore_Task, 'duration_secs': 0.014365} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.573869] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-88d6e2c7-d82c-4258-86fc-f6a1f4661cdb {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.579869] env[61957]: DEBUG oslo_vmware.api [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Waiting for the task: (returnval){ [ 856.579869] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52784c03-f583-06fb-a519-3ba387dd5486" [ 856.579869] env[61957]: _type = "Task" [ 856.579869] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.587889] env[61957]: DEBUG oslo_vmware.api [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52784c03-f583-06fb-a519-3ba387dd5486, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.716066] env[61957]: DEBUG oslo_vmware.api [None req-e4109c08-00a8-4666-8e84-183191cbc2fb tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': task-1277523, 'name': Rename_Task, 'duration_secs': 0.206022} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.716395] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-e4109c08-00a8-4666-8e84-183191cbc2fb tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: 0adae8e1-8c2f-4110-805b-1f286debc833] Powering on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 856.716395] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f3c8b613-ee00-4d66-a78f-1ebbe3eade32 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.722919] env[61957]: DEBUG oslo_vmware.api [None req-e4109c08-00a8-4666-8e84-183191cbc2fb tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Waiting for the task: (returnval){ [ 856.722919] env[61957]: value = "task-1277524" [ 856.722919] env[61957]: _type = "Task" [ 856.722919] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.730474] env[61957]: DEBUG oslo_vmware.api [None req-e4109c08-00a8-4666-8e84-183191cbc2fb tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': task-1277524, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.902863] env[61957]: DEBUG oslo_concurrency.lockutils [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Releasing lock "refresh_cache-f66db265-887e-4d61-b848-c609e5c884cb" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 856.903191] env[61957]: DEBUG nova.compute.manager [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: f66db265-887e-4d61-b848-c609e5c884cb] Instance network_info: |[{"id": "0289d1f8-a41a-44e5-9b66-6f92ce51debf", "address": "fa:16:3e:a0:ed:a1", "network": {"id": "2b5a7a9c-7143-4754-8d84-486c2ab13b21", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1508819166-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "55dd45b4291b456f83a446dfa87f550a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "572b7281-aad3-45fa-9cb2-fc1c70569948", "external-id": "nsx-vlan-transportzone-722", "segmentation_id": 722, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0289d1f8-a4", "ovs_interfaceid": "0289d1f8-a41a-44e5-9b66-6f92ce51debf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 856.903534] env[61957]: DEBUG oslo_concurrency.lockutils [req-353b6cf9-83b6-435f-8de2-a88dc3ed18c1 req-44c420f1-5db5-4352-b978-d531cda2b37e service nova] Acquired lock "refresh_cache-f66db265-887e-4d61-b848-c609e5c884cb" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 856.903726] env[61957]: DEBUG nova.network.neutron [req-353b6cf9-83b6-435f-8de2-a88dc3ed18c1 req-44c420f1-5db5-4352-b978-d531cda2b37e service nova] [instance: f66db265-887e-4d61-b848-c609e5c884cb] Refreshing network info cache for port 0289d1f8-a41a-44e5-9b66-6f92ce51debf {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 856.905069] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: f66db265-887e-4d61-b848-c609e5c884cb] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a0:ed:a1', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '572b7281-aad3-45fa-9cb2-fc1c70569948', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0289d1f8-a41a-44e5-9b66-6f92ce51debf', 'vif_model': 'vmxnet3'}] {{(pid=61957) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 856.912549] env[61957]: DEBUG oslo.service.loopingcall [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 856.915344] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f66db265-887e-4d61-b848-c609e5c884cb] Creating VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 856.915832] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2f21ed39-4593-4c26-bab2-7e9bf9094c72 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.936120] env[61957]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 856.936120] env[61957]: value = "task-1277525" [ 856.936120] env[61957]: _type = "Task" [ 856.936120] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.946779] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277525, 'name': CreateVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.066345] env[61957]: DEBUG nova.scheduler.client.report [None req-09e8eb69-1030-4228-8cb0-915d5cee1a52 tempest-AttachInterfacesV270Test-2032541732 tempest-AttachInterfacesV270Test-2032541732-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 857.090463] env[61957]: DEBUG oslo_vmware.api [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52784c03-f583-06fb-a519-3ba387dd5486, 'name': SearchDatastore_Task, 'duration_secs': 0.013418} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.091424] env[61957]: DEBUG oslo_concurrency.lockutils [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 857.091700] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore2] 6438fe40-046c-45d5-9986-8f182ecde49f/6438fe40-046c-45d5-9986-8f182ecde49f.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 857.091982] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d0c2903a-1b4d-4d4f-9d6e-ce0a45cde85f {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.098997] env[61957]: DEBUG oslo_vmware.api [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Waiting for the task: (returnval){ [ 857.098997] env[61957]: value = "task-1277526" [ 857.098997] env[61957]: _type = "Task" [ 857.098997] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.107441] env[61957]: DEBUG oslo_vmware.api [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Task: {'id': task-1277526, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.155198] env[61957]: DEBUG nova.compute.manager [None req-ac5c4b48-cfd7-420e-9aa1-225385350d11 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] [instance: c23141ee-0cbb-4d1b-8390-c3073fe354f1] Start spawning the instance on the hypervisor. {{(pid=61957) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 857.158250] env[61957]: DEBUG nova.network.neutron [req-353b6cf9-83b6-435f-8de2-a88dc3ed18c1 req-44c420f1-5db5-4352-b978-d531cda2b37e service nova] [instance: f66db265-887e-4d61-b848-c609e5c884cb] Updated VIF entry in instance network info cache for port 0289d1f8-a41a-44e5-9b66-6f92ce51debf. {{(pid=61957) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 857.158763] env[61957]: DEBUG nova.network.neutron [req-353b6cf9-83b6-435f-8de2-a88dc3ed18c1 req-44c420f1-5db5-4352-b978-d531cda2b37e service nova] [instance: f66db265-887e-4d61-b848-c609e5c884cb] Updating instance_info_cache with network_info: [{"id": "0289d1f8-a41a-44e5-9b66-6f92ce51debf", "address": "fa:16:3e:a0:ed:a1", "network": {"id": "2b5a7a9c-7143-4754-8d84-486c2ab13b21", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1508819166-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "55dd45b4291b456f83a446dfa87f550a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "572b7281-aad3-45fa-9cb2-fc1c70569948", "external-id": "nsx-vlan-transportzone-722", "segmentation_id": 722, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0289d1f8-a4", "ovs_interfaceid": "0289d1f8-a41a-44e5-9b66-6f92ce51debf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 857.180938] env[61957]: DEBUG nova.virt.hardware [None req-ac5c4b48-cfd7-420e-9aa1-225385350d11 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T17:22:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T17:21:50Z,direct_url=,disk_format='vmdk',id=11c76a2c-f705-470a-ba9d-4657858bab38,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='81d3096d0f094373913fc3dc8f5c3c6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T17:21:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 857.181301] env[61957]: DEBUG nova.virt.hardware [None req-ac5c4b48-cfd7-420e-9aa1-225385350d11 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Flavor limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 857.181548] env[61957]: DEBUG nova.virt.hardware [None req-ac5c4b48-cfd7-420e-9aa1-225385350d11 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Image limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 857.181813] env[61957]: DEBUG nova.virt.hardware [None req-ac5c4b48-cfd7-420e-9aa1-225385350d11 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Flavor pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 857.182103] env[61957]: DEBUG nova.virt.hardware [None req-ac5c4b48-cfd7-420e-9aa1-225385350d11 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Image pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 857.182338] env[61957]: DEBUG nova.virt.hardware [None req-ac5c4b48-cfd7-420e-9aa1-225385350d11 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 857.182688] env[61957]: DEBUG nova.virt.hardware [None req-ac5c4b48-cfd7-420e-9aa1-225385350d11 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 857.182792] env[61957]: DEBUG nova.virt.hardware [None req-ac5c4b48-cfd7-420e-9aa1-225385350d11 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 857.182969] env[61957]: DEBUG nova.virt.hardware [None req-ac5c4b48-cfd7-420e-9aa1-225385350d11 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Got 1 possible topologies {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 857.183151] env[61957]: DEBUG nova.virt.hardware [None req-ac5c4b48-cfd7-420e-9aa1-225385350d11 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 857.183328] env[61957]: DEBUG nova.virt.hardware [None req-ac5c4b48-cfd7-420e-9aa1-225385350d11 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 857.184704] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0957ee1-5683-4575-a3db-d16c38d49dbb {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.193807] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ae3ec58-5941-4c87-baeb-899f740bcb5a {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.233524] env[61957]: DEBUG oslo_vmware.api [None req-e4109c08-00a8-4666-8e84-183191cbc2fb tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': task-1277524, 'name': PowerOnVM_Task} progress is 1%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.448397] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277525, 'name': CreateVM_Task, 'duration_secs': 0.356699} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.448767] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f66db265-887e-4d61-b848-c609e5c884cb] Created VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 857.449682] env[61957]: DEBUG oslo_concurrency.lockutils [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 857.449903] env[61957]: DEBUG oslo_concurrency.lockutils [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Acquired lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 857.450315] env[61957]: DEBUG oslo_concurrency.lockutils [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 857.450696] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ed5ccd47-965b-47f6-adc9-9d8c067cc2d4 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.456323] env[61957]: DEBUG oslo_vmware.api [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Waiting for the task: (returnval){ [ 857.456323] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52d08f56-322c-541e-7ddc-cfbaa3fb1101" [ 857.456323] env[61957]: _type = "Task" [ 857.456323] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.466026] env[61957]: DEBUG oslo_vmware.api [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52d08f56-322c-541e-7ddc-cfbaa3fb1101, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.571903] env[61957]: DEBUG oslo_concurrency.lockutils [None req-09e8eb69-1030-4228-8cb0-915d5cee1a52 tempest-AttachInterfacesV270Test-2032541732 tempest-AttachInterfacesV270Test-2032541732-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.447s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 857.572563] env[61957]: DEBUG nova.compute.manager [None req-09e8eb69-1030-4228-8cb0-915d5cee1a52 tempest-AttachInterfacesV270Test-2032541732 tempest-AttachInterfacesV270Test-2032541732-project-member] [instance: 281b9644-0e7d-48b9-a7b6-45fd6102d558] Start building networks asynchronously for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 857.576050] env[61957]: DEBUG oslo_concurrency.lockutils [None req-5a8f4b06-9979-4e23-adff-500ed36e9a0d tempest-ServerGroupTestJSON-950728618 tempest-ServerGroupTestJSON-950728618-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 24.800s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 857.576389] env[61957]: DEBUG oslo_concurrency.lockutils [None req-5a8f4b06-9979-4e23-adff-500ed36e9a0d tempest-ServerGroupTestJSON-950728618 tempest-ServerGroupTestJSON-950728618-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 857.581311] env[61957]: DEBUG oslo_concurrency.lockutils [None req-47642d83-3473-4b34-8e43-c1548504523c tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 23.964s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 857.581588] env[61957]: DEBUG oslo_concurrency.lockutils [None req-47642d83-3473-4b34-8e43-c1548504523c tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 857.584867] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d61a7353-b862-4c7b-8175-41d701129e22 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 23.178s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 857.584867] env[61957]: DEBUG nova.objects.instance [None req-d61a7353-b862-4c7b-8175-41d701129e22 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Lazy-loading 'resources' on Instance uuid 14ba830d-4e2c-4e9a-a059-3c86209f0127 {{(pid=61957) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 857.610220] env[61957]: DEBUG oslo_vmware.api [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Task: {'id': task-1277526, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.612650] env[61957]: INFO nova.scheduler.client.report [None req-5a8f4b06-9979-4e23-adff-500ed36e9a0d tempest-ServerGroupTestJSON-950728618 tempest-ServerGroupTestJSON-950728618-project-member] Deleted allocations for instance e9ac118f-08b3-430b-848c-461c2b2e3e02 [ 857.615098] env[61957]: INFO nova.scheduler.client.report [None req-47642d83-3473-4b34-8e43-c1548504523c tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] Deleted allocations for instance 615dbbed-2b02-4351-9e03-8c13f424a133 [ 857.662575] env[61957]: DEBUG oslo_concurrency.lockutils [req-353b6cf9-83b6-435f-8de2-a88dc3ed18c1 req-44c420f1-5db5-4352-b978-d531cda2b37e service nova] Releasing lock "refresh_cache-f66db265-887e-4d61-b848-c609e5c884cb" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 857.734582] env[61957]: DEBUG oslo_vmware.api [None req-e4109c08-00a8-4666-8e84-183191cbc2fb tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': task-1277524, 'name': PowerOnVM_Task} progress is 82%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.744525] env[61957]: DEBUG oslo_concurrency.lockutils [None req-a0f6bb40-5b1e-4b57-987f-3c85bacc19c6 tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Acquiring lock "e0249936-d616-4ffb-8f77-d8107633c42a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 857.744951] env[61957]: DEBUG oslo_concurrency.lockutils [None req-a0f6bb40-5b1e-4b57-987f-3c85bacc19c6 tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Lock "e0249936-d616-4ffb-8f77-d8107633c42a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 857.745326] env[61957]: DEBUG oslo_concurrency.lockutils [None req-a0f6bb40-5b1e-4b57-987f-3c85bacc19c6 tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Acquiring lock "e0249936-d616-4ffb-8f77-d8107633c42a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 857.745628] env[61957]: DEBUG oslo_concurrency.lockutils [None req-a0f6bb40-5b1e-4b57-987f-3c85bacc19c6 tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Lock "e0249936-d616-4ffb-8f77-d8107633c42a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 857.745905] env[61957]: DEBUG oslo_concurrency.lockutils [None req-a0f6bb40-5b1e-4b57-987f-3c85bacc19c6 tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Lock "e0249936-d616-4ffb-8f77-d8107633c42a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 857.748353] env[61957]: INFO nova.compute.manager [None req-a0f6bb40-5b1e-4b57-987f-3c85bacc19c6 tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] [instance: e0249936-d616-4ffb-8f77-d8107633c42a] Terminating instance [ 857.751469] env[61957]: DEBUG nova.compute.manager [None req-a0f6bb40-5b1e-4b57-987f-3c85bacc19c6 tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] [instance: e0249936-d616-4ffb-8f77-d8107633c42a] Start destroying the instance on the hypervisor. {{(pid=61957) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 857.751783] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-a0f6bb40-5b1e-4b57-987f-3c85bacc19c6 tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] [instance: e0249936-d616-4ffb-8f77-d8107633c42a] Destroying instance {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 857.752755] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d0000ef-08c9-485b-a5ea-bb851088565c {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.761201] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-a0f6bb40-5b1e-4b57-987f-3c85bacc19c6 tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] [instance: e0249936-d616-4ffb-8f77-d8107633c42a] Powering off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 857.761201] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f6e39a59-06e1-491c-8194-8b1fe750e8ad {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.769276] env[61957]: DEBUG oslo_vmware.api [None req-a0f6bb40-5b1e-4b57-987f-3c85bacc19c6 tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Waiting for the task: (returnval){ [ 857.769276] env[61957]: value = "task-1277527" [ 857.769276] env[61957]: _type = "Task" [ 857.769276] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.774947] env[61957]: DEBUG oslo_vmware.api [None req-a0f6bb40-5b1e-4b57-987f-3c85bacc19c6 tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Task: {'id': task-1277527, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.801245] env[61957]: DEBUG nova.network.neutron [None req-ac5c4b48-cfd7-420e-9aa1-225385350d11 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] [instance: c23141ee-0cbb-4d1b-8390-c3073fe354f1] Successfully updated port: a9884dda-2fb3-4e26-84b1-877dd0e0459c {{(pid=61957) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 857.966883] env[61957]: DEBUG oslo_vmware.api [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52d08f56-322c-541e-7ddc-cfbaa3fb1101, 'name': SearchDatastore_Task, 'duration_secs': 0.020895} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.967259] env[61957]: DEBUG oslo_concurrency.lockutils [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Releasing lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 857.967502] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: f66db265-887e-4d61-b848-c609e5c884cb] Processing image 11c76a2c-f705-470a-ba9d-4657858bab38 {{(pid=61957) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 857.967822] env[61957]: DEBUG oslo_concurrency.lockutils [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 857.967906] env[61957]: DEBUG oslo_concurrency.lockutils [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Acquired lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 857.968109] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 857.968430] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-fe24e36b-eb47-49cd-984f-8a7c9439ab5b {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.981510] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 857.981713] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61957) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 857.982498] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5c0165ef-c65c-4ccf-990c-a4902e3ca9ae {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.988358] env[61957]: DEBUG oslo_vmware.api [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Waiting for the task: (returnval){ [ 857.988358] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]525fa3c1-a087-df22-b9de-aaaf7306ba30" [ 857.988358] env[61957]: _type = "Task" [ 857.988358] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.991693] env[61957]: DEBUG oslo_concurrency.lockutils [None req-95a860a5-ca00-4c71-b827-6e3af6a25042 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Acquiring lock "37ce46a1-8f9a-4d15-bd81-e40845a0e48a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 857.991946] env[61957]: DEBUG oslo_concurrency.lockutils [None req-95a860a5-ca00-4c71-b827-6e3af6a25042 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Lock "37ce46a1-8f9a-4d15-bd81-e40845a0e48a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 857.992183] env[61957]: DEBUG oslo_concurrency.lockutils [None req-95a860a5-ca00-4c71-b827-6e3af6a25042 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Acquiring lock "37ce46a1-8f9a-4d15-bd81-e40845a0e48a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 857.992440] env[61957]: DEBUG oslo_concurrency.lockutils [None req-95a860a5-ca00-4c71-b827-6e3af6a25042 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Lock "37ce46a1-8f9a-4d15-bd81-e40845a0e48a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 857.992655] env[61957]: DEBUG oslo_concurrency.lockutils [None req-95a860a5-ca00-4c71-b827-6e3af6a25042 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Lock "37ce46a1-8f9a-4d15-bd81-e40845a0e48a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 857.995472] env[61957]: INFO nova.compute.manager [None req-95a860a5-ca00-4c71-b827-6e3af6a25042 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: 37ce46a1-8f9a-4d15-bd81-e40845a0e48a] Terminating instance [ 857.999676] env[61957]: DEBUG oslo_vmware.api [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]525fa3c1-a087-df22-b9de-aaaf7306ba30, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.999828] env[61957]: DEBUG nova.compute.manager [None req-95a860a5-ca00-4c71-b827-6e3af6a25042 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: 37ce46a1-8f9a-4d15-bd81-e40845a0e48a] Start destroying the instance on the hypervisor. {{(pid=61957) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 857.999886] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-95a860a5-ca00-4c71-b827-6e3af6a25042 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: 37ce46a1-8f9a-4d15-bd81-e40845a0e48a] Destroying instance {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 858.000692] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e509de57-7633-434f-9e1a-e7df23130176 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.007423] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-95a860a5-ca00-4c71-b827-6e3af6a25042 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: 37ce46a1-8f9a-4d15-bd81-e40845a0e48a] Powering off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 858.007670] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-cae7e6a0-994b-4d92-b81b-a506e9376928 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.014251] env[61957]: DEBUG oslo_vmware.api [None req-95a860a5-ca00-4c71-b827-6e3af6a25042 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Waiting for the task: (returnval){ [ 858.014251] env[61957]: value = "task-1277528" [ 858.014251] env[61957]: _type = "Task" [ 858.014251] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.023331] env[61957]: DEBUG oslo_vmware.api [None req-95a860a5-ca00-4c71-b827-6e3af6a25042 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': task-1277528, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.088060] env[61957]: DEBUG nova.compute.utils [None req-09e8eb69-1030-4228-8cb0-915d5cee1a52 tempest-AttachInterfacesV270Test-2032541732 tempest-AttachInterfacesV270Test-2032541732-project-member] Using /dev/sd instead of None {{(pid=61957) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 858.093023] env[61957]: DEBUG nova.compute.manager [None req-09e8eb69-1030-4228-8cb0-915d5cee1a52 tempest-AttachInterfacesV270Test-2032541732 tempest-AttachInterfacesV270Test-2032541732-project-member] [instance: 281b9644-0e7d-48b9-a7b6-45fd6102d558] Allocating IP information in the background. {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 858.093023] env[61957]: DEBUG nova.network.neutron [None req-09e8eb69-1030-4228-8cb0-915d5cee1a52 tempest-AttachInterfacesV270Test-2032541732 tempest-AttachInterfacesV270Test-2032541732-project-member] [instance: 281b9644-0e7d-48b9-a7b6-45fd6102d558] allocate_for_instance() {{(pid=61957) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 858.112553] env[61957]: DEBUG oslo_vmware.api [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Task: {'id': task-1277526, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.522719} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.112553] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore2] 6438fe40-046c-45d5-9986-8f182ecde49f/6438fe40-046c-45d5-9986-8f182ecde49f.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 858.112553] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: 6438fe40-046c-45d5-9986-8f182ecde49f] Extending root virtual disk to 1048576 {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 858.112553] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2463e14d-40d0-41c6-8294-10c50334623c {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.121607] env[61957]: DEBUG oslo_vmware.api [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Waiting for the task: (returnval){ [ 858.121607] env[61957]: value = "task-1277529" [ 858.121607] env[61957]: _type = "Task" [ 858.121607] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.125784] env[61957]: DEBUG oslo_concurrency.lockutils [None req-5a8f4b06-9979-4e23-adff-500ed36e9a0d tempest-ServerGroupTestJSON-950728618 tempest-ServerGroupTestJSON-950728618-project-member] Lock "e9ac118f-08b3-430b-848c-461c2b2e3e02" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 29.336s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 858.129504] env[61957]: DEBUG oslo_concurrency.lockutils [None req-47642d83-3473-4b34-8e43-c1548504523c tempest-ServersTestMultiNic-1158900003 tempest-ServersTestMultiNic-1158900003-project-member] Lock "615dbbed-2b02-4351-9e03-8c13f424a133" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 28.812s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 858.142068] env[61957]: DEBUG oslo_vmware.api [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Task: {'id': task-1277529, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.150878] env[61957]: DEBUG nova.policy [None req-09e8eb69-1030-4228-8cb0-915d5cee1a52 tempest-AttachInterfacesV270Test-2032541732 tempest-AttachInterfacesV270Test-2032541732-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b25f68a471ec4f338c1e122d764de6dc', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd3bc71403439465a8df6b74e6d6ce0e1', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61957) authorize /opt/stack/nova/nova/policy.py:203}} [ 858.234169] env[61957]: DEBUG oslo_vmware.api [None req-e4109c08-00a8-4666-8e84-183191cbc2fb tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': task-1277524, 'name': PowerOnVM_Task, 'duration_secs': 1.23774} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.234540] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-e4109c08-00a8-4666-8e84-183191cbc2fb tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: 0adae8e1-8c2f-4110-805b-1f286debc833] Powered on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 858.234973] env[61957]: INFO nova.compute.manager [None req-e4109c08-00a8-4666-8e84-183191cbc2fb tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: 0adae8e1-8c2f-4110-805b-1f286debc833] Took 8.49 seconds to spawn the instance on the hypervisor. [ 858.234973] env[61957]: DEBUG nova.compute.manager [None req-e4109c08-00a8-4666-8e84-183191cbc2fb tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: 0adae8e1-8c2f-4110-805b-1f286debc833] Checking state {{(pid=61957) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 858.236434] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61059a88-3a62-44fc-87b5-660d852ff635 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.282963] env[61957]: DEBUG oslo_vmware.api [None req-a0f6bb40-5b1e-4b57-987f-3c85bacc19c6 tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Task: {'id': task-1277527, 'name': PowerOffVM_Task, 'duration_secs': 0.352249} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.283301] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-a0f6bb40-5b1e-4b57-987f-3c85bacc19c6 tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] [instance: e0249936-d616-4ffb-8f77-d8107633c42a] Powered off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 858.283661] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-a0f6bb40-5b1e-4b57-987f-3c85bacc19c6 tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] [instance: e0249936-d616-4ffb-8f77-d8107633c42a] Unregistering the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 858.284014] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5401c8da-6283-49ee-aead-1c4938239dc5 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.305291] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ac5c4b48-cfd7-420e-9aa1-225385350d11 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Acquiring lock "refresh_cache-c23141ee-0cbb-4d1b-8390-c3073fe354f1" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 858.305291] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ac5c4b48-cfd7-420e-9aa1-225385350d11 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Acquired lock "refresh_cache-c23141ee-0cbb-4d1b-8390-c3073fe354f1" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 858.305291] env[61957]: DEBUG nova.network.neutron [None req-ac5c4b48-cfd7-420e-9aa1-225385350d11 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] [instance: c23141ee-0cbb-4d1b-8390-c3073fe354f1] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 858.359587] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-a0f6bb40-5b1e-4b57-987f-3c85bacc19c6 tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] [instance: e0249936-d616-4ffb-8f77-d8107633c42a] Unregistered the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 858.359767] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-a0f6bb40-5b1e-4b57-987f-3c85bacc19c6 tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] [instance: e0249936-d616-4ffb-8f77-d8107633c42a] Deleting contents of the VM from datastore datastore2 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 858.359951] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-a0f6bb40-5b1e-4b57-987f-3c85bacc19c6 tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Deleting the datastore file [datastore2] e0249936-d616-4ffb-8f77-d8107633c42a {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 858.360267] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c7b839eb-89af-4f85-8d77-4597085c66d6 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.368487] env[61957]: DEBUG oslo_vmware.api [None req-a0f6bb40-5b1e-4b57-987f-3c85bacc19c6 tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Waiting for the task: (returnval){ [ 858.368487] env[61957]: value = "task-1277531" [ 858.368487] env[61957]: _type = "Task" [ 858.368487] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.377110] env[61957]: DEBUG oslo_vmware.api [None req-a0f6bb40-5b1e-4b57-987f-3c85bacc19c6 tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Task: {'id': task-1277531, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.481564] env[61957]: DEBUG nova.network.neutron [None req-09e8eb69-1030-4228-8cb0-915d5cee1a52 tempest-AttachInterfacesV270Test-2032541732 tempest-AttachInterfacesV270Test-2032541732-project-member] [instance: 281b9644-0e7d-48b9-a7b6-45fd6102d558] Successfully created port: 22b42651-5480-45d9-b363-9106dd6a7ce6 {{(pid=61957) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 858.503368] env[61957]: DEBUG oslo_vmware.api [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]525fa3c1-a087-df22-b9de-aaaf7306ba30, 'name': SearchDatastore_Task, 'duration_secs': 0.025092} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.509756] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17dc9b3d-fa31-401c-b801-c64f3efafddb {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.509756] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-499dee83-562b-4aed-9e91-9bdff6feee3f {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.522102] env[61957]: DEBUG oslo_vmware.api [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Waiting for the task: (returnval){ [ 858.522102] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]525417ab-4eac-1589-81b8-bbbadf168d21" [ 858.522102] env[61957]: _type = "Task" [ 858.522102] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.523894] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be52248a-0aee-42b8-b1d3-801dec40725f {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.534120] env[61957]: DEBUG oslo_vmware.api [None req-95a860a5-ca00-4c71-b827-6e3af6a25042 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': task-1277528, 'name': PowerOffVM_Task, 'duration_secs': 0.254702} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.534854] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-95a860a5-ca00-4c71-b827-6e3af6a25042 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: 37ce46a1-8f9a-4d15-bd81-e40845a0e48a] Powered off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 858.535173] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-95a860a5-ca00-4c71-b827-6e3af6a25042 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: 37ce46a1-8f9a-4d15-bd81-e40845a0e48a] Unregistering the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 858.535803] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c5df4ae1-fe33-451a-a9ed-c6f596cc8007 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.571923] env[61957]: DEBUG oslo_vmware.api [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]525417ab-4eac-1589-81b8-bbbadf168d21, 'name': SearchDatastore_Task, 'duration_secs': 0.018648} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.574227] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6168e320-da3b-42f0-b1e1-4017e108a294 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.576874] env[61957]: DEBUG oslo_concurrency.lockutils [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Releasing lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 858.577167] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore1] f66db265-887e-4d61-b848-c609e5c884cb/f66db265-887e-4d61-b848-c609e5c884cb.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 858.578283] env[61957]: DEBUG nova.compute.manager [req-519a0ffd-2b3b-4a40-875b-fcffde67f084 req-6e6a3808-8138-4d27-ae8b-52a5c50f2d9c service nova] [instance: c23141ee-0cbb-4d1b-8390-c3073fe354f1] Received event network-vif-plugged-a9884dda-2fb3-4e26-84b1-877dd0e0459c {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 858.578487] env[61957]: DEBUG oslo_concurrency.lockutils [req-519a0ffd-2b3b-4a40-875b-fcffde67f084 req-6e6a3808-8138-4d27-ae8b-52a5c50f2d9c service nova] Acquiring lock "c23141ee-0cbb-4d1b-8390-c3073fe354f1-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 858.578689] env[61957]: DEBUG oslo_concurrency.lockutils [req-519a0ffd-2b3b-4a40-875b-fcffde67f084 req-6e6a3808-8138-4d27-ae8b-52a5c50f2d9c service nova] Lock "c23141ee-0cbb-4d1b-8390-c3073fe354f1-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 858.578850] env[61957]: DEBUG oslo_concurrency.lockutils [req-519a0ffd-2b3b-4a40-875b-fcffde67f084 req-6e6a3808-8138-4d27-ae8b-52a5c50f2d9c service nova] Lock "c23141ee-0cbb-4d1b-8390-c3073fe354f1-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 858.579035] env[61957]: DEBUG nova.compute.manager [req-519a0ffd-2b3b-4a40-875b-fcffde67f084 req-6e6a3808-8138-4d27-ae8b-52a5c50f2d9c service nova] [instance: c23141ee-0cbb-4d1b-8390-c3073fe354f1] No waiting events found dispatching network-vif-plugged-a9884dda-2fb3-4e26-84b1-877dd0e0459c {{(pid=61957) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 858.579220] env[61957]: WARNING nova.compute.manager [req-519a0ffd-2b3b-4a40-875b-fcffde67f084 req-6e6a3808-8138-4d27-ae8b-52a5c50f2d9c service nova] [instance: c23141ee-0cbb-4d1b-8390-c3073fe354f1] Received unexpected event network-vif-plugged-a9884dda-2fb3-4e26-84b1-877dd0e0459c for instance with vm_state building and task_state spawning. [ 858.579567] env[61957]: DEBUG nova.compute.manager [req-519a0ffd-2b3b-4a40-875b-fcffde67f084 req-6e6a3808-8138-4d27-ae8b-52a5c50f2d9c service nova] [instance: c23141ee-0cbb-4d1b-8390-c3073fe354f1] Received event network-changed-a9884dda-2fb3-4e26-84b1-877dd0e0459c {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 858.579743] env[61957]: DEBUG nova.compute.manager [req-519a0ffd-2b3b-4a40-875b-fcffde67f084 req-6e6a3808-8138-4d27-ae8b-52a5c50f2d9c service nova] [instance: c23141ee-0cbb-4d1b-8390-c3073fe354f1] Refreshing instance network info cache due to event network-changed-a9884dda-2fb3-4e26-84b1-877dd0e0459c. {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 858.579743] env[61957]: DEBUG oslo_concurrency.lockutils [req-519a0ffd-2b3b-4a40-875b-fcffde67f084 req-6e6a3808-8138-4d27-ae8b-52a5c50f2d9c service nova] Acquiring lock "refresh_cache-c23141ee-0cbb-4d1b-8390-c3073fe354f1" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 858.580322] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-53f011a7-a66a-4960-adef-e6a47a5d6051 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.588084] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e06653b3-837a-478c-90a0-78ba0aa811e9 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.594117] env[61957]: DEBUG oslo_vmware.api [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Waiting for the task: (returnval){ [ 858.594117] env[61957]: value = "task-1277533" [ 858.594117] env[61957]: _type = "Task" [ 858.594117] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.595122] env[61957]: DEBUG nova.compute.manager [None req-09e8eb69-1030-4228-8cb0-915d5cee1a52 tempest-AttachInterfacesV270Test-2032541732 tempest-AttachInterfacesV270Test-2032541732-project-member] [instance: 281b9644-0e7d-48b9-a7b6-45fd6102d558] Start building block device mappings for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 858.610049] env[61957]: DEBUG nova.compute.provider_tree [None req-d61a7353-b862-4c7b-8175-41d701129e22 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 858.616974] env[61957]: DEBUG oslo_vmware.api [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Task: {'id': task-1277533, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.642309] env[61957]: DEBUG oslo_vmware.api [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Task: {'id': task-1277529, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071974} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.644713] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: 6438fe40-046c-45d5-9986-8f182ecde49f] Extended root virtual disk {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 858.647776] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74752ce3-9fee-4066-b3a0-bb1975453a40 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.651443] env[61957]: DEBUG oslo_concurrency.lockutils [None req-3c689a75-14e7-47a1-b918-cef3025704e3 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Acquiring lock "dbeeb200-70b9-4cb4-b5a4-182389d21918" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 858.651690] env[61957]: DEBUG oslo_concurrency.lockutils [None req-3c689a75-14e7-47a1-b918-cef3025704e3 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Lock "dbeeb200-70b9-4cb4-b5a4-182389d21918" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 858.652967] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-95a860a5-ca00-4c71-b827-6e3af6a25042 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: 37ce46a1-8f9a-4d15-bd81-e40845a0e48a] Unregistered the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 858.653191] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-95a860a5-ca00-4c71-b827-6e3af6a25042 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: 37ce46a1-8f9a-4d15-bd81-e40845a0e48a] Deleting contents of the VM from datastore datastore2 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 858.653378] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-95a860a5-ca00-4c71-b827-6e3af6a25042 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Deleting the datastore file [datastore2] 37ce46a1-8f9a-4d15-bd81-e40845a0e48a {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 858.654318] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-863d227a-8939-4442-a97f-b2f40af19167 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.674768] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: 6438fe40-046c-45d5-9986-8f182ecde49f] Reconfiguring VM instance instance-0000003d to attach disk [datastore2] 6438fe40-046c-45d5-9986-8f182ecde49f/6438fe40-046c-45d5-9986-8f182ecde49f.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 858.676481] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-59e2e09f-f43b-4d1e-83a5-16620df5dba7 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.693237] env[61957]: DEBUG oslo_vmware.api [None req-95a860a5-ca00-4c71-b827-6e3af6a25042 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Waiting for the task: (returnval){ [ 858.693237] env[61957]: value = "task-1277534" [ 858.693237] env[61957]: _type = "Task" [ 858.693237] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.698704] env[61957]: DEBUG oslo_vmware.api [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Waiting for the task: (returnval){ [ 858.698704] env[61957]: value = "task-1277535" [ 858.698704] env[61957]: _type = "Task" [ 858.698704] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.706843] env[61957]: DEBUG oslo_vmware.api [None req-95a860a5-ca00-4c71-b827-6e3af6a25042 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': task-1277534, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.709914] env[61957]: DEBUG oslo_vmware.api [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Task: {'id': task-1277535, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.757606] env[61957]: INFO nova.compute.manager [None req-e4109c08-00a8-4666-8e84-183191cbc2fb tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: 0adae8e1-8c2f-4110-805b-1f286debc833] Took 35.93 seconds to build instance. [ 858.880316] env[61957]: DEBUG oslo_vmware.api [None req-a0f6bb40-5b1e-4b57-987f-3c85bacc19c6 tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Task: {'id': task-1277531, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.315762} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.881403] env[61957]: DEBUG nova.network.neutron [None req-ac5c4b48-cfd7-420e-9aa1-225385350d11 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] [instance: c23141ee-0cbb-4d1b-8390-c3073fe354f1] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 858.883799] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-a0f6bb40-5b1e-4b57-987f-3c85bacc19c6 tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Deleted the datastore file {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 858.884078] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-a0f6bb40-5b1e-4b57-987f-3c85bacc19c6 tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] [instance: e0249936-d616-4ffb-8f77-d8107633c42a] Deleted contents of the VM from datastore datastore2 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 858.884328] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-a0f6bb40-5b1e-4b57-987f-3c85bacc19c6 tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] [instance: e0249936-d616-4ffb-8f77-d8107633c42a] Instance destroyed {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 858.884576] env[61957]: INFO nova.compute.manager [None req-a0f6bb40-5b1e-4b57-987f-3c85bacc19c6 tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] [instance: e0249936-d616-4ffb-8f77-d8107633c42a] Took 1.13 seconds to destroy the instance on the hypervisor. [ 858.884914] env[61957]: DEBUG oslo.service.loopingcall [None req-a0f6bb40-5b1e-4b57-987f-3c85bacc19c6 tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 858.885339] env[61957]: DEBUG nova.compute.manager [-] [instance: e0249936-d616-4ffb-8f77-d8107633c42a] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 858.885339] env[61957]: DEBUG nova.network.neutron [-] [instance: e0249936-d616-4ffb-8f77-d8107633c42a] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 859.106331] env[61957]: DEBUG oslo_vmware.api [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Task: {'id': task-1277533, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.117020] env[61957]: DEBUG nova.scheduler.client.report [None req-d61a7353-b862-4c7b-8175-41d701129e22 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 859.208021] env[61957]: DEBUG oslo_vmware.api [None req-95a860a5-ca00-4c71-b827-6e3af6a25042 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': task-1277534, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.241163} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.208021] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-95a860a5-ca00-4c71-b827-6e3af6a25042 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Deleted the datastore file {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 859.208021] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-95a860a5-ca00-4c71-b827-6e3af6a25042 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: 37ce46a1-8f9a-4d15-bd81-e40845a0e48a] Deleted contents of the VM from datastore datastore2 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 859.208021] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-95a860a5-ca00-4c71-b827-6e3af6a25042 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: 37ce46a1-8f9a-4d15-bd81-e40845a0e48a] Instance destroyed {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 859.208021] env[61957]: INFO nova.compute.manager [None req-95a860a5-ca00-4c71-b827-6e3af6a25042 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: 37ce46a1-8f9a-4d15-bd81-e40845a0e48a] Took 1.21 seconds to destroy the instance on the hypervisor. [ 859.208021] env[61957]: DEBUG oslo.service.loopingcall [None req-95a860a5-ca00-4c71-b827-6e3af6a25042 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 859.208021] env[61957]: DEBUG nova.compute.manager [-] [instance: 37ce46a1-8f9a-4d15-bd81-e40845a0e48a] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 859.208021] env[61957]: DEBUG nova.network.neutron [-] [instance: 37ce46a1-8f9a-4d15-bd81-e40845a0e48a] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 859.211454] env[61957]: DEBUG oslo_vmware.api [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Task: {'id': task-1277535, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.259469] env[61957]: DEBUG oslo_concurrency.lockutils [None req-e4109c08-00a8-4666-8e84-183191cbc2fb tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Lock "0adae8e1-8c2f-4110-805b-1f286debc833" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 109.243s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 859.362258] env[61957]: DEBUG nova.network.neutron [None req-ac5c4b48-cfd7-420e-9aa1-225385350d11 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] [instance: c23141ee-0cbb-4d1b-8390-c3073fe354f1] Updating instance_info_cache with network_info: [{"id": "a9884dda-2fb3-4e26-84b1-877dd0e0459c", "address": "fa:16:3e:f9:46:03", "network": {"id": "82fd2357-53b6-4903-963e-a2866736643e", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-612681178-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1673c854272f4b14a623a73b8f4b687b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6eb7e3e9-5cc2-40f1-a6eb-f70f06531667", "external-id": "nsx-vlan-transportzone-938", "segmentation_id": 938, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa9884dda-2f", "ovs_interfaceid": "a9884dda-2fb3-4e26-84b1-877dd0e0459c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 859.507656] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b24ed70-d221-4345-b4bb-a8ba1820d94c {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.515768] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-0f86f5c0-66c7-4015-9f77-1a170b4bd5bc tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: 0adae8e1-8c2f-4110-805b-1f286debc833] Suspending the VM {{(pid=61957) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1163}} [ 859.516108] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-ee6a58f4-dca2-4c8c-9921-7ea93cc95bb9 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.522280] env[61957]: DEBUG oslo_vmware.api [None req-0f86f5c0-66c7-4015-9f77-1a170b4bd5bc tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Waiting for the task: (returnval){ [ 859.522280] env[61957]: value = "task-1277536" [ 859.522280] env[61957]: _type = "Task" [ 859.522280] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.530477] env[61957]: DEBUG oslo_vmware.api [None req-0f86f5c0-66c7-4015-9f77-1a170b4bd5bc tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': task-1277536, 'name': SuspendVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.606653] env[61957]: DEBUG oslo_vmware.api [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Task: {'id': task-1277533, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.568816} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.607082] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore1] f66db265-887e-4d61-b848-c609e5c884cb/f66db265-887e-4d61-b848-c609e5c884cb.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 859.607315] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: f66db265-887e-4d61-b848-c609e5c884cb] Extending root virtual disk to 1048576 {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 859.607592] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0ebd6d8c-db29-44ab-a720-1103c331a8e7 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.617656] env[61957]: DEBUG nova.compute.manager [None req-09e8eb69-1030-4228-8cb0-915d5cee1a52 tempest-AttachInterfacesV270Test-2032541732 tempest-AttachInterfacesV270Test-2032541732-project-member] [instance: 281b9644-0e7d-48b9-a7b6-45fd6102d558] Start spawning the instance on the hypervisor. {{(pid=61957) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 859.625118] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d61a7353-b862-4c7b-8175-41d701129e22 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.040s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 859.627251] env[61957]: DEBUG oslo_vmware.api [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Waiting for the task: (returnval){ [ 859.627251] env[61957]: value = "task-1277537" [ 859.627251] env[61957]: _type = "Task" [ 859.627251] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.629194] env[61957]: DEBUG oslo_concurrency.lockutils [None req-902199c7-5b38-4a0e-a88f-80539210637b tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.167s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 859.630904] env[61957]: INFO nova.compute.claims [None req-902199c7-5b38-4a0e-a88f-80539210637b tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] [instance: 526ff179-62a6-4763-ab25-797617c4ed57] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 859.646104] env[61957]: DEBUG oslo_vmware.api [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Task: {'id': task-1277537, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.658794] env[61957]: DEBUG nova.virt.hardware [None req-09e8eb69-1030-4228-8cb0-915d5cee1a52 tempest-AttachInterfacesV270Test-2032541732 tempest-AttachInterfacesV270Test-2032541732-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T17:22:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T17:21:50Z,direct_url=,disk_format='vmdk',id=11c76a2c-f705-470a-ba9d-4657858bab38,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='81d3096d0f094373913fc3dc8f5c3c6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T17:21:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 859.659299] env[61957]: DEBUG nova.virt.hardware [None req-09e8eb69-1030-4228-8cb0-915d5cee1a52 tempest-AttachInterfacesV270Test-2032541732 tempest-AttachInterfacesV270Test-2032541732-project-member] Flavor limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 859.659485] env[61957]: DEBUG nova.virt.hardware [None req-09e8eb69-1030-4228-8cb0-915d5cee1a52 tempest-AttachInterfacesV270Test-2032541732 tempest-AttachInterfacesV270Test-2032541732-project-member] Image limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 859.659687] env[61957]: DEBUG nova.virt.hardware [None req-09e8eb69-1030-4228-8cb0-915d5cee1a52 tempest-AttachInterfacesV270Test-2032541732 tempest-AttachInterfacesV270Test-2032541732-project-member] Flavor pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 859.659834] env[61957]: DEBUG nova.virt.hardware [None req-09e8eb69-1030-4228-8cb0-915d5cee1a52 tempest-AttachInterfacesV270Test-2032541732 tempest-AttachInterfacesV270Test-2032541732-project-member] Image pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 859.659983] env[61957]: DEBUG nova.virt.hardware [None req-09e8eb69-1030-4228-8cb0-915d5cee1a52 tempest-AttachInterfacesV270Test-2032541732 tempest-AttachInterfacesV270Test-2032541732-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 859.660307] env[61957]: DEBUG nova.virt.hardware [None req-09e8eb69-1030-4228-8cb0-915d5cee1a52 tempest-AttachInterfacesV270Test-2032541732 tempest-AttachInterfacesV270Test-2032541732-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 859.660479] env[61957]: DEBUG nova.virt.hardware [None req-09e8eb69-1030-4228-8cb0-915d5cee1a52 tempest-AttachInterfacesV270Test-2032541732 tempest-AttachInterfacesV270Test-2032541732-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 859.660652] env[61957]: DEBUG nova.virt.hardware [None req-09e8eb69-1030-4228-8cb0-915d5cee1a52 tempest-AttachInterfacesV270Test-2032541732 tempest-AttachInterfacesV270Test-2032541732-project-member] Got 1 possible topologies {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 859.660819] env[61957]: DEBUG nova.virt.hardware [None req-09e8eb69-1030-4228-8cb0-915d5cee1a52 tempest-AttachInterfacesV270Test-2032541732 tempest-AttachInterfacesV270Test-2032541732-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 859.661846] env[61957]: DEBUG nova.virt.hardware [None req-09e8eb69-1030-4228-8cb0-915d5cee1a52 tempest-AttachInterfacesV270Test-2032541732 tempest-AttachInterfacesV270Test-2032541732-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 859.662152] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f78cec7-5260-48ee-bad2-b44e0ff4b2bb {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.667299] env[61957]: INFO nova.scheduler.client.report [None req-d61a7353-b862-4c7b-8175-41d701129e22 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Deleted allocations for instance 14ba830d-4e2c-4e9a-a059-3c86209f0127 [ 859.679913] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce083bb3-7d2d-4c5e-9257-b5c960391282 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.708303] env[61957]: DEBUG oslo_vmware.api [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Task: {'id': task-1277535, 'name': ReconfigVM_Task, 'duration_secs': 0.552147} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.708636] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: 6438fe40-046c-45d5-9986-8f182ecde49f] Reconfigured VM instance instance-0000003d to attach disk [datastore2] 6438fe40-046c-45d5-9986-8f182ecde49f/6438fe40-046c-45d5-9986-8f182ecde49f.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 859.709278] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4ebbe4d6-685f-4079-b103-bb7c3bfb7ac0 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.715559] env[61957]: DEBUG oslo_vmware.api [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Waiting for the task: (returnval){ [ 859.715559] env[61957]: value = "task-1277538" [ 859.715559] env[61957]: _type = "Task" [ 859.715559] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.725022] env[61957]: DEBUG oslo_vmware.api [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Task: {'id': task-1277538, 'name': Rename_Task} progress is 5%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.765765] env[61957]: DEBUG nova.compute.manager [None req-bafb915b-a249-4473-bf3a-4c41863ba98f tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] [instance: 28ab7d23-487f-4ae1-8fe4-58db55b59918] Starting instance... {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 859.865864] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ac5c4b48-cfd7-420e-9aa1-225385350d11 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Releasing lock "refresh_cache-c23141ee-0cbb-4d1b-8390-c3073fe354f1" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 859.866252] env[61957]: DEBUG nova.compute.manager [None req-ac5c4b48-cfd7-420e-9aa1-225385350d11 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] [instance: c23141ee-0cbb-4d1b-8390-c3073fe354f1] Instance network_info: |[{"id": "a9884dda-2fb3-4e26-84b1-877dd0e0459c", "address": "fa:16:3e:f9:46:03", "network": {"id": "82fd2357-53b6-4903-963e-a2866736643e", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-612681178-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1673c854272f4b14a623a73b8f4b687b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6eb7e3e9-5cc2-40f1-a6eb-f70f06531667", "external-id": "nsx-vlan-transportzone-938", "segmentation_id": 938, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa9884dda-2f", "ovs_interfaceid": "a9884dda-2fb3-4e26-84b1-877dd0e0459c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 859.866624] env[61957]: DEBUG oslo_concurrency.lockutils [req-519a0ffd-2b3b-4a40-875b-fcffde67f084 req-6e6a3808-8138-4d27-ae8b-52a5c50f2d9c service nova] Acquired lock "refresh_cache-c23141ee-0cbb-4d1b-8390-c3073fe354f1" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 859.866843] env[61957]: DEBUG nova.network.neutron [req-519a0ffd-2b3b-4a40-875b-fcffde67f084 req-6e6a3808-8138-4d27-ae8b-52a5c50f2d9c service nova] [instance: c23141ee-0cbb-4d1b-8390-c3073fe354f1] Refreshing network info cache for port a9884dda-2fb3-4e26-84b1-877dd0e0459c {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 859.868249] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-ac5c4b48-cfd7-420e-9aa1-225385350d11 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] [instance: c23141ee-0cbb-4d1b-8390-c3073fe354f1] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f9:46:03', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '6eb7e3e9-5cc2-40f1-a6eb-f70f06531667', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a9884dda-2fb3-4e26-84b1-877dd0e0459c', 'vif_model': 'vmxnet3'}] {{(pid=61957) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 859.876951] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-ac5c4b48-cfd7-420e-9aa1-225385350d11 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Creating folder: Project (1673c854272f4b14a623a73b8f4b687b). Parent ref: group-v274445. {{(pid=61957) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 859.877522] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-866eb245-e03f-4624-9306-044cda3a7e2f {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.888814] env[61957]: INFO nova.virt.vmwareapi.vm_util [None req-ac5c4b48-cfd7-420e-9aa1-225385350d11 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Created folder: Project (1673c854272f4b14a623a73b8f4b687b) in parent group-v274445. [ 859.889089] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-ac5c4b48-cfd7-420e-9aa1-225385350d11 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Creating folder: Instances. Parent ref: group-v274512. {{(pid=61957) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 859.889648] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-00cfb499-a76e-4a9f-a09a-3865bd8879d2 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.898425] env[61957]: INFO nova.virt.vmwareapi.vm_util [None req-ac5c4b48-cfd7-420e-9aa1-225385350d11 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Created folder: Instances in parent group-v274512. [ 859.898615] env[61957]: DEBUG oslo.service.loopingcall [None req-ac5c4b48-cfd7-420e-9aa1-225385350d11 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 859.898807] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c23141ee-0cbb-4d1b-8390-c3073fe354f1] Creating VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 859.899039] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d6170f32-1be8-40ad-97b2-cde99cfbfbac {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.920101] env[61957]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 859.920101] env[61957]: value = "task-1277541" [ 859.920101] env[61957]: _type = "Task" [ 859.920101] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.929085] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277541, 'name': CreateVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.034594] env[61957]: DEBUG oslo_vmware.api [None req-0f86f5c0-66c7-4015-9f77-1a170b4bd5bc tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': task-1277536, 'name': SuspendVM_Task} progress is 33%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.117179] env[61957]: DEBUG nova.network.neutron [-] [instance: e0249936-d616-4ffb-8f77-d8107633c42a] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 860.151982] env[61957]: DEBUG oslo_vmware.api [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Task: {'id': task-1277537, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071046} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.151982] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: f66db265-887e-4d61-b848-c609e5c884cb] Extended root virtual disk {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 860.151982] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a5b4ffb-52c5-46ce-8ee0-995b235ce158 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.177907] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: f66db265-887e-4d61-b848-c609e5c884cb] Reconfiguring VM instance instance-0000003e to attach disk [datastore1] f66db265-887e-4d61-b848-c609e5c884cb/f66db265-887e-4d61-b848-c609e5c884cb.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 860.180721] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-64010d0b-06cc-4ad1-a6c8-73b0b518f45e {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.198668] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d61a7353-b862-4c7b-8175-41d701129e22 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Lock "14ba830d-4e2c-4e9a-a059-3c86209f0127" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 28.857s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 860.206716] env[61957]: DEBUG oslo_vmware.api [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Waiting for the task: (returnval){ [ 860.206716] env[61957]: value = "task-1277542" [ 860.206716] env[61957]: _type = "Task" [ 860.206716] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.219484] env[61957]: DEBUG oslo_vmware.api [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Task: {'id': task-1277542, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.229582] env[61957]: DEBUG oslo_vmware.api [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Task: {'id': task-1277538, 'name': Rename_Task} progress is 99%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.287766] env[61957]: DEBUG nova.network.neutron [-] [instance: 37ce46a1-8f9a-4d15-bd81-e40845a0e48a] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 860.299908] env[61957]: DEBUG oslo_concurrency.lockutils [None req-bafb915b-a249-4473-bf3a-4c41863ba98f tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 860.432762] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277541, 'name': CreateVM_Task} progress is 99%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.540125] env[61957]: DEBUG oslo_vmware.api [None req-0f86f5c0-66c7-4015-9f77-1a170b4bd5bc tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': task-1277536, 'name': SuspendVM_Task, 'duration_secs': 1.001586} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.540655] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-0f86f5c0-66c7-4015-9f77-1a170b4bd5bc tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: 0adae8e1-8c2f-4110-805b-1f286debc833] Suspended the VM {{(pid=61957) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1167}} [ 860.541932] env[61957]: DEBUG nova.compute.manager [None req-0f86f5c0-66c7-4015-9f77-1a170b4bd5bc tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: 0adae8e1-8c2f-4110-805b-1f286debc833] Checking state {{(pid=61957) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 860.542997] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4cbdf963-7ab6-436b-8e1a-a383d7e1ac76 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.620467] env[61957]: INFO nova.compute.manager [-] [instance: e0249936-d616-4ffb-8f77-d8107633c42a] Took 1.73 seconds to deallocate network for instance. [ 860.724102] env[61957]: DEBUG oslo_vmware.api [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Task: {'id': task-1277542, 'name': ReconfigVM_Task, 'duration_secs': 0.416738} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.728971] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: f66db265-887e-4d61-b848-c609e5c884cb] Reconfigured VM instance instance-0000003e to attach disk [datastore1] f66db265-887e-4d61-b848-c609e5c884cb/f66db265-887e-4d61-b848-c609e5c884cb.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 860.732456] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d55591ae-5590-4a5e-aa57-9ba09b2aeca8 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.740193] env[61957]: DEBUG oslo_vmware.api [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Task: {'id': task-1277538, 'name': Rename_Task, 'duration_secs': 0.553504} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.742126] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: 6438fe40-046c-45d5-9986-8f182ecde49f] Powering on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 860.742487] env[61957]: DEBUG oslo_vmware.api [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Waiting for the task: (returnval){ [ 860.742487] env[61957]: value = "task-1277543" [ 860.742487] env[61957]: _type = "Task" [ 860.742487] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.742735] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c9678585-994c-4b90-8853-02230dcdf144 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.759393] env[61957]: DEBUG oslo_vmware.api [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Waiting for the task: (returnval){ [ 860.759393] env[61957]: value = "task-1277544" [ 860.759393] env[61957]: _type = "Task" [ 860.759393] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.764983] env[61957]: DEBUG oslo_vmware.api [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Task: {'id': task-1277543, 'name': Rename_Task} progress is 14%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.773971] env[61957]: DEBUG oslo_vmware.api [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Task: {'id': task-1277544, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.790499] env[61957]: INFO nova.compute.manager [-] [instance: 37ce46a1-8f9a-4d15-bd81-e40845a0e48a] Took 1.58 seconds to deallocate network for instance. [ 860.818580] env[61957]: DEBUG nova.network.neutron [None req-09e8eb69-1030-4228-8cb0-915d5cee1a52 tempest-AttachInterfacesV270Test-2032541732 tempest-AttachInterfacesV270Test-2032541732-project-member] [instance: 281b9644-0e7d-48b9-a7b6-45fd6102d558] Successfully updated port: 22b42651-5480-45d9-b363-9106dd6a7ce6 {{(pid=61957) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 860.934986] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277541, 'name': CreateVM_Task, 'duration_secs': 0.598981} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.939256] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c23141ee-0cbb-4d1b-8390-c3073fe354f1] Created VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 860.939256] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ac5c4b48-cfd7-420e-9aa1-225385350d11 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 860.939442] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ac5c4b48-cfd7-420e-9aa1-225385350d11 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Acquired lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 860.939874] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ac5c4b48-cfd7-420e-9aa1-225385350d11 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 860.940342] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-451d0f1a-2611-457a-84a0-53154ed6d11f {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.947612] env[61957]: DEBUG oslo_vmware.api [None req-ac5c4b48-cfd7-420e-9aa1-225385350d11 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Waiting for the task: (returnval){ [ 860.947612] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]528518bc-34f4-461c-6399-579a40cb33c1" [ 860.947612] env[61957]: _type = "Task" [ 860.947612] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.959669] env[61957]: DEBUG oslo_vmware.api [None req-ac5c4b48-cfd7-420e-9aa1-225385350d11 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]528518bc-34f4-461c-6399-579a40cb33c1, 'name': SearchDatastore_Task, 'duration_secs': 0.011004} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.962272] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ac5c4b48-cfd7-420e-9aa1-225385350d11 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Releasing lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 860.962272] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-ac5c4b48-cfd7-420e-9aa1-225385350d11 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] [instance: c23141ee-0cbb-4d1b-8390-c3073fe354f1] Processing image 11c76a2c-f705-470a-ba9d-4657858bab38 {{(pid=61957) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 860.962272] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ac5c4b48-cfd7-420e-9aa1-225385350d11 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 860.962272] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ac5c4b48-cfd7-420e-9aa1-225385350d11 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Acquired lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 860.962272] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-ac5c4b48-cfd7-420e-9aa1-225385350d11 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 860.963592] env[61957]: DEBUG nova.compute.manager [req-ccdf11d1-c2c4-448f-9ecb-3769b32c76af req-05255898-fe0e-4f25-bf85-98c657342a88 service nova] [instance: e0249936-d616-4ffb-8f77-d8107633c42a] Received event network-vif-deleted-3852bc01-a0a4-4b76-9461-9eaf58378500 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 860.963774] env[61957]: DEBUG nova.compute.manager [req-ccdf11d1-c2c4-448f-9ecb-3769b32c76af req-05255898-fe0e-4f25-bf85-98c657342a88 service nova] [instance: 37ce46a1-8f9a-4d15-bd81-e40845a0e48a] Received event network-vif-deleted-7234f796-0abc-4e81-ba04-29607283618c {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 860.963963] env[61957]: DEBUG nova.compute.manager [req-ccdf11d1-c2c4-448f-9ecb-3769b32c76af req-05255898-fe0e-4f25-bf85-98c657342a88 service nova] [instance: 281b9644-0e7d-48b9-a7b6-45fd6102d558] Received event network-vif-plugged-22b42651-5480-45d9-b363-9106dd6a7ce6 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 860.964370] env[61957]: DEBUG oslo_concurrency.lockutils [req-ccdf11d1-c2c4-448f-9ecb-3769b32c76af req-05255898-fe0e-4f25-bf85-98c657342a88 service nova] Acquiring lock "281b9644-0e7d-48b9-a7b6-45fd6102d558-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 860.964370] env[61957]: DEBUG oslo_concurrency.lockutils [req-ccdf11d1-c2c4-448f-9ecb-3769b32c76af req-05255898-fe0e-4f25-bf85-98c657342a88 service nova] Lock "281b9644-0e7d-48b9-a7b6-45fd6102d558-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 860.964485] env[61957]: DEBUG oslo_concurrency.lockutils [req-ccdf11d1-c2c4-448f-9ecb-3769b32c76af req-05255898-fe0e-4f25-bf85-98c657342a88 service nova] Lock "281b9644-0e7d-48b9-a7b6-45fd6102d558-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 860.965193] env[61957]: DEBUG nova.compute.manager [req-ccdf11d1-c2c4-448f-9ecb-3769b32c76af req-05255898-fe0e-4f25-bf85-98c657342a88 service nova] [instance: 281b9644-0e7d-48b9-a7b6-45fd6102d558] No waiting events found dispatching network-vif-plugged-22b42651-5480-45d9-b363-9106dd6a7ce6 {{(pid=61957) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 860.965193] env[61957]: WARNING nova.compute.manager [req-ccdf11d1-c2c4-448f-9ecb-3769b32c76af req-05255898-fe0e-4f25-bf85-98c657342a88 service nova] [instance: 281b9644-0e7d-48b9-a7b6-45fd6102d558] Received unexpected event network-vif-plugged-22b42651-5480-45d9-b363-9106dd6a7ce6 for instance with vm_state building and task_state spawning. [ 860.966033] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-192e3b87-48ed-473c-a9e9-78fc1b8f19d6 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.977326] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-ac5c4b48-cfd7-420e-9aa1-225385350d11 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 860.977562] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-ac5c4b48-cfd7-420e-9aa1-225385350d11 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61957) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 860.978371] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2e3b59ec-515b-4b45-9b89-c5ecc953ef61 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.986896] env[61957]: DEBUG oslo_vmware.api [None req-ac5c4b48-cfd7-420e-9aa1-225385350d11 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Waiting for the task: (returnval){ [ 860.986896] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52a9a392-3c50-f984-f812-b8cb50554ea2" [ 860.986896] env[61957]: _type = "Task" [ 860.986896] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.998478] env[61957]: DEBUG oslo_vmware.api [None req-ac5c4b48-cfd7-420e-9aa1-225385350d11 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52a9a392-3c50-f984-f812-b8cb50554ea2, 'name': SearchDatastore_Task, 'duration_secs': 0.007933} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.002921] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7c9160a5-bc62-4f91-8d3c-49e13874613f {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.008986] env[61957]: DEBUG oslo_vmware.api [None req-ac5c4b48-cfd7-420e-9aa1-225385350d11 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Waiting for the task: (returnval){ [ 861.008986] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52e6000a-4e8f-79a8-6f7f-daad83db4aab" [ 861.008986] env[61957]: _type = "Task" [ 861.008986] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.022705] env[61957]: DEBUG oslo_vmware.api [None req-ac5c4b48-cfd7-420e-9aa1-225385350d11 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52e6000a-4e8f-79a8-6f7f-daad83db4aab, 'name': SearchDatastore_Task} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.025782] env[61957]: DEBUG nova.network.neutron [req-519a0ffd-2b3b-4a40-875b-fcffde67f084 req-6e6a3808-8138-4d27-ae8b-52a5c50f2d9c service nova] [instance: c23141ee-0cbb-4d1b-8390-c3073fe354f1] Updated VIF entry in instance network info cache for port a9884dda-2fb3-4e26-84b1-877dd0e0459c. {{(pid=61957) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 861.026142] env[61957]: DEBUG nova.network.neutron [req-519a0ffd-2b3b-4a40-875b-fcffde67f084 req-6e6a3808-8138-4d27-ae8b-52a5c50f2d9c service nova] [instance: c23141ee-0cbb-4d1b-8390-c3073fe354f1] Updating instance_info_cache with network_info: [{"id": "a9884dda-2fb3-4e26-84b1-877dd0e0459c", "address": "fa:16:3e:f9:46:03", "network": {"id": "82fd2357-53b6-4903-963e-a2866736643e", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-612681178-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1673c854272f4b14a623a73b8f4b687b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6eb7e3e9-5cc2-40f1-a6eb-f70f06531667", "external-id": "nsx-vlan-transportzone-938", "segmentation_id": 938, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa9884dda-2f", "ovs_interfaceid": "a9884dda-2fb3-4e26-84b1-877dd0e0459c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 861.028492] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ac5c4b48-cfd7-420e-9aa1-225385350d11 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Releasing lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 861.028492] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-ac5c4b48-cfd7-420e-9aa1-225385350d11 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore1] c23141ee-0cbb-4d1b-8390-c3073fe354f1/c23141ee-0cbb-4d1b-8390-c3073fe354f1.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 861.028492] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-31a30b4c-bd68-4161-bdc2-1fde15c31f74 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.035768] env[61957]: DEBUG oslo_vmware.api [None req-ac5c4b48-cfd7-420e-9aa1-225385350d11 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Waiting for the task: (returnval){ [ 861.035768] env[61957]: value = "task-1277545" [ 861.035768] env[61957]: _type = "Task" [ 861.035768] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.045748] env[61957]: DEBUG oslo_vmware.api [None req-ac5c4b48-cfd7-420e-9aa1-225385350d11 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Task: {'id': task-1277545, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.132593] env[61957]: DEBUG oslo_concurrency.lockutils [None req-a0f6bb40-5b1e-4b57-987f-3c85bacc19c6 tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 861.154780] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd23d0c1-5393-48d9-b1d2-2bec48d2613f {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.162322] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-262b3f9f-0c3d-4986-8707-d62aeda7e582 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.195092] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1875f6fa-dc89-497b-818d-9f2423fa0c28 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.205071] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0de39750-36b8-4a73-b37d-ee59e2f0c4b9 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.219664] env[61957]: DEBUG nova.compute.provider_tree [None req-902199c7-5b38-4a0e-a88f-80539210637b tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 861.255688] env[61957]: DEBUG oslo_vmware.api [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Task: {'id': task-1277543, 'name': Rename_Task, 'duration_secs': 0.219078} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.256135] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: f66db265-887e-4d61-b848-c609e5c884cb] Powering on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 861.256457] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-125f364d-fbfc-4bba-b20c-8b0cc1377f65 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.269252] env[61957]: DEBUG oslo_vmware.api [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Task: {'id': task-1277544, 'name': PowerOnVM_Task, 'duration_secs': 0.465164} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.273240] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: 6438fe40-046c-45d5-9986-8f182ecde49f] Powered on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 861.273240] env[61957]: INFO nova.compute.manager [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: 6438fe40-046c-45d5-9986-8f182ecde49f] Took 9.09 seconds to spawn the instance on the hypervisor. [ 861.273240] env[61957]: DEBUG nova.compute.manager [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: 6438fe40-046c-45d5-9986-8f182ecde49f] Checking state {{(pid=61957) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 861.273240] env[61957]: DEBUG oslo_vmware.api [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Waiting for the task: (returnval){ [ 861.273240] env[61957]: value = "task-1277546" [ 861.273240] env[61957]: _type = "Task" [ 861.273240] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.273240] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd5c29d1-f2bd-4189-a5a5-1c7d5c27b64c {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.293074] env[61957]: DEBUG oslo_vmware.api [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Task: {'id': task-1277546, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.301055] env[61957]: DEBUG oslo_concurrency.lockutils [None req-95a860a5-ca00-4c71-b827-6e3af6a25042 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 861.324306] env[61957]: DEBUG oslo_concurrency.lockutils [None req-09e8eb69-1030-4228-8cb0-915d5cee1a52 tempest-AttachInterfacesV270Test-2032541732 tempest-AttachInterfacesV270Test-2032541732-project-member] Acquiring lock "refresh_cache-281b9644-0e7d-48b9-a7b6-45fd6102d558" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 861.324487] env[61957]: DEBUG oslo_concurrency.lockutils [None req-09e8eb69-1030-4228-8cb0-915d5cee1a52 tempest-AttachInterfacesV270Test-2032541732 tempest-AttachInterfacesV270Test-2032541732-project-member] Acquired lock "refresh_cache-281b9644-0e7d-48b9-a7b6-45fd6102d558" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 861.324615] env[61957]: DEBUG nova.network.neutron [None req-09e8eb69-1030-4228-8cb0-915d5cee1a52 tempest-AttachInterfacesV270Test-2032541732 tempest-AttachInterfacesV270Test-2032541732-project-member] [instance: 281b9644-0e7d-48b9-a7b6-45fd6102d558] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 861.531558] env[61957]: DEBUG oslo_concurrency.lockutils [req-519a0ffd-2b3b-4a40-875b-fcffde67f084 req-6e6a3808-8138-4d27-ae8b-52a5c50f2d9c service nova] Releasing lock "refresh_cache-c23141ee-0cbb-4d1b-8390-c3073fe354f1" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 861.548586] env[61957]: DEBUG oslo_vmware.api [None req-ac5c4b48-cfd7-420e-9aa1-225385350d11 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Task: {'id': task-1277545, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.451035} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.548901] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-ac5c4b48-cfd7-420e-9aa1-225385350d11 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore1] c23141ee-0cbb-4d1b-8390-c3073fe354f1/c23141ee-0cbb-4d1b-8390-c3073fe354f1.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 861.549149] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-ac5c4b48-cfd7-420e-9aa1-225385350d11 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] [instance: c23141ee-0cbb-4d1b-8390-c3073fe354f1] Extending root virtual disk to 1048576 {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 861.549418] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4a88aa79-46d2-4c77-a76b-e8b1c944d5ac {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.556773] env[61957]: DEBUG oslo_vmware.api [None req-ac5c4b48-cfd7-420e-9aa1-225385350d11 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Waiting for the task: (returnval){ [ 861.556773] env[61957]: value = "task-1277547" [ 861.556773] env[61957]: _type = "Task" [ 861.556773] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.565544] env[61957]: DEBUG oslo_vmware.api [None req-ac5c4b48-cfd7-420e-9aa1-225385350d11 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Task: {'id': task-1277547, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.725190] env[61957]: DEBUG nova.scheduler.client.report [None req-902199c7-5b38-4a0e-a88f-80539210637b tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 861.787252] env[61957]: DEBUG oslo_vmware.api [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Task: {'id': task-1277546, 'name': PowerOnVM_Task} progress is 1%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.805456] env[61957]: INFO nova.compute.manager [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: 6438fe40-046c-45d5-9986-8f182ecde49f] Took 36.25 seconds to build instance. [ 861.865784] env[61957]: DEBUG nova.network.neutron [None req-09e8eb69-1030-4228-8cb0-915d5cee1a52 tempest-AttachInterfacesV270Test-2032541732 tempest-AttachInterfacesV270Test-2032541732-project-member] [instance: 281b9644-0e7d-48b9-a7b6-45fd6102d558] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 862.053296] env[61957]: DEBUG nova.network.neutron [None req-09e8eb69-1030-4228-8cb0-915d5cee1a52 tempest-AttachInterfacesV270Test-2032541732 tempest-AttachInterfacesV270Test-2032541732-project-member] [instance: 281b9644-0e7d-48b9-a7b6-45fd6102d558] Updating instance_info_cache with network_info: [{"id": "22b42651-5480-45d9-b363-9106dd6a7ce6", "address": "fa:16:3e:ab:76:20", "network": {"id": "88b1e2da-32a9-41ef-8498-389712fe4f5a", "bridge": "br-int", "label": "tempest-AttachInterfacesV270Test-1957691659-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d3bc71403439465a8df6b74e6d6ce0e1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ea00b53a-9c9b-4592-ab95-7e10473f338d", "external-id": "nsx-vlan-transportzone-235", "segmentation_id": 235, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap22b42651-54", "ovs_interfaceid": "22b42651-5480-45d9-b363-9106dd6a7ce6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 862.066382] env[61957]: DEBUG oslo_vmware.api [None req-ac5c4b48-cfd7-420e-9aa1-225385350d11 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Task: {'id': task-1277547, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06706} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.067550] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-ac5c4b48-cfd7-420e-9aa1-225385350d11 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] [instance: c23141ee-0cbb-4d1b-8390-c3073fe354f1] Extended root virtual disk {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 862.067550] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fa81312-c967-4f7a-8d5e-4c27cbc181ef {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.091532] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-ac5c4b48-cfd7-420e-9aa1-225385350d11 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] [instance: c23141ee-0cbb-4d1b-8390-c3073fe354f1] Reconfiguring VM instance instance-0000003f to attach disk [datastore1] c23141ee-0cbb-4d1b-8390-c3073fe354f1/c23141ee-0cbb-4d1b-8390-c3073fe354f1.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 862.092185] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0409e495-43fd-47fc-86c6-7f12e56be29e {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.114012] env[61957]: DEBUG oslo_vmware.api [None req-ac5c4b48-cfd7-420e-9aa1-225385350d11 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Waiting for the task: (returnval){ [ 862.114012] env[61957]: value = "task-1277548" [ 862.114012] env[61957]: _type = "Task" [ 862.114012] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.123462] env[61957]: DEBUG oslo_vmware.api [None req-ac5c4b48-cfd7-420e-9aa1-225385350d11 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Task: {'id': task-1277548, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.231407] env[61957]: DEBUG oslo_concurrency.lockutils [None req-902199c7-5b38-4a0e-a88f-80539210637b tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.602s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 862.231956] env[61957]: DEBUG nova.compute.manager [None req-902199c7-5b38-4a0e-a88f-80539210637b tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] [instance: 526ff179-62a6-4763-ab25-797617c4ed57] Start building networks asynchronously for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 862.234865] env[61957]: DEBUG oslo_concurrency.lockutils [None req-c13ea078-3782-4688-93ee-408e1ac828c0 tempest-ServerTagsTestJSON-969152625 tempest-ServerTagsTestJSON-969152625-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.026s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 862.236385] env[61957]: INFO nova.compute.claims [None req-c13ea078-3782-4688-93ee-408e1ac828c0 tempest-ServerTagsTestJSON-969152625 tempest-ServerTagsTestJSON-969152625-project-member] [instance: 8e04a098-25de-4d57-9e3c-ea44d234d57e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 862.287411] env[61957]: DEBUG oslo_vmware.api [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Task: {'id': task-1277546, 'name': PowerOnVM_Task} progress is 91%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.311764] env[61957]: DEBUG oslo_concurrency.lockutils [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Lock "6438fe40-046c-45d5-9986-8f182ecde49f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 111.431s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 862.391558] env[61957]: DEBUG oslo_concurrency.lockutils [None req-a7399dbc-0983-4a91-82d1-9c1b173d4f53 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Acquiring lock "39b8b1df-efdd-4c22-9f31-85c85be4f0eb" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 862.391805] env[61957]: DEBUG oslo_concurrency.lockutils [None req-a7399dbc-0983-4a91-82d1-9c1b173d4f53 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Lock "39b8b1df-efdd-4c22-9f31-85c85be4f0eb" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 862.555864] env[61957]: DEBUG oslo_concurrency.lockutils [None req-09e8eb69-1030-4228-8cb0-915d5cee1a52 tempest-AttachInterfacesV270Test-2032541732 tempest-AttachInterfacesV270Test-2032541732-project-member] Releasing lock "refresh_cache-281b9644-0e7d-48b9-a7b6-45fd6102d558" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 862.556225] env[61957]: DEBUG nova.compute.manager [None req-09e8eb69-1030-4228-8cb0-915d5cee1a52 tempest-AttachInterfacesV270Test-2032541732 tempest-AttachInterfacesV270Test-2032541732-project-member] [instance: 281b9644-0e7d-48b9-a7b6-45fd6102d558] Instance network_info: |[{"id": "22b42651-5480-45d9-b363-9106dd6a7ce6", "address": "fa:16:3e:ab:76:20", "network": {"id": "88b1e2da-32a9-41ef-8498-389712fe4f5a", "bridge": "br-int", "label": "tempest-AttachInterfacesV270Test-1957691659-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d3bc71403439465a8df6b74e6d6ce0e1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ea00b53a-9c9b-4592-ab95-7e10473f338d", "external-id": "nsx-vlan-transportzone-235", "segmentation_id": 235, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap22b42651-54", "ovs_interfaceid": "22b42651-5480-45d9-b363-9106dd6a7ce6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 862.556687] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-09e8eb69-1030-4228-8cb0-915d5cee1a52 tempest-AttachInterfacesV270Test-2032541732 tempest-AttachInterfacesV270Test-2032541732-project-member] [instance: 281b9644-0e7d-48b9-a7b6-45fd6102d558] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ab:76:20', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ea00b53a-9c9b-4592-ab95-7e10473f338d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '22b42651-5480-45d9-b363-9106dd6a7ce6', 'vif_model': 'vmxnet3'}] {{(pid=61957) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 862.564170] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-09e8eb69-1030-4228-8cb0-915d5cee1a52 tempest-AttachInterfacesV270Test-2032541732 tempest-AttachInterfacesV270Test-2032541732-project-member] Creating folder: Project (d3bc71403439465a8df6b74e6d6ce0e1). Parent ref: group-v274445. {{(pid=61957) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 862.564757] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-51484476-59eb-42ba-bb4c-01bd295b7340 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.575812] env[61957]: INFO nova.virt.vmwareapi.vm_util [None req-09e8eb69-1030-4228-8cb0-915d5cee1a52 tempest-AttachInterfacesV270Test-2032541732 tempest-AttachInterfacesV270Test-2032541732-project-member] Created folder: Project (d3bc71403439465a8df6b74e6d6ce0e1) in parent group-v274445. [ 862.575812] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-09e8eb69-1030-4228-8cb0-915d5cee1a52 tempest-AttachInterfacesV270Test-2032541732 tempest-AttachInterfacesV270Test-2032541732-project-member] Creating folder: Instances. Parent ref: group-v274515. {{(pid=61957) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 862.575812] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e89a6ce4-b971-4783-b52d-4d74eeae13df {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.585536] env[61957]: INFO nova.virt.vmwareapi.vm_util [None req-09e8eb69-1030-4228-8cb0-915d5cee1a52 tempest-AttachInterfacesV270Test-2032541732 tempest-AttachInterfacesV270Test-2032541732-project-member] Created folder: Instances in parent group-v274515. [ 862.585791] env[61957]: DEBUG oslo.service.loopingcall [None req-09e8eb69-1030-4228-8cb0-915d5cee1a52 tempest-AttachInterfacesV270Test-2032541732 tempest-AttachInterfacesV270Test-2032541732-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 862.586269] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 281b9644-0e7d-48b9-a7b6-45fd6102d558] Creating VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 862.586269] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b1d5532f-8f9a-482a-b0cc-a4e73bf6c252 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.607171] env[61957]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 862.607171] env[61957]: value = "task-1277551" [ 862.607171] env[61957]: _type = "Task" [ 862.607171] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.615610] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277551, 'name': CreateVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.624611] env[61957]: DEBUG oslo_vmware.api [None req-ac5c4b48-cfd7-420e-9aa1-225385350d11 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Task: {'id': task-1277548, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.741476] env[61957]: DEBUG nova.compute.utils [None req-902199c7-5b38-4a0e-a88f-80539210637b tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Using /dev/sd instead of None {{(pid=61957) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 862.745262] env[61957]: DEBUG nova.compute.manager [None req-902199c7-5b38-4a0e-a88f-80539210637b tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] [instance: 526ff179-62a6-4763-ab25-797617c4ed57] Allocating IP information in the background. {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 862.745441] env[61957]: DEBUG nova.network.neutron [None req-902199c7-5b38-4a0e-a88f-80539210637b tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] [instance: 526ff179-62a6-4763-ab25-797617c4ed57] allocate_for_instance() {{(pid=61957) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 862.787283] env[61957]: DEBUG oslo_vmware.api [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Task: {'id': task-1277546, 'name': PowerOnVM_Task, 'duration_secs': 1.022845} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.787565] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: f66db265-887e-4d61-b848-c609e5c884cb] Powered on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 862.787728] env[61957]: INFO nova.compute.manager [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: f66db265-887e-4d61-b848-c609e5c884cb] Took 8.13 seconds to spawn the instance on the hypervisor. [ 862.787915] env[61957]: DEBUG nova.compute.manager [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: f66db265-887e-4d61-b848-c609e5c884cb] Checking state {{(pid=61957) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 862.788764] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0305e41-4bac-40a3-823d-79aae0eef85c {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.798758] env[61957]: DEBUG nova.policy [None req-902199c7-5b38-4a0e-a88f-80539210637b tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '965923b6f23744a484629c752b0d11d1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ea955bd4d84548ff8e6a0c21542d6ee7', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61957) authorize /opt/stack/nova/nova/policy.py:203}} [ 862.815099] env[61957]: DEBUG nova.compute.manager [None req-ca59f931-a2d6-4c9e-954c-a383ffdeea61 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] [instance: d5d10a31-0e78-4ed7-b944-9208138a4861] Starting instance... {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 863.071242] env[61957]: DEBUG nova.compute.manager [req-b66bae22-c89b-4efc-8cd2-574ef88a848e req-4a8ee172-dd53-46b7-81c3-21d930355acc service nova] [instance: 281b9644-0e7d-48b9-a7b6-45fd6102d558] Received event network-changed-22b42651-5480-45d9-b363-9106dd6a7ce6 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 863.071573] env[61957]: DEBUG nova.compute.manager [req-b66bae22-c89b-4efc-8cd2-574ef88a848e req-4a8ee172-dd53-46b7-81c3-21d930355acc service nova] [instance: 281b9644-0e7d-48b9-a7b6-45fd6102d558] Refreshing instance network info cache due to event network-changed-22b42651-5480-45d9-b363-9106dd6a7ce6. {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 863.071679] env[61957]: DEBUG oslo_concurrency.lockutils [req-b66bae22-c89b-4efc-8cd2-574ef88a848e req-4a8ee172-dd53-46b7-81c3-21d930355acc service nova] Acquiring lock "refresh_cache-281b9644-0e7d-48b9-a7b6-45fd6102d558" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 863.071906] env[61957]: DEBUG oslo_concurrency.lockutils [req-b66bae22-c89b-4efc-8cd2-574ef88a848e req-4a8ee172-dd53-46b7-81c3-21d930355acc service nova] Acquired lock "refresh_cache-281b9644-0e7d-48b9-a7b6-45fd6102d558" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 863.071987] env[61957]: DEBUG nova.network.neutron [req-b66bae22-c89b-4efc-8cd2-574ef88a848e req-4a8ee172-dd53-46b7-81c3-21d930355acc service nova] [instance: 281b9644-0e7d-48b9-a7b6-45fd6102d558] Refreshing network info cache for port 22b42651-5480-45d9-b363-9106dd6a7ce6 {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 863.118095] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277551, 'name': CreateVM_Task} progress is 99%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.125965] env[61957]: DEBUG oslo_vmware.api [None req-ac5c4b48-cfd7-420e-9aa1-225385350d11 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Task: {'id': task-1277548, 'name': ReconfigVM_Task, 'duration_secs': 0.735902} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.126092] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-ac5c4b48-cfd7-420e-9aa1-225385350d11 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] [instance: c23141ee-0cbb-4d1b-8390-c3073fe354f1] Reconfigured VM instance instance-0000003f to attach disk [datastore1] c23141ee-0cbb-4d1b-8390-c3073fe354f1/c23141ee-0cbb-4d1b-8390-c3073fe354f1.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 863.126754] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c1623649-f464-49a0-8712-473f4f1f92b8 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.132763] env[61957]: DEBUG oslo_vmware.api [None req-ac5c4b48-cfd7-420e-9aa1-225385350d11 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Waiting for the task: (returnval){ [ 863.132763] env[61957]: value = "task-1277552" [ 863.132763] env[61957]: _type = "Task" [ 863.132763] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.144795] env[61957]: DEBUG oslo_vmware.api [None req-ac5c4b48-cfd7-420e-9aa1-225385350d11 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Task: {'id': task-1277552, 'name': Rename_Task} progress is 6%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.232146] env[61957]: DEBUG nova.network.neutron [None req-902199c7-5b38-4a0e-a88f-80539210637b tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] [instance: 526ff179-62a6-4763-ab25-797617c4ed57] Successfully created port: 25abeea6-dd92-49e0-b014-d35dd1c0f8ee {{(pid=61957) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 863.246386] env[61957]: DEBUG nova.compute.manager [None req-902199c7-5b38-4a0e-a88f-80539210637b tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] [instance: 526ff179-62a6-4763-ab25-797617c4ed57] Start building block device mappings for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 863.314809] env[61957]: INFO nova.compute.manager [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: f66db265-887e-4d61-b848-c609e5c884cb] Took 34.16 seconds to build instance. [ 863.339333] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ca59f931-a2d6-4c9e-954c-a383ffdeea61 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 863.496456] env[61957]: DEBUG nova.compute.manager [None req-22255302-8924-4370-98de-304a0c0dd7ba tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: 0adae8e1-8c2f-4110-805b-1f286debc833] Checking state {{(pid=61957) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 863.497923] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebac7f6a-5bb9-40a9-b5e0-796f588e1edd {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.509803] env[61957]: DEBUG oslo_service.periodic_task [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61957) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 863.511113] env[61957]: DEBUG oslo_service.periodic_task [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61957) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 863.511113] env[61957]: DEBUG nova.compute.manager [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Starting heal instance info cache {{(pid=61957) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 863.511113] env[61957]: DEBUG nova.compute.manager [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Rebuilding the list of instances to heal {{(pid=61957) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9930}} [ 863.607761] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4e0efd6-fac0-41dd-a8fc-1428da8e186b {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.623677] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-708f6d82-caa1-4f85-bb9d-427cfe40ca08 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.630705] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277551, 'name': CreateVM_Task, 'duration_secs': 0.558225} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.630920] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 281b9644-0e7d-48b9-a7b6-45fd6102d558] Created VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 863.632037] env[61957]: DEBUG oslo_concurrency.lockutils [None req-09e8eb69-1030-4228-8cb0-915d5cee1a52 tempest-AttachInterfacesV270Test-2032541732 tempest-AttachInterfacesV270Test-2032541732-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 863.632037] env[61957]: DEBUG oslo_concurrency.lockutils [None req-09e8eb69-1030-4228-8cb0-915d5cee1a52 tempest-AttachInterfacesV270Test-2032541732 tempest-AttachInterfacesV270Test-2032541732-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 863.632183] env[61957]: DEBUG oslo_concurrency.lockutils [None req-09e8eb69-1030-4228-8cb0-915d5cee1a52 tempest-AttachInterfacesV270Test-2032541732 tempest-AttachInterfacesV270Test-2032541732-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 863.633767] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8448f6ac-ba6e-4c05-ac26-4ec0928ada22 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.661716] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3c9dd6c-24c2-46c7-a57d-6dd7da10b733 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.665672] env[61957]: DEBUG oslo_vmware.api [None req-09e8eb69-1030-4228-8cb0-915d5cee1a52 tempest-AttachInterfacesV270Test-2032541732 tempest-AttachInterfacesV270Test-2032541732-project-member] Waiting for the task: (returnval){ [ 863.665672] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]527cf6bd-afcc-08ef-ad5e-d857e20edd90" [ 863.665672] env[61957]: _type = "Task" [ 863.665672] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.674328] env[61957]: DEBUG oslo_vmware.api [None req-ac5c4b48-cfd7-420e-9aa1-225385350d11 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Task: {'id': task-1277552, 'name': Rename_Task, 'duration_secs': 0.247187} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.675091] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-ac5c4b48-cfd7-420e-9aa1-225385350d11 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] [instance: c23141ee-0cbb-4d1b-8390-c3073fe354f1] Powering on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 863.676324] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21cffaf7-c71f-46ba-99b9-79719f1a5443 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.683465] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-09e844e1-d726-4dfb-a882-705704d4c3e3 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.684968] env[61957]: DEBUG oslo_vmware.api [None req-09e8eb69-1030-4228-8cb0-915d5cee1a52 tempest-AttachInterfacesV270Test-2032541732 tempest-AttachInterfacesV270Test-2032541732-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]527cf6bd-afcc-08ef-ad5e-d857e20edd90, 'name': SearchDatastore_Task, 'duration_secs': 0.011135} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.687521] env[61957]: DEBUG oslo_concurrency.lockutils [None req-09e8eb69-1030-4228-8cb0-915d5cee1a52 tempest-AttachInterfacesV270Test-2032541732 tempest-AttachInterfacesV270Test-2032541732-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 863.687840] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-09e8eb69-1030-4228-8cb0-915d5cee1a52 tempest-AttachInterfacesV270Test-2032541732 tempest-AttachInterfacesV270Test-2032541732-project-member] [instance: 281b9644-0e7d-48b9-a7b6-45fd6102d558] Processing image 11c76a2c-f705-470a-ba9d-4657858bab38 {{(pid=61957) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 863.688128] env[61957]: DEBUG oslo_concurrency.lockutils [None req-09e8eb69-1030-4228-8cb0-915d5cee1a52 tempest-AttachInterfacesV270Test-2032541732 tempest-AttachInterfacesV270Test-2032541732-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 863.688291] env[61957]: DEBUG oslo_concurrency.lockutils [None req-09e8eb69-1030-4228-8cb0-915d5cee1a52 tempest-AttachInterfacesV270Test-2032541732 tempest-AttachInterfacesV270Test-2032541732-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 863.688477] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-09e8eb69-1030-4228-8cb0-915d5cee1a52 tempest-AttachInterfacesV270Test-2032541732 tempest-AttachInterfacesV270Test-2032541732-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 863.689098] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a2c703e9-88fc-43f4-bcbe-bda8d1a37a72 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.700766] env[61957]: DEBUG nova.compute.provider_tree [None req-c13ea078-3782-4688-93ee-408e1ac828c0 tempest-ServerTagsTestJSON-969152625 tempest-ServerTagsTestJSON-969152625-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 863.703755] env[61957]: DEBUG oslo_vmware.api [None req-ac5c4b48-cfd7-420e-9aa1-225385350d11 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Waiting for the task: (returnval){ [ 863.703755] env[61957]: value = "task-1277553" [ 863.703755] env[61957]: _type = "Task" [ 863.703755] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.715430] env[61957]: DEBUG oslo_vmware.api [None req-ac5c4b48-cfd7-420e-9aa1-225385350d11 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Task: {'id': task-1277553, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.715715] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-09e8eb69-1030-4228-8cb0-915d5cee1a52 tempest-AttachInterfacesV270Test-2032541732 tempest-AttachInterfacesV270Test-2032541732-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 863.715880] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-09e8eb69-1030-4228-8cb0-915d5cee1a52 tempest-AttachInterfacesV270Test-2032541732 tempest-AttachInterfacesV270Test-2032541732-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61957) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 863.716829] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-16c47f79-780f-42f0-92d6-3cb4eb080b57 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.722470] env[61957]: DEBUG oslo_vmware.api [None req-09e8eb69-1030-4228-8cb0-915d5cee1a52 tempest-AttachInterfacesV270Test-2032541732 tempest-AttachInterfacesV270Test-2032541732-project-member] Waiting for the task: (returnval){ [ 863.722470] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52144292-701d-bcc3-84ae-5eedab1691cb" [ 863.722470] env[61957]: _type = "Task" [ 863.722470] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.730877] env[61957]: DEBUG oslo_vmware.api [None req-09e8eb69-1030-4228-8cb0-915d5cee1a52 tempest-AttachInterfacesV270Test-2032541732 tempest-AttachInterfacesV270Test-2032541732-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52144292-701d-bcc3-84ae-5eedab1691cb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.817394] env[61957]: DEBUG oslo_concurrency.lockutils [None req-a93d2f7b-4844-4b35-b3df-b0f16046ca27 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Lock "f66db265-887e-4d61-b848-c609e5c884cb" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 112.908s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 863.923961] env[61957]: DEBUG nova.network.neutron [req-b66bae22-c89b-4efc-8cd2-574ef88a848e req-4a8ee172-dd53-46b7-81c3-21d930355acc service nova] [instance: 281b9644-0e7d-48b9-a7b6-45fd6102d558] Updated VIF entry in instance network info cache for port 22b42651-5480-45d9-b363-9106dd6a7ce6. {{(pid=61957) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 863.924372] env[61957]: DEBUG nova.network.neutron [req-b66bae22-c89b-4efc-8cd2-574ef88a848e req-4a8ee172-dd53-46b7-81c3-21d930355acc service nova] [instance: 281b9644-0e7d-48b9-a7b6-45fd6102d558] Updating instance_info_cache with network_info: [{"id": "22b42651-5480-45d9-b363-9106dd6a7ce6", "address": "fa:16:3e:ab:76:20", "network": {"id": "88b1e2da-32a9-41ef-8498-389712fe4f5a", "bridge": "br-int", "label": "tempest-AttachInterfacesV270Test-1957691659-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d3bc71403439465a8df6b74e6d6ce0e1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ea00b53a-9c9b-4592-ab95-7e10473f338d", "external-id": "nsx-vlan-transportzone-235", "segmentation_id": 235, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap22b42651-54", "ovs_interfaceid": "22b42651-5480-45d9-b363-9106dd6a7ce6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 864.012751] env[61957]: INFO nova.compute.manager [None req-22255302-8924-4370-98de-304a0c0dd7ba tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: 0adae8e1-8c2f-4110-805b-1f286debc833] instance snapshotting [ 864.013076] env[61957]: WARNING nova.compute.manager [None req-22255302-8924-4370-98de-304a0c0dd7ba tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: 0adae8e1-8c2f-4110-805b-1f286debc833] trying to snapshot a non-running instance: (state: 7 expected: 1) [ 864.017151] env[61957]: DEBUG nova.compute.manager [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] [instance: c23141ee-0cbb-4d1b-8390-c3073fe354f1] Skipping network cache update for instance because it is Building. {{(pid=61957) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 864.017332] env[61957]: DEBUG nova.compute.manager [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] [instance: 281b9644-0e7d-48b9-a7b6-45fd6102d558] Skipping network cache update for instance because it is Building. {{(pid=61957) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 864.017495] env[61957]: DEBUG nova.compute.manager [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] [instance: 526ff179-62a6-4763-ab25-797617c4ed57] Skipping network cache update for instance because it is Building. {{(pid=61957) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 864.017697] env[61957]: DEBUG nova.compute.manager [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] [instance: 8e04a098-25de-4d57-9e3c-ea44d234d57e] Skipping network cache update for instance because it is Building. {{(pid=61957) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 864.020674] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c54bdf0a-1397-41bc-8170-5e48fb106d9b {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.044049] env[61957]: DEBUG oslo_concurrency.lockutils [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Acquiring lock "refresh_cache-e0249936-d616-4ffb-8f77-d8107633c42a" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 864.044197] env[61957]: DEBUG oslo_concurrency.lockutils [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Acquired lock "refresh_cache-e0249936-d616-4ffb-8f77-d8107633c42a" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 864.044266] env[61957]: DEBUG nova.network.neutron [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] [instance: e0249936-d616-4ffb-8f77-d8107633c42a] Forcefully refreshing network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 864.044392] env[61957]: DEBUG nova.objects.instance [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Lazy-loading 'info_cache' on Instance uuid e0249936-d616-4ffb-8f77-d8107633c42a {{(pid=61957) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 864.046192] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76275d6b-bb5a-4b36-b085-3b74204303f6 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.205838] env[61957]: DEBUG nova.scheduler.client.report [None req-c13ea078-3782-4688-93ee-408e1ac828c0 tempest-ServerTagsTestJSON-969152625 tempest-ServerTagsTestJSON-969152625-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 864.219510] env[61957]: DEBUG oslo_vmware.api [None req-ac5c4b48-cfd7-420e-9aa1-225385350d11 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Task: {'id': task-1277553, 'name': PowerOnVM_Task} progress is 90%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.234195] env[61957]: DEBUG oslo_vmware.api [None req-09e8eb69-1030-4228-8cb0-915d5cee1a52 tempest-AttachInterfacesV270Test-2032541732 tempest-AttachInterfacesV270Test-2032541732-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52144292-701d-bcc3-84ae-5eedab1691cb, 'name': SearchDatastore_Task, 'duration_secs': 0.012749} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.234410] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d645ba30-72ee-485c-af5e-a2ad9efdbbc5 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.241504] env[61957]: DEBUG oslo_vmware.api [None req-09e8eb69-1030-4228-8cb0-915d5cee1a52 tempest-AttachInterfacesV270Test-2032541732 tempest-AttachInterfacesV270Test-2032541732-project-member] Waiting for the task: (returnval){ [ 864.241504] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52216fce-0ea1-d359-8f5e-6d2d3b556a8f" [ 864.241504] env[61957]: _type = "Task" [ 864.241504] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.251396] env[61957]: DEBUG oslo_vmware.api [None req-09e8eb69-1030-4228-8cb0-915d5cee1a52 tempest-AttachInterfacesV270Test-2032541732 tempest-AttachInterfacesV270Test-2032541732-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52216fce-0ea1-d359-8f5e-6d2d3b556a8f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.260929] env[61957]: DEBUG nova.compute.manager [None req-902199c7-5b38-4a0e-a88f-80539210637b tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] [instance: 526ff179-62a6-4763-ab25-797617c4ed57] Start spawning the instance on the hypervisor. {{(pid=61957) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 864.297458] env[61957]: DEBUG nova.virt.hardware [None req-902199c7-5b38-4a0e-a88f-80539210637b tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T17:22:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T17:21:50Z,direct_url=,disk_format='vmdk',id=11c76a2c-f705-470a-ba9d-4657858bab38,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='81d3096d0f094373913fc3dc8f5c3c6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T17:21:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 864.297746] env[61957]: DEBUG nova.virt.hardware [None req-902199c7-5b38-4a0e-a88f-80539210637b tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Flavor limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 864.297914] env[61957]: DEBUG nova.virt.hardware [None req-902199c7-5b38-4a0e-a88f-80539210637b tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Image limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 864.298150] env[61957]: DEBUG nova.virt.hardware [None req-902199c7-5b38-4a0e-a88f-80539210637b tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Flavor pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 864.298740] env[61957]: DEBUG nova.virt.hardware [None req-902199c7-5b38-4a0e-a88f-80539210637b tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Image pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 864.298740] env[61957]: DEBUG nova.virt.hardware [None req-902199c7-5b38-4a0e-a88f-80539210637b tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 864.298740] env[61957]: DEBUG nova.virt.hardware [None req-902199c7-5b38-4a0e-a88f-80539210637b tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 864.298869] env[61957]: DEBUG nova.virt.hardware [None req-902199c7-5b38-4a0e-a88f-80539210637b tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 864.299288] env[61957]: DEBUG nova.virt.hardware [None req-902199c7-5b38-4a0e-a88f-80539210637b tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Got 1 possible topologies {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 864.299288] env[61957]: DEBUG nova.virt.hardware [None req-902199c7-5b38-4a0e-a88f-80539210637b tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 864.299466] env[61957]: DEBUG nova.virt.hardware [None req-902199c7-5b38-4a0e-a88f-80539210637b tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 864.300773] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bf07c51-259f-4b29-8f7a-5e931149c94a {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.310982] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8d429ff-ca44-49fb-972c-2929dd0a9282 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.325856] env[61957]: DEBUG nova.compute.manager [None req-7916de58-8954-4a2d-b3f9-5ec9b9a16797 tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] [instance: 0d0f2d34-de35-4e80-8d9f-12693add0786] Starting instance... {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 864.426861] env[61957]: DEBUG oslo_concurrency.lockutils [req-b66bae22-c89b-4efc-8cd2-574ef88a848e req-4a8ee172-dd53-46b7-81c3-21d930355acc service nova] Releasing lock "refresh_cache-281b9644-0e7d-48b9-a7b6-45fd6102d558" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 864.557094] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-22255302-8924-4370-98de-304a0c0dd7ba tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: 0adae8e1-8c2f-4110-805b-1f286debc833] Creating Snapshot of the VM instance {{(pid=61957) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 864.557437] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-6b6a3428-6402-4ef6-81f6-c3db8c162ba7 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.564663] env[61957]: DEBUG oslo_vmware.api [None req-22255302-8924-4370-98de-304a0c0dd7ba tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Waiting for the task: (returnval){ [ 864.564663] env[61957]: value = "task-1277554" [ 864.564663] env[61957]: _type = "Task" [ 864.564663] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.573809] env[61957]: DEBUG oslo_vmware.api [None req-22255302-8924-4370-98de-304a0c0dd7ba tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': task-1277554, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.714051] env[61957]: DEBUG oslo_concurrency.lockutils [None req-c13ea078-3782-4688-93ee-408e1ac828c0 tempest-ServerTagsTestJSON-969152625 tempest-ServerTagsTestJSON-969152625-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.479s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 864.715082] env[61957]: DEBUG nova.compute.manager [None req-c13ea078-3782-4688-93ee-408e1ac828c0 tempest-ServerTagsTestJSON-969152625 tempest-ServerTagsTestJSON-969152625-project-member] [instance: 8e04a098-25de-4d57-9e3c-ea44d234d57e] Start building networks asynchronously for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 864.717711] env[61957]: DEBUG oslo_concurrency.lockutils [None req-85c79741-463e-4bc0-9ed2-4682b88fcc00 tempest-ServersV294TestFqdnHostnames-1426486100 tempest-ServersV294TestFqdnHostnames-1426486100-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.359s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 864.719254] env[61957]: INFO nova.compute.claims [None req-85c79741-463e-4bc0-9ed2-4682b88fcc00 tempest-ServersV294TestFqdnHostnames-1426486100 tempest-ServersV294TestFqdnHostnames-1426486100-project-member] [instance: 2e4a5344-600f-4b61-826e-c15f96b50af2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 864.731465] env[61957]: DEBUG oslo_vmware.api [None req-ac5c4b48-cfd7-420e-9aa1-225385350d11 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Task: {'id': task-1277553, 'name': PowerOnVM_Task, 'duration_secs': 0.562491} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.735243] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-ac5c4b48-cfd7-420e-9aa1-225385350d11 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] [instance: c23141ee-0cbb-4d1b-8390-c3073fe354f1] Powered on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 864.735369] env[61957]: INFO nova.compute.manager [None req-ac5c4b48-cfd7-420e-9aa1-225385350d11 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] [instance: c23141ee-0cbb-4d1b-8390-c3073fe354f1] Took 7.58 seconds to spawn the instance on the hypervisor. [ 864.735549] env[61957]: DEBUG nova.compute.manager [None req-ac5c4b48-cfd7-420e-9aa1-225385350d11 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] [instance: c23141ee-0cbb-4d1b-8390-c3073fe354f1] Checking state {{(pid=61957) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 864.736417] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-daaa58c5-bf00-4d83-8466-18a0716fd883 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.757442] env[61957]: DEBUG oslo_vmware.api [None req-09e8eb69-1030-4228-8cb0-915d5cee1a52 tempest-AttachInterfacesV270Test-2032541732 tempest-AttachInterfacesV270Test-2032541732-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52216fce-0ea1-d359-8f5e-6d2d3b556a8f, 'name': SearchDatastore_Task, 'duration_secs': 0.012246} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.757738] env[61957]: DEBUG oslo_concurrency.lockutils [None req-09e8eb69-1030-4228-8cb0-915d5cee1a52 tempest-AttachInterfacesV270Test-2032541732 tempest-AttachInterfacesV270Test-2032541732-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 864.758100] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-09e8eb69-1030-4228-8cb0-915d5cee1a52 tempest-AttachInterfacesV270Test-2032541732 tempest-AttachInterfacesV270Test-2032541732-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore2] 281b9644-0e7d-48b9-a7b6-45fd6102d558/281b9644-0e7d-48b9-a7b6-45fd6102d558.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 864.758401] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a58987e6-8132-4a7e-b18f-2bee9119063f {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.766472] env[61957]: DEBUG oslo_vmware.api [None req-09e8eb69-1030-4228-8cb0-915d5cee1a52 tempest-AttachInterfacesV270Test-2032541732 tempest-AttachInterfacesV270Test-2032541732-project-member] Waiting for the task: (returnval){ [ 864.766472] env[61957]: value = "task-1277555" [ 864.766472] env[61957]: _type = "Task" [ 864.766472] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.775595] env[61957]: DEBUG oslo_vmware.api [None req-09e8eb69-1030-4228-8cb0-915d5cee1a52 tempest-AttachInterfacesV270Test-2032541732 tempest-AttachInterfacesV270Test-2032541732-project-member] Task: {'id': task-1277555, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.853241] env[61957]: DEBUG oslo_concurrency.lockutils [None req-7916de58-8954-4a2d-b3f9-5ec9b9a16797 tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 864.945742] env[61957]: DEBUG oslo_concurrency.lockutils [None req-61ad082c-18ce-4de4-a90d-5e84d15388ff tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Acquiring lock "6438fe40-046c-45d5-9986-8f182ecde49f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 864.946432] env[61957]: DEBUG oslo_concurrency.lockutils [None req-61ad082c-18ce-4de4-a90d-5e84d15388ff tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Lock "6438fe40-046c-45d5-9986-8f182ecde49f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 864.946432] env[61957]: DEBUG oslo_concurrency.lockutils [None req-61ad082c-18ce-4de4-a90d-5e84d15388ff tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Acquiring lock "6438fe40-046c-45d5-9986-8f182ecde49f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 864.946562] env[61957]: DEBUG oslo_concurrency.lockutils [None req-61ad082c-18ce-4de4-a90d-5e84d15388ff tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Lock "6438fe40-046c-45d5-9986-8f182ecde49f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 864.946748] env[61957]: DEBUG oslo_concurrency.lockutils [None req-61ad082c-18ce-4de4-a90d-5e84d15388ff tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Lock "6438fe40-046c-45d5-9986-8f182ecde49f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 864.949815] env[61957]: INFO nova.compute.manager [None req-61ad082c-18ce-4de4-a90d-5e84d15388ff tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: 6438fe40-046c-45d5-9986-8f182ecde49f] Terminating instance [ 864.957020] env[61957]: DEBUG nova.compute.manager [None req-61ad082c-18ce-4de4-a90d-5e84d15388ff tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: 6438fe40-046c-45d5-9986-8f182ecde49f] Start destroying the instance on the hypervisor. {{(pid=61957) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 864.957020] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-61ad082c-18ce-4de4-a90d-5e84d15388ff tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: 6438fe40-046c-45d5-9986-8f182ecde49f] Destroying instance {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 864.957020] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56d264a4-9293-4d9c-8e56-062ef599dc35 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.968206] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-61ad082c-18ce-4de4-a90d-5e84d15388ff tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: 6438fe40-046c-45d5-9986-8f182ecde49f] Powering off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 864.968825] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-39f5aea6-910f-47fd-974f-5f6731a12eca {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.978578] env[61957]: DEBUG oslo_vmware.api [None req-61ad082c-18ce-4de4-a90d-5e84d15388ff tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Waiting for the task: (returnval){ [ 864.978578] env[61957]: value = "task-1277556" [ 864.978578] env[61957]: _type = "Task" [ 864.978578] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.995870] env[61957]: DEBUG oslo_vmware.api [None req-61ad082c-18ce-4de4-a90d-5e84d15388ff tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Task: {'id': task-1277556, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.998550] env[61957]: DEBUG nova.network.neutron [None req-902199c7-5b38-4a0e-a88f-80539210637b tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] [instance: 526ff179-62a6-4763-ab25-797617c4ed57] Successfully updated port: 25abeea6-dd92-49e0-b014-d35dd1c0f8ee {{(pid=61957) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 865.024047] env[61957]: DEBUG oslo_concurrency.lockutils [None req-99665f7b-964c-4087-97b8-943747701eca tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Acquiring lock "f66db265-887e-4d61-b848-c609e5c884cb" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 865.024336] env[61957]: DEBUG oslo_concurrency.lockutils [None req-99665f7b-964c-4087-97b8-943747701eca tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Lock "f66db265-887e-4d61-b848-c609e5c884cb" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 865.024546] env[61957]: DEBUG oslo_concurrency.lockutils [None req-99665f7b-964c-4087-97b8-943747701eca tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Acquiring lock "f66db265-887e-4d61-b848-c609e5c884cb-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 865.024732] env[61957]: DEBUG oslo_concurrency.lockutils [None req-99665f7b-964c-4087-97b8-943747701eca tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Lock "f66db265-887e-4d61-b848-c609e5c884cb-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 865.024904] env[61957]: DEBUG oslo_concurrency.lockutils [None req-99665f7b-964c-4087-97b8-943747701eca tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Lock "f66db265-887e-4d61-b848-c609e5c884cb-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 865.028118] env[61957]: INFO nova.compute.manager [None req-99665f7b-964c-4087-97b8-943747701eca tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: f66db265-887e-4d61-b848-c609e5c884cb] Terminating instance [ 865.030280] env[61957]: DEBUG nova.compute.manager [None req-99665f7b-964c-4087-97b8-943747701eca tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: f66db265-887e-4d61-b848-c609e5c884cb] Start destroying the instance on the hypervisor. {{(pid=61957) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 865.030531] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-99665f7b-964c-4087-97b8-943747701eca tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: f66db265-887e-4d61-b848-c609e5c884cb] Destroying instance {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 865.034202] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-690df590-6d89-4060-9fdb-4100ec6db3cb {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.040215] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-99665f7b-964c-4087-97b8-943747701eca tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: f66db265-887e-4d61-b848-c609e5c884cb] Powering off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 865.040580] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-34df5eef-3b99-4387-8e5c-68055c71a016 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.050919] env[61957]: DEBUG oslo_vmware.api [None req-99665f7b-964c-4087-97b8-943747701eca tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Waiting for the task: (returnval){ [ 865.050919] env[61957]: value = "task-1277557" [ 865.050919] env[61957]: _type = "Task" [ 865.050919] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.066964] env[61957]: DEBUG oslo_vmware.api [None req-99665f7b-964c-4087-97b8-943747701eca tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Task: {'id': task-1277557, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.076884] env[61957]: DEBUG oslo_vmware.api [None req-22255302-8924-4370-98de-304a0c0dd7ba tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': task-1277554, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.078336] env[61957]: DEBUG nova.network.neutron [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] [instance: e0249936-d616-4ffb-8f77-d8107633c42a] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 865.216952] env[61957]: DEBUG nova.compute.manager [req-951496d1-7e9b-453a-8415-224ba408dd0c req-f6af9023-805c-455c-9d4f-9c3dd1580bfc service nova] [instance: 526ff179-62a6-4763-ab25-797617c4ed57] Received event network-vif-plugged-25abeea6-dd92-49e0-b014-d35dd1c0f8ee {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 865.217322] env[61957]: DEBUG oslo_concurrency.lockutils [req-951496d1-7e9b-453a-8415-224ba408dd0c req-f6af9023-805c-455c-9d4f-9c3dd1580bfc service nova] Acquiring lock "526ff179-62a6-4763-ab25-797617c4ed57-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 865.217426] env[61957]: DEBUG oslo_concurrency.lockutils [req-951496d1-7e9b-453a-8415-224ba408dd0c req-f6af9023-805c-455c-9d4f-9c3dd1580bfc service nova] Lock "526ff179-62a6-4763-ab25-797617c4ed57-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 865.217640] env[61957]: DEBUG oslo_concurrency.lockutils [req-951496d1-7e9b-453a-8415-224ba408dd0c req-f6af9023-805c-455c-9d4f-9c3dd1580bfc service nova] Lock "526ff179-62a6-4763-ab25-797617c4ed57-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 865.217778] env[61957]: DEBUG nova.compute.manager [req-951496d1-7e9b-453a-8415-224ba408dd0c req-f6af9023-805c-455c-9d4f-9c3dd1580bfc service nova] [instance: 526ff179-62a6-4763-ab25-797617c4ed57] No waiting events found dispatching network-vif-plugged-25abeea6-dd92-49e0-b014-d35dd1c0f8ee {{(pid=61957) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 865.217924] env[61957]: WARNING nova.compute.manager [req-951496d1-7e9b-453a-8415-224ba408dd0c req-f6af9023-805c-455c-9d4f-9c3dd1580bfc service nova] [instance: 526ff179-62a6-4763-ab25-797617c4ed57] Received unexpected event network-vif-plugged-25abeea6-dd92-49e0-b014-d35dd1c0f8ee for instance with vm_state building and task_state spawning. [ 865.218563] env[61957]: DEBUG nova.compute.manager [req-951496d1-7e9b-453a-8415-224ba408dd0c req-f6af9023-805c-455c-9d4f-9c3dd1580bfc service nova] [instance: 526ff179-62a6-4763-ab25-797617c4ed57] Received event network-changed-25abeea6-dd92-49e0-b014-d35dd1c0f8ee {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 865.218980] env[61957]: DEBUG nova.compute.manager [req-951496d1-7e9b-453a-8415-224ba408dd0c req-f6af9023-805c-455c-9d4f-9c3dd1580bfc service nova] [instance: 526ff179-62a6-4763-ab25-797617c4ed57] Refreshing instance network info cache due to event network-changed-25abeea6-dd92-49e0-b014-d35dd1c0f8ee. {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 865.218980] env[61957]: DEBUG oslo_concurrency.lockutils [req-951496d1-7e9b-453a-8415-224ba408dd0c req-f6af9023-805c-455c-9d4f-9c3dd1580bfc service nova] Acquiring lock "refresh_cache-526ff179-62a6-4763-ab25-797617c4ed57" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 865.219310] env[61957]: DEBUG oslo_concurrency.lockutils [req-951496d1-7e9b-453a-8415-224ba408dd0c req-f6af9023-805c-455c-9d4f-9c3dd1580bfc service nova] Acquired lock "refresh_cache-526ff179-62a6-4763-ab25-797617c4ed57" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 865.219310] env[61957]: DEBUG nova.network.neutron [req-951496d1-7e9b-453a-8415-224ba408dd0c req-f6af9023-805c-455c-9d4f-9c3dd1580bfc service nova] [instance: 526ff179-62a6-4763-ab25-797617c4ed57] Refreshing network info cache for port 25abeea6-dd92-49e0-b014-d35dd1c0f8ee {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 865.224650] env[61957]: DEBUG nova.compute.utils [None req-c13ea078-3782-4688-93ee-408e1ac828c0 tempest-ServerTagsTestJSON-969152625 tempest-ServerTagsTestJSON-969152625-project-member] Using /dev/sd instead of None {{(pid=61957) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 865.228718] env[61957]: DEBUG nova.compute.manager [None req-c13ea078-3782-4688-93ee-408e1ac828c0 tempest-ServerTagsTestJSON-969152625 tempest-ServerTagsTestJSON-969152625-project-member] [instance: 8e04a098-25de-4d57-9e3c-ea44d234d57e] Allocating IP information in the background. {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 865.229030] env[61957]: DEBUG nova.network.neutron [None req-c13ea078-3782-4688-93ee-408e1ac828c0 tempest-ServerTagsTestJSON-969152625 tempest-ServerTagsTestJSON-969152625-project-member] [instance: 8e04a098-25de-4d57-9e3c-ea44d234d57e] allocate_for_instance() {{(pid=61957) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 865.264054] env[61957]: INFO nova.compute.manager [None req-ac5c4b48-cfd7-420e-9aa1-225385350d11 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] [instance: c23141ee-0cbb-4d1b-8390-c3073fe354f1] Took 34.48 seconds to build instance. [ 865.278019] env[61957]: DEBUG oslo_vmware.api [None req-09e8eb69-1030-4228-8cb0-915d5cee1a52 tempest-AttachInterfacesV270Test-2032541732 tempest-AttachInterfacesV270Test-2032541732-project-member] Task: {'id': task-1277555, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.48282} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.278471] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-09e8eb69-1030-4228-8cb0-915d5cee1a52 tempest-AttachInterfacesV270Test-2032541732 tempest-AttachInterfacesV270Test-2032541732-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore2] 281b9644-0e7d-48b9-a7b6-45fd6102d558/281b9644-0e7d-48b9-a7b6-45fd6102d558.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 865.278717] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-09e8eb69-1030-4228-8cb0-915d5cee1a52 tempest-AttachInterfacesV270Test-2032541732 tempest-AttachInterfacesV270Test-2032541732-project-member] [instance: 281b9644-0e7d-48b9-a7b6-45fd6102d558] Extending root virtual disk to 1048576 {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 865.278994] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b5435784-ea05-4c9d-be52-810786c1c566 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.287268] env[61957]: DEBUG oslo_vmware.api [None req-09e8eb69-1030-4228-8cb0-915d5cee1a52 tempest-AttachInterfacesV270Test-2032541732 tempest-AttachInterfacesV270Test-2032541732-project-member] Waiting for the task: (returnval){ [ 865.287268] env[61957]: value = "task-1277558" [ 865.287268] env[61957]: _type = "Task" [ 865.287268] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.293795] env[61957]: DEBUG nova.policy [None req-c13ea078-3782-4688-93ee-408e1ac828c0 tempest-ServerTagsTestJSON-969152625 tempest-ServerTagsTestJSON-969152625-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9f025854f256455196bfae74638a83db', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e8d71e68f10c4ee3a816f73cf9add0a5', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61957) authorize /opt/stack/nova/nova/policy.py:203}} [ 865.302269] env[61957]: DEBUG oslo_vmware.api [None req-09e8eb69-1030-4228-8cb0-915d5cee1a52 tempest-AttachInterfacesV270Test-2032541732 tempest-AttachInterfacesV270Test-2032541732-project-member] Task: {'id': task-1277558, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.492040] env[61957]: DEBUG oslo_vmware.api [None req-61ad082c-18ce-4de4-a90d-5e84d15388ff tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Task: {'id': task-1277556, 'name': PowerOffVM_Task, 'duration_secs': 0.263223} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.492622] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-61ad082c-18ce-4de4-a90d-5e84d15388ff tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: 6438fe40-046c-45d5-9986-8f182ecde49f] Powered off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 865.492895] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-61ad082c-18ce-4de4-a90d-5e84d15388ff tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: 6438fe40-046c-45d5-9986-8f182ecde49f] Unregistering the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 865.493817] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1e129294-96f1-4182-9e64-da97325c5cff {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.499602] env[61957]: DEBUG oslo_concurrency.lockutils [None req-902199c7-5b38-4a0e-a88f-80539210637b tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Acquiring lock "refresh_cache-526ff179-62a6-4763-ab25-797617c4ed57" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 865.563753] env[61957]: DEBUG oslo_vmware.api [None req-99665f7b-964c-4087-97b8-943747701eca tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Task: {'id': task-1277557, 'name': PowerOffVM_Task, 'duration_secs': 0.273737} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.564123] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-99665f7b-964c-4087-97b8-943747701eca tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: f66db265-887e-4d61-b848-c609e5c884cb] Powered off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 865.564687] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-99665f7b-964c-4087-97b8-943747701eca tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: f66db265-887e-4d61-b848-c609e5c884cb] Unregistering the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 865.564687] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-81f83b29-f9f5-44f6-8b42-733b0e0b35ad {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.571554] env[61957]: DEBUG nova.network.neutron [None req-c13ea078-3782-4688-93ee-408e1ac828c0 tempest-ServerTagsTestJSON-969152625 tempest-ServerTagsTestJSON-969152625-project-member] [instance: 8e04a098-25de-4d57-9e3c-ea44d234d57e] Successfully created port: 55344e12-1398-4752-a0f4-fdea985786bc {{(pid=61957) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 865.578459] env[61957]: DEBUG oslo_vmware.api [None req-22255302-8924-4370-98de-304a0c0dd7ba tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': task-1277554, 'name': CreateSnapshot_Task, 'duration_secs': 0.811938} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.579203] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-22255302-8924-4370-98de-304a0c0dd7ba tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: 0adae8e1-8c2f-4110-805b-1f286debc833] Created Snapshot of the VM instance {{(pid=61957) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 865.579726] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-61ad082c-18ce-4de4-a90d-5e84d15388ff tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: 6438fe40-046c-45d5-9986-8f182ecde49f] Unregistered the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 865.580063] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-61ad082c-18ce-4de4-a90d-5e84d15388ff tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: 6438fe40-046c-45d5-9986-8f182ecde49f] Deleting contents of the VM from datastore datastore2 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 865.580317] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-61ad082c-18ce-4de4-a90d-5e84d15388ff tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Deleting the datastore file [datastore2] 6438fe40-046c-45d5-9986-8f182ecde49f {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 865.581268] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d22d427-dfa6-4363-b0f3-57e64eef2e83 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.585499] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-71aa9fb3-8cab-499a-8008-840afda9a1b6 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.596810] env[61957]: DEBUG oslo_vmware.api [None req-61ad082c-18ce-4de4-a90d-5e84d15388ff tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Waiting for the task: (returnval){ [ 865.596810] env[61957]: value = "task-1277561" [ 865.596810] env[61957]: _type = "Task" [ 865.596810] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.604525] env[61957]: DEBUG oslo_vmware.api [None req-61ad082c-18ce-4de4-a90d-5e84d15388ff tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Task: {'id': task-1277561, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.662822] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-99665f7b-964c-4087-97b8-943747701eca tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: f66db265-887e-4d61-b848-c609e5c884cb] Unregistered the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 865.662822] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-99665f7b-964c-4087-97b8-943747701eca tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: f66db265-887e-4d61-b848-c609e5c884cb] Deleting contents of the VM from datastore datastore1 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 865.662822] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-99665f7b-964c-4087-97b8-943747701eca tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Deleting the datastore file [datastore1] f66db265-887e-4d61-b848-c609e5c884cb {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 865.665127] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6e8e7c5c-a007-4ba8-9fc8-91df1284f478 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.675615] env[61957]: DEBUG oslo_vmware.api [None req-99665f7b-964c-4087-97b8-943747701eca tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Waiting for the task: (returnval){ [ 865.675615] env[61957]: value = "task-1277562" [ 865.675615] env[61957]: _type = "Task" [ 865.675615] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.686075] env[61957]: DEBUG oslo_vmware.api [None req-99665f7b-964c-4087-97b8-943747701eca tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Task: {'id': task-1277562, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.725953] env[61957]: DEBUG nova.compute.manager [None req-c13ea078-3782-4688-93ee-408e1ac828c0 tempest-ServerTagsTestJSON-969152625 tempest-ServerTagsTestJSON-969152625-project-member] [instance: 8e04a098-25de-4d57-9e3c-ea44d234d57e] Start building block device mappings for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 865.754869] env[61957]: DEBUG nova.network.neutron [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] [instance: e0249936-d616-4ffb-8f77-d8107633c42a] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 865.767966] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ac5c4b48-cfd7-420e-9aa1-225385350d11 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Lock "c23141ee-0cbb-4d1b-8390-c3073fe354f1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 105.651s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 865.784775] env[61957]: DEBUG nova.network.neutron [req-951496d1-7e9b-453a-8415-224ba408dd0c req-f6af9023-805c-455c-9d4f-9c3dd1580bfc service nova] [instance: 526ff179-62a6-4763-ab25-797617c4ed57] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 865.801190] env[61957]: DEBUG oslo_vmware.api [None req-09e8eb69-1030-4228-8cb0-915d5cee1a52 tempest-AttachInterfacesV270Test-2032541732 tempest-AttachInterfacesV270Test-2032541732-project-member] Task: {'id': task-1277558, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072203} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.801518] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-09e8eb69-1030-4228-8cb0-915d5cee1a52 tempest-AttachInterfacesV270Test-2032541732 tempest-AttachInterfacesV270Test-2032541732-project-member] [instance: 281b9644-0e7d-48b9-a7b6-45fd6102d558] Extended root virtual disk {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 865.802511] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e4009ea-7e63-40ce-a10a-58898e46307a {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.825495] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-09e8eb69-1030-4228-8cb0-915d5cee1a52 tempest-AttachInterfacesV270Test-2032541732 tempest-AttachInterfacesV270Test-2032541732-project-member] [instance: 281b9644-0e7d-48b9-a7b6-45fd6102d558] Reconfiguring VM instance instance-00000040 to attach disk [datastore2] 281b9644-0e7d-48b9-a7b6-45fd6102d558/281b9644-0e7d-48b9-a7b6-45fd6102d558.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 865.828501] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7c7450ae-7029-40ed-b152-eb3a04503191 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.850740] env[61957]: DEBUG oslo_vmware.api [None req-09e8eb69-1030-4228-8cb0-915d5cee1a52 tempest-AttachInterfacesV270Test-2032541732 tempest-AttachInterfacesV270Test-2032541732-project-member] Waiting for the task: (returnval){ [ 865.850740] env[61957]: value = "task-1277563" [ 865.850740] env[61957]: _type = "Task" [ 865.850740] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.863860] env[61957]: DEBUG oslo_vmware.api [None req-09e8eb69-1030-4228-8cb0-915d5cee1a52 tempest-AttachInterfacesV270Test-2032541732 tempest-AttachInterfacesV270Test-2032541732-project-member] Task: {'id': task-1277563, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.978266] env[61957]: DEBUG nova.network.neutron [req-951496d1-7e9b-453a-8415-224ba408dd0c req-f6af9023-805c-455c-9d4f-9c3dd1580bfc service nova] [instance: 526ff179-62a6-4763-ab25-797617c4ed57] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 866.103467] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-22255302-8924-4370-98de-304a0c0dd7ba tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: 0adae8e1-8c2f-4110-805b-1f286debc833] Creating linked-clone VM from snapshot {{(pid=61957) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 866.107188] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-42af245f-723b-432a-aca1-d74ce30e76d3 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.118385] env[61957]: DEBUG oslo_vmware.api [None req-61ad082c-18ce-4de4-a90d-5e84d15388ff tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Task: {'id': task-1277561, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.178414} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.119812] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-61ad082c-18ce-4de4-a90d-5e84d15388ff tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Deleted the datastore file {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 866.121048] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-61ad082c-18ce-4de4-a90d-5e84d15388ff tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: 6438fe40-046c-45d5-9986-8f182ecde49f] Deleted contents of the VM from datastore datastore2 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 866.121048] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-61ad082c-18ce-4de4-a90d-5e84d15388ff tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: 6438fe40-046c-45d5-9986-8f182ecde49f] Instance destroyed {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 866.121048] env[61957]: INFO nova.compute.manager [None req-61ad082c-18ce-4de4-a90d-5e84d15388ff tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: 6438fe40-046c-45d5-9986-8f182ecde49f] Took 1.17 seconds to destroy the instance on the hypervisor. [ 866.121048] env[61957]: DEBUG oslo.service.loopingcall [None req-61ad082c-18ce-4de4-a90d-5e84d15388ff tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 866.121048] env[61957]: DEBUG oslo_vmware.api [None req-22255302-8924-4370-98de-304a0c0dd7ba tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Waiting for the task: (returnval){ [ 866.121048] env[61957]: value = "task-1277564" [ 866.121048] env[61957]: _type = "Task" [ 866.121048] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.124029] env[61957]: DEBUG nova.compute.manager [-] [instance: 6438fe40-046c-45d5-9986-8f182ecde49f] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 866.124029] env[61957]: DEBUG nova.network.neutron [-] [instance: 6438fe40-046c-45d5-9986-8f182ecde49f] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 866.134142] env[61957]: DEBUG oslo_vmware.api [None req-22255302-8924-4370-98de-304a0c0dd7ba tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': task-1277564, 'name': CloneVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.186763] env[61957]: DEBUG oslo_vmware.api [None req-99665f7b-964c-4087-97b8-943747701eca tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Task: {'id': task-1277562, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.167437} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.188591] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-99665f7b-964c-4087-97b8-943747701eca tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Deleted the datastore file {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 866.188854] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-99665f7b-964c-4087-97b8-943747701eca tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: f66db265-887e-4d61-b848-c609e5c884cb] Deleted contents of the VM from datastore datastore1 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 866.189054] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-99665f7b-964c-4087-97b8-943747701eca tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: f66db265-887e-4d61-b848-c609e5c884cb] Instance destroyed {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 866.189344] env[61957]: INFO nova.compute.manager [None req-99665f7b-964c-4087-97b8-943747701eca tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: f66db265-887e-4d61-b848-c609e5c884cb] Took 1.16 seconds to destroy the instance on the hypervisor. [ 866.189471] env[61957]: DEBUG oslo.service.loopingcall [None req-99665f7b-964c-4087-97b8-943747701eca tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 866.190234] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fba679d2-bc96-43e8-88c4-57587d942411 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.193178] env[61957]: DEBUG nova.compute.manager [-] [instance: f66db265-887e-4d61-b848-c609e5c884cb] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 866.193178] env[61957]: DEBUG nova.network.neutron [-] [instance: f66db265-887e-4d61-b848-c609e5c884cb] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 866.199301] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4ab0f81-c826-4a3e-9b52-a43e0fabac03 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.233567] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6590adf-8311-4c59-a3a8-6675fe60da03 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.245279] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6be7675-ff1a-48ad-98f6-8307e325b155 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.263265] env[61957]: DEBUG oslo_concurrency.lockutils [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Releasing lock "refresh_cache-e0249936-d616-4ffb-8f77-d8107633c42a" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 866.264066] env[61957]: DEBUG nova.compute.manager [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] [instance: e0249936-d616-4ffb-8f77-d8107633c42a] Updated the network info_cache for instance {{(pid=61957) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9997}} [ 866.264515] env[61957]: DEBUG nova.compute.provider_tree [None req-85c79741-463e-4bc0-9ed2-4682b88fcc00 tempest-ServersV294TestFqdnHostnames-1426486100 tempest-ServersV294TestFqdnHostnames-1426486100-project-member] Updating inventory in ProviderTree for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 866.266854] env[61957]: DEBUG oslo_service.periodic_task [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61957) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 866.270020] env[61957]: DEBUG oslo_service.periodic_task [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61957) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 866.270020] env[61957]: DEBUG oslo_service.periodic_task [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61957) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 866.270020] env[61957]: DEBUG oslo_service.periodic_task [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61957) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 866.270020] env[61957]: DEBUG oslo_service.periodic_task [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61957) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 866.270020] env[61957]: DEBUG oslo_service.periodic_task [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61957) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 866.270020] env[61957]: DEBUG nova.compute.manager [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61957) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 866.270020] env[61957]: DEBUG oslo_service.periodic_task [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61957) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 866.274399] env[61957]: DEBUG nova.compute.manager [None req-accdd3a1-4d57-48f0-a206-1602b908905b tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] [instance: a8ca1ad6-d636-4fa2-b0ac-53b020e392ce] Starting instance... {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 866.365802] env[61957]: DEBUG oslo_vmware.api [None req-09e8eb69-1030-4228-8cb0-915d5cee1a52 tempest-AttachInterfacesV270Test-2032541732 tempest-AttachInterfacesV270Test-2032541732-project-member] Task: {'id': task-1277563, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.482610] env[61957]: DEBUG oslo_concurrency.lockutils [req-951496d1-7e9b-453a-8415-224ba408dd0c req-f6af9023-805c-455c-9d4f-9c3dd1580bfc service nova] Releasing lock "refresh_cache-526ff179-62a6-4763-ab25-797617c4ed57" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 866.482711] env[61957]: DEBUG oslo_concurrency.lockutils [None req-902199c7-5b38-4a0e-a88f-80539210637b tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Acquired lock "refresh_cache-526ff179-62a6-4763-ab25-797617c4ed57" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 866.482851] env[61957]: DEBUG nova.network.neutron [None req-902199c7-5b38-4a0e-a88f-80539210637b tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] [instance: 526ff179-62a6-4763-ab25-797617c4ed57] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 866.635679] env[61957]: DEBUG oslo_vmware.api [None req-22255302-8924-4370-98de-304a0c0dd7ba tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': task-1277564, 'name': CloneVM_Task} progress is 94%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.742557] env[61957]: DEBUG nova.compute.manager [None req-c13ea078-3782-4688-93ee-408e1ac828c0 tempest-ServerTagsTestJSON-969152625 tempest-ServerTagsTestJSON-969152625-project-member] [instance: 8e04a098-25de-4d57-9e3c-ea44d234d57e] Start spawning the instance on the hypervisor. {{(pid=61957) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 866.774475] env[61957]: DEBUG oslo_concurrency.lockutils [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 866.781564] env[61957]: DEBUG nova.virt.hardware [None req-c13ea078-3782-4688-93ee-408e1ac828c0 tempest-ServerTagsTestJSON-969152625 tempest-ServerTagsTestJSON-969152625-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T17:22:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T17:21:50Z,direct_url=,disk_format='vmdk',id=11c76a2c-f705-470a-ba9d-4657858bab38,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='81d3096d0f094373913fc3dc8f5c3c6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T17:21:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 866.781874] env[61957]: DEBUG nova.virt.hardware [None req-c13ea078-3782-4688-93ee-408e1ac828c0 tempest-ServerTagsTestJSON-969152625 tempest-ServerTagsTestJSON-969152625-project-member] Flavor limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 866.782148] env[61957]: DEBUG nova.virt.hardware [None req-c13ea078-3782-4688-93ee-408e1ac828c0 tempest-ServerTagsTestJSON-969152625 tempest-ServerTagsTestJSON-969152625-project-member] Image limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 866.782324] env[61957]: DEBUG nova.virt.hardware [None req-c13ea078-3782-4688-93ee-408e1ac828c0 tempest-ServerTagsTestJSON-969152625 tempest-ServerTagsTestJSON-969152625-project-member] Flavor pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 866.782520] env[61957]: DEBUG nova.virt.hardware [None req-c13ea078-3782-4688-93ee-408e1ac828c0 tempest-ServerTagsTestJSON-969152625 tempest-ServerTagsTestJSON-969152625-project-member] Image pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 866.782698] env[61957]: DEBUG nova.virt.hardware [None req-c13ea078-3782-4688-93ee-408e1ac828c0 tempest-ServerTagsTestJSON-969152625 tempest-ServerTagsTestJSON-969152625-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 866.782915] env[61957]: DEBUG nova.virt.hardware [None req-c13ea078-3782-4688-93ee-408e1ac828c0 tempest-ServerTagsTestJSON-969152625 tempest-ServerTagsTestJSON-969152625-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 866.783132] env[61957]: DEBUG nova.virt.hardware [None req-c13ea078-3782-4688-93ee-408e1ac828c0 tempest-ServerTagsTestJSON-969152625 tempest-ServerTagsTestJSON-969152625-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 866.783357] env[61957]: DEBUG nova.virt.hardware [None req-c13ea078-3782-4688-93ee-408e1ac828c0 tempest-ServerTagsTestJSON-969152625 tempest-ServerTagsTestJSON-969152625-project-member] Got 1 possible topologies {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 866.783572] env[61957]: DEBUG nova.virt.hardware [None req-c13ea078-3782-4688-93ee-408e1ac828c0 tempest-ServerTagsTestJSON-969152625 tempest-ServerTagsTestJSON-969152625-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 866.783792] env[61957]: DEBUG nova.virt.hardware [None req-c13ea078-3782-4688-93ee-408e1ac828c0 tempest-ServerTagsTestJSON-969152625 tempest-ServerTagsTestJSON-969152625-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 866.789014] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d8fd23b-cb4b-402d-9595-d49f21cdd32f {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.794257] env[61957]: ERROR nova.scheduler.client.report [None req-85c79741-463e-4bc0-9ed2-4682b88fcc00 tempest-ServersV294TestFqdnHostnames-1426486100 tempest-ServersV294TestFqdnHostnames-1426486100-project-member] [req-6cd9fa3b-f487-427c-bb81-f2c5d8665b26] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 0ceb6c9e-61c6-496d-8579-9d32627e96da. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-6cd9fa3b-f487-427c-bb81-f2c5d8665b26"}]} [ 866.799870] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcf3d334-a154-4a65-8ee6-0db5d6e8eb8d {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.816164] env[61957]: DEBUG nova.scheduler.client.report [None req-85c79741-463e-4bc0-9ed2-4682b88fcc00 tempest-ServersV294TestFqdnHostnames-1426486100 tempest-ServersV294TestFqdnHostnames-1426486100-project-member] Refreshing inventories for resource provider 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 866.818675] env[61957]: DEBUG oslo_concurrency.lockutils [None req-accdd3a1-4d57-48f0-a206-1602b908905b tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 866.835201] env[61957]: DEBUG nova.scheduler.client.report [None req-85c79741-463e-4bc0-9ed2-4682b88fcc00 tempest-ServersV294TestFqdnHostnames-1426486100 tempest-ServersV294TestFqdnHostnames-1426486100-project-member] Updating ProviderTree inventory for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 866.835429] env[61957]: DEBUG nova.compute.provider_tree [None req-85c79741-463e-4bc0-9ed2-4682b88fcc00 tempest-ServersV294TestFqdnHostnames-1426486100 tempest-ServersV294TestFqdnHostnames-1426486100-project-member] Updating inventory in ProviderTree for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 866.852313] env[61957]: DEBUG nova.scheduler.client.report [None req-85c79741-463e-4bc0-9ed2-4682b88fcc00 tempest-ServersV294TestFqdnHostnames-1426486100 tempest-ServersV294TestFqdnHostnames-1426486100-project-member] Refreshing aggregate associations for resource provider 0ceb6c9e-61c6-496d-8579-9d32627e96da, aggregates: None {{(pid=61957) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 866.864729] env[61957]: DEBUG oslo_vmware.api [None req-09e8eb69-1030-4228-8cb0-915d5cee1a52 tempest-AttachInterfacesV270Test-2032541732 tempest-AttachInterfacesV270Test-2032541732-project-member] Task: {'id': task-1277563, 'name': ReconfigVM_Task, 'duration_secs': 0.758414} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.865169] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-09e8eb69-1030-4228-8cb0-915d5cee1a52 tempest-AttachInterfacesV270Test-2032541732 tempest-AttachInterfacesV270Test-2032541732-project-member] [instance: 281b9644-0e7d-48b9-a7b6-45fd6102d558] Reconfigured VM instance instance-00000040 to attach disk [datastore2] 281b9644-0e7d-48b9-a7b6-45fd6102d558/281b9644-0e7d-48b9-a7b6-45fd6102d558.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 866.866079] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a652e371-a90a-435f-aa41-c8e46d6aff7d {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.873058] env[61957]: DEBUG oslo_vmware.api [None req-09e8eb69-1030-4228-8cb0-915d5cee1a52 tempest-AttachInterfacesV270Test-2032541732 tempest-AttachInterfacesV270Test-2032541732-project-member] Waiting for the task: (returnval){ [ 866.873058] env[61957]: value = "task-1277565" [ 866.873058] env[61957]: _type = "Task" [ 866.873058] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.874071] env[61957]: DEBUG nova.scheduler.client.report [None req-85c79741-463e-4bc0-9ed2-4682b88fcc00 tempest-ServersV294TestFqdnHostnames-1426486100 tempest-ServersV294TestFqdnHostnames-1426486100-project-member] Refreshing trait associations for resource provider 0ceb6c9e-61c6-496d-8579-9d32627e96da, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO,HW_ARCH_X86_64,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=61957) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 866.885695] env[61957]: DEBUG oslo_vmware.api [None req-09e8eb69-1030-4228-8cb0-915d5cee1a52 tempest-AttachInterfacesV270Test-2032541732 tempest-AttachInterfacesV270Test-2032541732-project-member] Task: {'id': task-1277565, 'name': Rename_Task} progress is 6%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.070289] env[61957]: DEBUG nova.network.neutron [None req-902199c7-5b38-4a0e-a88f-80539210637b tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] [instance: 526ff179-62a6-4763-ab25-797617c4ed57] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 867.137772] env[61957]: DEBUG oslo_vmware.api [None req-22255302-8924-4370-98de-304a0c0dd7ba tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': task-1277564, 'name': CloneVM_Task} progress is 94%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.267951] env[61957]: DEBUG nova.compute.manager [req-f0c037a2-4476-4f2b-b120-bd7f1469fc85 req-944c8641-3d42-4bbe-80ff-3ed14452df0a service nova] [instance: 6438fe40-046c-45d5-9986-8f182ecde49f] Received event network-vif-deleted-e247dac1-1a7a-4b92-a2c0-66cfbe089af3 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 867.268117] env[61957]: INFO nova.compute.manager [req-f0c037a2-4476-4f2b-b120-bd7f1469fc85 req-944c8641-3d42-4bbe-80ff-3ed14452df0a service nova] [instance: 6438fe40-046c-45d5-9986-8f182ecde49f] Neutron deleted interface e247dac1-1a7a-4b92-a2c0-66cfbe089af3; detaching it from the instance and deleting it from the info cache [ 867.268302] env[61957]: DEBUG nova.network.neutron [req-f0c037a2-4476-4f2b-b120-bd7f1469fc85 req-944c8641-3d42-4bbe-80ff-3ed14452df0a service nova] [instance: 6438fe40-046c-45d5-9986-8f182ecde49f] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 867.291843] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-972e696f-6945-4208-8333-d00a956396e8 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.299929] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5722a77-2eaa-4270-8938-fdd869fe70a0 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.333692] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d697c4e-1b8d-4517-b1ea-d982256f4f0a {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.337187] env[61957]: DEBUG nova.network.neutron [None req-902199c7-5b38-4a0e-a88f-80539210637b tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] [instance: 526ff179-62a6-4763-ab25-797617c4ed57] Updating instance_info_cache with network_info: [{"id": "25abeea6-dd92-49e0-b014-d35dd1c0f8ee", "address": "fa:16:3e:0a:44:84", "network": {"id": "8e60ab13-6dd3-4cda-9cdf-974ad8bebe1f", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1404238496-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ea955bd4d84548ff8e6a0c21542d6ee7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9f208df-1fb5-4403-9796-7fd19e4bfb85", "external-id": "cl2-zone-400", "segmentation_id": 400, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap25abeea6-dd", "ovs_interfaceid": "25abeea6-dd92-49e0-b014-d35dd1c0f8ee", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 867.345988] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06401226-1bf3-4f9d-9e2b-8cc1efcbd251 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.359520] env[61957]: DEBUG nova.compute.provider_tree [None req-85c79741-463e-4bc0-9ed2-4682b88fcc00 tempest-ServersV294TestFqdnHostnames-1426486100 tempest-ServersV294TestFqdnHostnames-1426486100-project-member] Updating inventory in ProviderTree for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 867.392867] env[61957]: DEBUG oslo_vmware.api [None req-09e8eb69-1030-4228-8cb0-915d5cee1a52 tempest-AttachInterfacesV270Test-2032541732 tempest-AttachInterfacesV270Test-2032541732-project-member] Task: {'id': task-1277565, 'name': Rename_Task, 'duration_secs': 0.142697} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.392867] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-09e8eb69-1030-4228-8cb0-915d5cee1a52 tempest-AttachInterfacesV270Test-2032541732 tempest-AttachInterfacesV270Test-2032541732-project-member] [instance: 281b9644-0e7d-48b9-a7b6-45fd6102d558] Powering on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 867.392867] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-29b4d098-32e6-461b-97e5-42defac28948 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.398749] env[61957]: DEBUG oslo_vmware.api [None req-09e8eb69-1030-4228-8cb0-915d5cee1a52 tempest-AttachInterfacesV270Test-2032541732 tempest-AttachInterfacesV270Test-2032541732-project-member] Waiting for the task: (returnval){ [ 867.398749] env[61957]: value = "task-1277566" [ 867.398749] env[61957]: _type = "Task" [ 867.398749] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.412315] env[61957]: DEBUG oslo_vmware.api [None req-09e8eb69-1030-4228-8cb0-915d5cee1a52 tempest-AttachInterfacesV270Test-2032541732 tempest-AttachInterfacesV270Test-2032541732-project-member] Task: {'id': task-1277566, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.563917] env[61957]: DEBUG nova.network.neutron [None req-c13ea078-3782-4688-93ee-408e1ac828c0 tempest-ServerTagsTestJSON-969152625 tempest-ServerTagsTestJSON-969152625-project-member] [instance: 8e04a098-25de-4d57-9e3c-ea44d234d57e] Successfully updated port: 55344e12-1398-4752-a0f4-fdea985786bc {{(pid=61957) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 867.637298] env[61957]: DEBUG oslo_vmware.api [None req-22255302-8924-4370-98de-304a0c0dd7ba tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': task-1277564, 'name': CloneVM_Task} progress is 100%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.642018] env[61957]: DEBUG nova.network.neutron [-] [instance: 6438fe40-046c-45d5-9986-8f182ecde49f] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 867.650270] env[61957]: DEBUG nova.network.neutron [-] [instance: f66db265-887e-4d61-b848-c609e5c884cb] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 867.774338] env[61957]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7bcdcb37-e39c-414e-9a0f-79c0cc18c33e {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.784297] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c69cfd8-f3b3-4111-b8af-6389445b56d5 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.810651] env[61957]: DEBUG nova.compute.manager [req-f0c037a2-4476-4f2b-b120-bd7f1469fc85 req-944c8641-3d42-4bbe-80ff-3ed14452df0a service nova] [instance: 6438fe40-046c-45d5-9986-8f182ecde49f] Detach interface failed, port_id=e247dac1-1a7a-4b92-a2c0-66cfbe089af3, reason: Instance 6438fe40-046c-45d5-9986-8f182ecde49f could not be found. {{(pid=61957) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 867.810905] env[61957]: DEBUG nova.compute.manager [req-f0c037a2-4476-4f2b-b120-bd7f1469fc85 req-944c8641-3d42-4bbe-80ff-3ed14452df0a service nova] [instance: f66db265-887e-4d61-b848-c609e5c884cb] Received event network-vif-deleted-0289d1f8-a41a-44e5-9b66-6f92ce51debf {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 867.811105] env[61957]: INFO nova.compute.manager [req-f0c037a2-4476-4f2b-b120-bd7f1469fc85 req-944c8641-3d42-4bbe-80ff-3ed14452df0a service nova] [instance: f66db265-887e-4d61-b848-c609e5c884cb] Neutron deleted interface 0289d1f8-a41a-44e5-9b66-6f92ce51debf; detaching it from the instance and deleting it from the info cache [ 867.811603] env[61957]: DEBUG nova.network.neutron [req-f0c037a2-4476-4f2b-b120-bd7f1469fc85 req-944c8641-3d42-4bbe-80ff-3ed14452df0a service nova] [instance: f66db265-887e-4d61-b848-c609e5c884cb] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 867.843014] env[61957]: DEBUG oslo_concurrency.lockutils [None req-902199c7-5b38-4a0e-a88f-80539210637b tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Releasing lock "refresh_cache-526ff179-62a6-4763-ab25-797617c4ed57" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 867.843014] env[61957]: DEBUG nova.compute.manager [None req-902199c7-5b38-4a0e-a88f-80539210637b tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] [instance: 526ff179-62a6-4763-ab25-797617c4ed57] Instance network_info: |[{"id": "25abeea6-dd92-49e0-b014-d35dd1c0f8ee", "address": "fa:16:3e:0a:44:84", "network": {"id": "8e60ab13-6dd3-4cda-9cdf-974ad8bebe1f", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1404238496-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ea955bd4d84548ff8e6a0c21542d6ee7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9f208df-1fb5-4403-9796-7fd19e4bfb85", "external-id": "cl2-zone-400", "segmentation_id": 400, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap25abeea6-dd", "ovs_interfaceid": "25abeea6-dd92-49e0-b014-d35dd1c0f8ee", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 867.843014] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-902199c7-5b38-4a0e-a88f-80539210637b tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] [instance: 526ff179-62a6-4763-ab25-797617c4ed57] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:0a:44:84', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c9f208df-1fb5-4403-9796-7fd19e4bfb85', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '25abeea6-dd92-49e0-b014-d35dd1c0f8ee', 'vif_model': 'vmxnet3'}] {{(pid=61957) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 867.848789] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-902199c7-5b38-4a0e-a88f-80539210637b tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Creating folder: Project (ea955bd4d84548ff8e6a0c21542d6ee7). Parent ref: group-v274445. {{(pid=61957) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 867.849248] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f4bafcac-1bff-446b-82ca-fa3589d0ec1e {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.862141] env[61957]: INFO nova.virt.vmwareapi.vm_util [None req-902199c7-5b38-4a0e-a88f-80539210637b tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Created folder: Project (ea955bd4d84548ff8e6a0c21542d6ee7) in parent group-v274445. [ 867.862141] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-902199c7-5b38-4a0e-a88f-80539210637b tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Creating folder: Instances. Parent ref: group-v274520. {{(pid=61957) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 867.862141] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-229fbecb-0ea1-4681-9e32-2b4a9baa85d8 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.883021] env[61957]: INFO nova.virt.vmwareapi.vm_util [None req-902199c7-5b38-4a0e-a88f-80539210637b tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Created folder: Instances in parent group-v274520. [ 867.883021] env[61957]: DEBUG oslo.service.loopingcall [None req-902199c7-5b38-4a0e-a88f-80539210637b tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 867.883021] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 526ff179-62a6-4763-ab25-797617c4ed57] Creating VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 867.883021] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b2982180-57aa-4c96-aba5-1bb80424d06a {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.907070] env[61957]: DEBUG nova.scheduler.client.report [None req-85c79741-463e-4bc0-9ed2-4682b88fcc00 tempest-ServersV294TestFqdnHostnames-1426486100 tempest-ServersV294TestFqdnHostnames-1426486100-project-member] Updated inventory for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da with generation 84 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 867.907583] env[61957]: DEBUG nova.compute.provider_tree [None req-85c79741-463e-4bc0-9ed2-4682b88fcc00 tempest-ServersV294TestFqdnHostnames-1426486100 tempest-ServersV294TestFqdnHostnames-1426486100-project-member] Updating resource provider 0ceb6c9e-61c6-496d-8579-9d32627e96da generation from 84 to 85 during operation: update_inventory {{(pid=61957) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 867.908087] env[61957]: DEBUG nova.compute.provider_tree [None req-85c79741-463e-4bc0-9ed2-4682b88fcc00 tempest-ServersV294TestFqdnHostnames-1426486100 tempest-ServersV294TestFqdnHostnames-1426486100-project-member] Updating inventory in ProviderTree for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 867.912366] env[61957]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 867.912366] env[61957]: value = "task-1277569" [ 867.912366] env[61957]: _type = "Task" [ 867.912366] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.921197] env[61957]: DEBUG oslo_vmware.api [None req-09e8eb69-1030-4228-8cb0-915d5cee1a52 tempest-AttachInterfacesV270Test-2032541732 tempest-AttachInterfacesV270Test-2032541732-project-member] Task: {'id': task-1277566, 'name': PowerOnVM_Task, 'duration_secs': 0.49523} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.922281] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-09e8eb69-1030-4228-8cb0-915d5cee1a52 tempest-AttachInterfacesV270Test-2032541732 tempest-AttachInterfacesV270Test-2032541732-project-member] [instance: 281b9644-0e7d-48b9-a7b6-45fd6102d558] Powered on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 867.922281] env[61957]: INFO nova.compute.manager [None req-09e8eb69-1030-4228-8cb0-915d5cee1a52 tempest-AttachInterfacesV270Test-2032541732 tempest-AttachInterfacesV270Test-2032541732-project-member] [instance: 281b9644-0e7d-48b9-a7b6-45fd6102d558] Took 8.30 seconds to spawn the instance on the hypervisor. [ 867.923156] env[61957]: DEBUG nova.compute.manager [None req-09e8eb69-1030-4228-8cb0-915d5cee1a52 tempest-AttachInterfacesV270Test-2032541732 tempest-AttachInterfacesV270Test-2032541732-project-member] [instance: 281b9644-0e7d-48b9-a7b6-45fd6102d558] Checking state {{(pid=61957) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 867.926852] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2fa09bd-5d94-43cf-ac35-6104b45fb3b3 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.929627] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277569, 'name': CreateVM_Task} progress is 6%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.066480] env[61957]: DEBUG oslo_concurrency.lockutils [None req-c13ea078-3782-4688-93ee-408e1ac828c0 tempest-ServerTagsTestJSON-969152625 tempest-ServerTagsTestJSON-969152625-project-member] Acquiring lock "refresh_cache-8e04a098-25de-4d57-9e3c-ea44d234d57e" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 868.066684] env[61957]: DEBUG oslo_concurrency.lockutils [None req-c13ea078-3782-4688-93ee-408e1ac828c0 tempest-ServerTagsTestJSON-969152625 tempest-ServerTagsTestJSON-969152625-project-member] Acquired lock "refresh_cache-8e04a098-25de-4d57-9e3c-ea44d234d57e" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 868.067443] env[61957]: DEBUG nova.network.neutron [None req-c13ea078-3782-4688-93ee-408e1ac828c0 tempest-ServerTagsTestJSON-969152625 tempest-ServerTagsTestJSON-969152625-project-member] [instance: 8e04a098-25de-4d57-9e3c-ea44d234d57e] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 868.142073] env[61957]: INFO nova.compute.manager [-] [instance: 6438fe40-046c-45d5-9986-8f182ecde49f] Took 2.02 seconds to deallocate network for instance. [ 868.142427] env[61957]: DEBUG oslo_vmware.api [None req-22255302-8924-4370-98de-304a0c0dd7ba tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': task-1277564, 'name': CloneVM_Task, 'duration_secs': 1.558392} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.145762] env[61957]: INFO nova.virt.vmwareapi.vmops [None req-22255302-8924-4370-98de-304a0c0dd7ba tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: 0adae8e1-8c2f-4110-805b-1f286debc833] Created linked-clone VM from snapshot [ 868.150952] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47c05eab-bbbf-4918-84ce-c73e6f94f3dd {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.154169] env[61957]: INFO nova.compute.manager [-] [instance: f66db265-887e-4d61-b848-c609e5c884cb] Took 1.96 seconds to deallocate network for instance. [ 868.163939] env[61957]: DEBUG nova.virt.vmwareapi.images [None req-22255302-8924-4370-98de-304a0c0dd7ba tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: 0adae8e1-8c2f-4110-805b-1f286debc833] Uploading image 64157260-64c5-4eb6-9410-9be776e1d28d {{(pid=61957) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 868.201998] env[61957]: DEBUG oslo_vmware.rw_handles [None req-22255302-8924-4370-98de-304a0c0dd7ba tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 868.201998] env[61957]: value = "vm-274519" [ 868.201998] env[61957]: _type = "VirtualMachine" [ 868.201998] env[61957]: }. {{(pid=61957) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 868.201998] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-21e84b2e-c7f9-4938-b163-0edd97de2ffb {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.210253] env[61957]: DEBUG oslo_vmware.rw_handles [None req-22255302-8924-4370-98de-304a0c0dd7ba tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Lease: (returnval){ [ 868.210253] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]525d8446-06a6-3db7-1cdc-d57a8ee7d764" [ 868.210253] env[61957]: _type = "HttpNfcLease" [ 868.210253] env[61957]: } obtained for exporting VM: (result){ [ 868.210253] env[61957]: value = "vm-274519" [ 868.210253] env[61957]: _type = "VirtualMachine" [ 868.210253] env[61957]: }. {{(pid=61957) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 868.210537] env[61957]: DEBUG oslo_vmware.api [None req-22255302-8924-4370-98de-304a0c0dd7ba tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Waiting for the lease: (returnval){ [ 868.210537] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]525d8446-06a6-3db7-1cdc-d57a8ee7d764" [ 868.210537] env[61957]: _type = "HttpNfcLease" [ 868.210537] env[61957]: } to be ready. {{(pid=61957) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 868.219525] env[61957]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 868.219525] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]525d8446-06a6-3db7-1cdc-d57a8ee7d764" [ 868.219525] env[61957]: _type = "HttpNfcLease" [ 868.219525] env[61957]: } is initializing. {{(pid=61957) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 868.313802] env[61957]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ecd05a75-6e12-43d1-bc10-34cd7128613f {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.323836] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-022fee91-d285-4499-aee3-5fcac8546a9f {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.353055] env[61957]: DEBUG nova.compute.manager [req-f0c037a2-4476-4f2b-b120-bd7f1469fc85 req-944c8641-3d42-4bbe-80ff-3ed14452df0a service nova] [instance: f66db265-887e-4d61-b848-c609e5c884cb] Detach interface failed, port_id=0289d1f8-a41a-44e5-9b66-6f92ce51debf, reason: Instance f66db265-887e-4d61-b848-c609e5c884cb could not be found. {{(pid=61957) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 868.353322] env[61957]: DEBUG nova.compute.manager [req-f0c037a2-4476-4f2b-b120-bd7f1469fc85 req-944c8641-3d42-4bbe-80ff-3ed14452df0a service nova] [instance: c23141ee-0cbb-4d1b-8390-c3073fe354f1] Received event network-changed-a9884dda-2fb3-4e26-84b1-877dd0e0459c {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 868.353497] env[61957]: DEBUG nova.compute.manager [req-f0c037a2-4476-4f2b-b120-bd7f1469fc85 req-944c8641-3d42-4bbe-80ff-3ed14452df0a service nova] [instance: c23141ee-0cbb-4d1b-8390-c3073fe354f1] Refreshing instance network info cache due to event network-changed-a9884dda-2fb3-4e26-84b1-877dd0e0459c. {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 868.356784] env[61957]: DEBUG oslo_concurrency.lockutils [req-f0c037a2-4476-4f2b-b120-bd7f1469fc85 req-944c8641-3d42-4bbe-80ff-3ed14452df0a service nova] Acquiring lock "refresh_cache-c23141ee-0cbb-4d1b-8390-c3073fe354f1" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 868.356784] env[61957]: DEBUG oslo_concurrency.lockutils [req-f0c037a2-4476-4f2b-b120-bd7f1469fc85 req-944c8641-3d42-4bbe-80ff-3ed14452df0a service nova] Acquired lock "refresh_cache-c23141ee-0cbb-4d1b-8390-c3073fe354f1" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 868.356784] env[61957]: DEBUG nova.network.neutron [req-f0c037a2-4476-4f2b-b120-bd7f1469fc85 req-944c8641-3d42-4bbe-80ff-3ed14452df0a service nova] [instance: c23141ee-0cbb-4d1b-8390-c3073fe354f1] Refreshing network info cache for port a9884dda-2fb3-4e26-84b1-877dd0e0459c {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 868.416455] env[61957]: DEBUG oslo_concurrency.lockutils [None req-85c79741-463e-4bc0-9ed2-4682b88fcc00 tempest-ServersV294TestFqdnHostnames-1426486100 tempest-ServersV294TestFqdnHostnames-1426486100-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.696s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 868.416455] env[61957]: DEBUG nova.compute.manager [None req-85c79741-463e-4bc0-9ed2-4682b88fcc00 tempest-ServersV294TestFqdnHostnames-1426486100 tempest-ServersV294TestFqdnHostnames-1426486100-project-member] [instance: 2e4a5344-600f-4b61-826e-c15f96b50af2] Start building networks asynchronously for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 868.419063] env[61957]: DEBUG oslo_concurrency.lockutils [None req-82a5645a-fb19-4ad1-aa47-e12b7f048b77 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 24.137s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 868.419343] env[61957]: DEBUG nova.objects.instance [None req-82a5645a-fb19-4ad1-aa47-e12b7f048b77 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] [instance: 17f00b20-9d3b-45e6-919d-6fab9999ec77] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61957) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 868.435428] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277569, 'name': CreateVM_Task, 'duration_secs': 0.433988} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.435721] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 526ff179-62a6-4763-ab25-797617c4ed57] Created VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 868.436407] env[61957]: DEBUG oslo_concurrency.lockutils [None req-902199c7-5b38-4a0e-a88f-80539210637b tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 868.436629] env[61957]: DEBUG oslo_concurrency.lockutils [None req-902199c7-5b38-4a0e-a88f-80539210637b tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 868.437060] env[61957]: DEBUG oslo_concurrency.lockutils [None req-902199c7-5b38-4a0e-a88f-80539210637b tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 868.438120] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6de82ec4-14eb-4e57-b3bc-5cf614d28d49 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.448830] env[61957]: DEBUG oslo_vmware.api [None req-902199c7-5b38-4a0e-a88f-80539210637b tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Waiting for the task: (returnval){ [ 868.448830] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52a774bf-22f6-09cc-3b9a-f74ee3a8f8f9" [ 868.448830] env[61957]: _type = "Task" [ 868.448830] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.454601] env[61957]: INFO nova.compute.manager [None req-09e8eb69-1030-4228-8cb0-915d5cee1a52 tempest-AttachInterfacesV270Test-2032541732 tempest-AttachInterfacesV270Test-2032541732-project-member] [instance: 281b9644-0e7d-48b9-a7b6-45fd6102d558] Took 36.10 seconds to build instance. [ 868.461901] env[61957]: DEBUG oslo_vmware.api [None req-902199c7-5b38-4a0e-a88f-80539210637b tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52a774bf-22f6-09cc-3b9a-f74ee3a8f8f9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.656484] env[61957]: DEBUG oslo_concurrency.lockutils [None req-61ad082c-18ce-4de4-a90d-5e84d15388ff tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 868.670418] env[61957]: DEBUG oslo_concurrency.lockutils [None req-99665f7b-964c-4087-97b8-943747701eca tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 868.722678] env[61957]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 868.722678] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]525d8446-06a6-3db7-1cdc-d57a8ee7d764" [ 868.722678] env[61957]: _type = "HttpNfcLease" [ 868.722678] env[61957]: } is ready. {{(pid=61957) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 868.722678] env[61957]: DEBUG oslo_vmware.rw_handles [None req-22255302-8924-4370-98de-304a0c0dd7ba tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 868.722678] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]525d8446-06a6-3db7-1cdc-d57a8ee7d764" [ 868.722678] env[61957]: _type = "HttpNfcLease" [ 868.722678] env[61957]: }. {{(pid=61957) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 868.724289] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-047df7a4-75f5-4c94-9128-673b70f320d9 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.731888] env[61957]: DEBUG oslo_vmware.rw_handles [None req-22255302-8924-4370-98de-304a0c0dd7ba tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5210ccde-3f01-e031-c858-b8a459135f72/disk-0.vmdk from lease info. {{(pid=61957) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 868.732366] env[61957]: DEBUG oslo_vmware.rw_handles [None req-22255302-8924-4370-98de-304a0c0dd7ba tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5210ccde-3f01-e031-c858-b8a459135f72/disk-0.vmdk for reading. {{(pid=61957) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 868.734598] env[61957]: DEBUG nova.network.neutron [None req-c13ea078-3782-4688-93ee-408e1ac828c0 tempest-ServerTagsTestJSON-969152625 tempest-ServerTagsTestJSON-969152625-project-member] [instance: 8e04a098-25de-4d57-9e3c-ea44d234d57e] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 868.932860] env[61957]: DEBUG nova.compute.utils [None req-85c79741-463e-4bc0-9ed2-4682b88fcc00 tempest-ServersV294TestFqdnHostnames-1426486100 tempest-ServersV294TestFqdnHostnames-1426486100-project-member] Using /dev/sd instead of None {{(pid=61957) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 868.934455] env[61957]: DEBUG nova.compute.manager [None req-85c79741-463e-4bc0-9ed2-4682b88fcc00 tempest-ServersV294TestFqdnHostnames-1426486100 tempest-ServersV294TestFqdnHostnames-1426486100-project-member] [instance: 2e4a5344-600f-4b61-826e-c15f96b50af2] Allocating IP information in the background. {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 868.934455] env[61957]: DEBUG nova.network.neutron [None req-85c79741-463e-4bc0-9ed2-4682b88fcc00 tempest-ServersV294TestFqdnHostnames-1426486100 tempest-ServersV294TestFqdnHostnames-1426486100-project-member] [instance: 2e4a5344-600f-4b61-826e-c15f96b50af2] allocate_for_instance() {{(pid=61957) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 868.959051] env[61957]: DEBUG oslo_concurrency.lockutils [None req-09e8eb69-1030-4228-8cb0-915d5cee1a52 tempest-AttachInterfacesV270Test-2032541732 tempest-AttachInterfacesV270Test-2032541732-project-member] Lock "281b9644-0e7d-48b9-a7b6-45fd6102d558" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 87.074s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 868.964792] env[61957]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-d44d5f16-b6df-47b4-a42f-f98af7e97f7f {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.969386] env[61957]: DEBUG oslo_vmware.api [None req-902199c7-5b38-4a0e-a88f-80539210637b tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52a774bf-22f6-09cc-3b9a-f74ee3a8f8f9, 'name': SearchDatastore_Task, 'duration_secs': 0.019454} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.969386] env[61957]: DEBUG oslo_concurrency.lockutils [None req-902199c7-5b38-4a0e-a88f-80539210637b tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 868.969386] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-902199c7-5b38-4a0e-a88f-80539210637b tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] [instance: 526ff179-62a6-4763-ab25-797617c4ed57] Processing image 11c76a2c-f705-470a-ba9d-4657858bab38 {{(pid=61957) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 868.969386] env[61957]: DEBUG oslo_concurrency.lockutils [None req-902199c7-5b38-4a0e-a88f-80539210637b tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 868.969386] env[61957]: DEBUG oslo_concurrency.lockutils [None req-902199c7-5b38-4a0e-a88f-80539210637b tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 868.969611] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-902199c7-5b38-4a0e-a88f-80539210637b tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 868.970083] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-033631c2-618a-44ea-9749-d9fcdf632aab {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.981294] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-902199c7-5b38-4a0e-a88f-80539210637b tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 868.981883] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-902199c7-5b38-4a0e-a88f-80539210637b tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61957) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 868.982645] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8d8814de-e5f2-4974-804c-3237a25e0d0d {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.990471] env[61957]: DEBUG oslo_vmware.api [None req-902199c7-5b38-4a0e-a88f-80539210637b tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Waiting for the task: (returnval){ [ 868.990471] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52f04f24-d275-e4b9-0807-195c728f16c5" [ 868.990471] env[61957]: _type = "Task" [ 868.990471] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.997524] env[61957]: DEBUG oslo_vmware.api [None req-902199c7-5b38-4a0e-a88f-80539210637b tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52f04f24-d275-e4b9-0807-195c728f16c5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.024378] env[61957]: DEBUG nova.policy [None req-85c79741-463e-4bc0-9ed2-4682b88fcc00 tempest-ServersV294TestFqdnHostnames-1426486100 tempest-ServersV294TestFqdnHostnames-1426486100-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3941a643182e4f6e818297158f8647c7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ecf709e5c49146238729af50f362b1ea', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61957) authorize /opt/stack/nova/nova/policy.py:203}} [ 869.034416] env[61957]: DEBUG nova.network.neutron [None req-c13ea078-3782-4688-93ee-408e1ac828c0 tempest-ServerTagsTestJSON-969152625 tempest-ServerTagsTestJSON-969152625-project-member] [instance: 8e04a098-25de-4d57-9e3c-ea44d234d57e] Updating instance_info_cache with network_info: [{"id": "55344e12-1398-4752-a0f4-fdea985786bc", "address": "fa:16:3e:f8:bc:bc", "network": {"id": "2153fc60-b52d-4ed2-b36e-de231997cbe9", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-2097535952-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e8d71e68f10c4ee3a816f73cf9add0a5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4055505f-97ab-400b-969c-43d99b38fd48", "external-id": "nsx-vlan-transportzone-952", "segmentation_id": 952, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap55344e12-13", "ovs_interfaceid": "55344e12-1398-4752-a0f4-fdea985786bc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 869.291653] env[61957]: DEBUG nova.network.neutron [req-f0c037a2-4476-4f2b-b120-bd7f1469fc85 req-944c8641-3d42-4bbe-80ff-3ed14452df0a service nova] [instance: c23141ee-0cbb-4d1b-8390-c3073fe354f1] Updated VIF entry in instance network info cache for port a9884dda-2fb3-4e26-84b1-877dd0e0459c. {{(pid=61957) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 869.292029] env[61957]: DEBUG nova.network.neutron [req-f0c037a2-4476-4f2b-b120-bd7f1469fc85 req-944c8641-3d42-4bbe-80ff-3ed14452df0a service nova] [instance: c23141ee-0cbb-4d1b-8390-c3073fe354f1] Updating instance_info_cache with network_info: [{"id": "a9884dda-2fb3-4e26-84b1-877dd0e0459c", "address": "fa:16:3e:f9:46:03", "network": {"id": "82fd2357-53b6-4903-963e-a2866736643e", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-612681178-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.165", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1673c854272f4b14a623a73b8f4b687b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6eb7e3e9-5cc2-40f1-a6eb-f70f06531667", "external-id": "nsx-vlan-transportzone-938", "segmentation_id": 938, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa9884dda-2f", "ovs_interfaceid": "a9884dda-2fb3-4e26-84b1-877dd0e0459c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 869.337046] env[61957]: DEBUG oslo_concurrency.lockutils [None req-c7666870-1954-4260-b0a0-b151a9e3e160 tempest-AttachInterfacesV270Test-2032541732 tempest-AttachInterfacesV270Test-2032541732-project-member] Acquiring lock "interface-281b9644-0e7d-48b9-a7b6-45fd6102d558-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 869.337046] env[61957]: DEBUG oslo_concurrency.lockutils [None req-c7666870-1954-4260-b0a0-b151a9e3e160 tempest-AttachInterfacesV270Test-2032541732 tempest-AttachInterfacesV270Test-2032541732-project-member] Lock "interface-281b9644-0e7d-48b9-a7b6-45fd6102d558-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 869.337192] env[61957]: DEBUG nova.objects.instance [None req-c7666870-1954-4260-b0a0-b151a9e3e160 tempest-AttachInterfacesV270Test-2032541732 tempest-AttachInterfacesV270Test-2032541732-project-member] Lazy-loading 'flavor' on Instance uuid 281b9644-0e7d-48b9-a7b6-45fd6102d558 {{(pid=61957) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 869.435534] env[61957]: DEBUG oslo_concurrency.lockutils [None req-82a5645a-fb19-4ad1-aa47-e12b7f048b77 tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.016s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 869.436718] env[61957]: DEBUG oslo_concurrency.lockutils [None req-582b6bd7-3488-48cf-a2f8-f44952b10345 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.066s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 869.438334] env[61957]: INFO nova.compute.claims [None req-582b6bd7-3488-48cf-a2f8-f44952b10345 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] [instance: e40fa112-4648-428f-a403-b3e3b8319ea8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 869.444013] env[61957]: DEBUG nova.compute.manager [None req-85c79741-463e-4bc0-9ed2-4682b88fcc00 tempest-ServersV294TestFqdnHostnames-1426486100 tempest-ServersV294TestFqdnHostnames-1426486100-project-member] [instance: 2e4a5344-600f-4b61-826e-c15f96b50af2] Start building block device mappings for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 869.471817] env[61957]: DEBUG nova.compute.manager [None req-ca2fb2e2-e99f-422a-be65-983bd81b6e15 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] [instance: 736eabcf-70f8-4e1e-9830-270d9971fc9a] Starting instance... {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 869.476121] env[61957]: DEBUG nova.network.neutron [None req-85c79741-463e-4bc0-9ed2-4682b88fcc00 tempest-ServersV294TestFqdnHostnames-1426486100 tempest-ServersV294TestFqdnHostnames-1426486100-project-member] [instance: 2e4a5344-600f-4b61-826e-c15f96b50af2] Successfully created port: a1026b09-e6b6-47f4-b5b2-567a15abfa41 {{(pid=61957) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 869.506286] env[61957]: DEBUG oslo_vmware.api [None req-902199c7-5b38-4a0e-a88f-80539210637b tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52f04f24-d275-e4b9-0807-195c728f16c5, 'name': SearchDatastore_Task, 'duration_secs': 0.026506} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.508612] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-84389cd6-aa66-4309-8825-9c7400db971c {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.518780] env[61957]: DEBUG oslo_vmware.api [None req-902199c7-5b38-4a0e-a88f-80539210637b tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Waiting for the task: (returnval){ [ 869.518780] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52b9ba82-0936-d5ac-e83c-03c85b134a27" [ 869.518780] env[61957]: _type = "Task" [ 869.518780] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.530297] env[61957]: DEBUG oslo_vmware.api [None req-902199c7-5b38-4a0e-a88f-80539210637b tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52b9ba82-0936-d5ac-e83c-03c85b134a27, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.537227] env[61957]: DEBUG oslo_concurrency.lockutils [None req-c13ea078-3782-4688-93ee-408e1ac828c0 tempest-ServerTagsTestJSON-969152625 tempest-ServerTagsTestJSON-969152625-project-member] Releasing lock "refresh_cache-8e04a098-25de-4d57-9e3c-ea44d234d57e" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 869.538624] env[61957]: DEBUG nova.compute.manager [None req-c13ea078-3782-4688-93ee-408e1ac828c0 tempest-ServerTagsTestJSON-969152625 tempest-ServerTagsTestJSON-969152625-project-member] [instance: 8e04a098-25de-4d57-9e3c-ea44d234d57e] Instance network_info: |[{"id": "55344e12-1398-4752-a0f4-fdea985786bc", "address": "fa:16:3e:f8:bc:bc", "network": {"id": "2153fc60-b52d-4ed2-b36e-de231997cbe9", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-2097535952-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e8d71e68f10c4ee3a816f73cf9add0a5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4055505f-97ab-400b-969c-43d99b38fd48", "external-id": "nsx-vlan-transportzone-952", "segmentation_id": 952, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap55344e12-13", "ovs_interfaceid": "55344e12-1398-4752-a0f4-fdea985786bc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 869.538624] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-c13ea078-3782-4688-93ee-408e1ac828c0 tempest-ServerTagsTestJSON-969152625 tempest-ServerTagsTestJSON-969152625-project-member] [instance: 8e04a098-25de-4d57-9e3c-ea44d234d57e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f8:bc:bc', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '4055505f-97ab-400b-969c-43d99b38fd48', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '55344e12-1398-4752-a0f4-fdea985786bc', 'vif_model': 'vmxnet3'}] {{(pid=61957) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 869.546928] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-c13ea078-3782-4688-93ee-408e1ac828c0 tempest-ServerTagsTestJSON-969152625 tempest-ServerTagsTestJSON-969152625-project-member] Creating folder: Project (e8d71e68f10c4ee3a816f73cf9add0a5). Parent ref: group-v274445. {{(pid=61957) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 869.547594] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-dd6fb933-759b-49e3-bc53-83f964f0a2a1 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.556119] env[61957]: DEBUG nova.compute.manager [req-c5327e90-2700-4ba7-8e61-d5ecac3bf62a req-82136c0f-b90e-4f96-a4d8-6cb77a3dd41b service nova] [instance: 8e04a098-25de-4d57-9e3c-ea44d234d57e] Received event network-vif-plugged-55344e12-1398-4752-a0f4-fdea985786bc {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 869.556413] env[61957]: DEBUG oslo_concurrency.lockutils [req-c5327e90-2700-4ba7-8e61-d5ecac3bf62a req-82136c0f-b90e-4f96-a4d8-6cb77a3dd41b service nova] Acquiring lock "8e04a098-25de-4d57-9e3c-ea44d234d57e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 869.556751] env[61957]: DEBUG oslo_concurrency.lockutils [req-c5327e90-2700-4ba7-8e61-d5ecac3bf62a req-82136c0f-b90e-4f96-a4d8-6cb77a3dd41b service nova] Lock "8e04a098-25de-4d57-9e3c-ea44d234d57e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 869.557242] env[61957]: DEBUG oslo_concurrency.lockutils [req-c5327e90-2700-4ba7-8e61-d5ecac3bf62a req-82136c0f-b90e-4f96-a4d8-6cb77a3dd41b service nova] Lock "8e04a098-25de-4d57-9e3c-ea44d234d57e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 869.557591] env[61957]: DEBUG nova.compute.manager [req-c5327e90-2700-4ba7-8e61-d5ecac3bf62a req-82136c0f-b90e-4f96-a4d8-6cb77a3dd41b service nova] [instance: 8e04a098-25de-4d57-9e3c-ea44d234d57e] No waiting events found dispatching network-vif-plugged-55344e12-1398-4752-a0f4-fdea985786bc {{(pid=61957) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 869.557895] env[61957]: WARNING nova.compute.manager [req-c5327e90-2700-4ba7-8e61-d5ecac3bf62a req-82136c0f-b90e-4f96-a4d8-6cb77a3dd41b service nova] [instance: 8e04a098-25de-4d57-9e3c-ea44d234d57e] Received unexpected event network-vif-plugged-55344e12-1398-4752-a0f4-fdea985786bc for instance with vm_state building and task_state spawning. [ 869.558344] env[61957]: DEBUG nova.compute.manager [req-c5327e90-2700-4ba7-8e61-d5ecac3bf62a req-82136c0f-b90e-4f96-a4d8-6cb77a3dd41b service nova] [instance: 8e04a098-25de-4d57-9e3c-ea44d234d57e] Received event network-changed-55344e12-1398-4752-a0f4-fdea985786bc {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 869.559620] env[61957]: DEBUG nova.compute.manager [req-c5327e90-2700-4ba7-8e61-d5ecac3bf62a req-82136c0f-b90e-4f96-a4d8-6cb77a3dd41b service nova] [instance: 8e04a098-25de-4d57-9e3c-ea44d234d57e] Refreshing instance network info cache due to event network-changed-55344e12-1398-4752-a0f4-fdea985786bc. {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 869.559935] env[61957]: DEBUG oslo_concurrency.lockutils [req-c5327e90-2700-4ba7-8e61-d5ecac3bf62a req-82136c0f-b90e-4f96-a4d8-6cb77a3dd41b service nova] Acquiring lock "refresh_cache-8e04a098-25de-4d57-9e3c-ea44d234d57e" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 869.560171] env[61957]: DEBUG oslo_concurrency.lockutils [req-c5327e90-2700-4ba7-8e61-d5ecac3bf62a req-82136c0f-b90e-4f96-a4d8-6cb77a3dd41b service nova] Acquired lock "refresh_cache-8e04a098-25de-4d57-9e3c-ea44d234d57e" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 869.560389] env[61957]: DEBUG nova.network.neutron [req-c5327e90-2700-4ba7-8e61-d5ecac3bf62a req-82136c0f-b90e-4f96-a4d8-6cb77a3dd41b service nova] [instance: 8e04a098-25de-4d57-9e3c-ea44d234d57e] Refreshing network info cache for port 55344e12-1398-4752-a0f4-fdea985786bc {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 869.563645] env[61957]: INFO nova.virt.vmwareapi.vm_util [None req-c13ea078-3782-4688-93ee-408e1ac828c0 tempest-ServerTagsTestJSON-969152625 tempest-ServerTagsTestJSON-969152625-project-member] Created folder: Project (e8d71e68f10c4ee3a816f73cf9add0a5) in parent group-v274445. [ 869.563958] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-c13ea078-3782-4688-93ee-408e1ac828c0 tempest-ServerTagsTestJSON-969152625 tempest-ServerTagsTestJSON-969152625-project-member] Creating folder: Instances. Parent ref: group-v274523. {{(pid=61957) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 869.566589] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-350cb238-82a9-4e15-89e9-685c9979a2dc {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.795964] env[61957]: DEBUG oslo_concurrency.lockutils [req-f0c037a2-4476-4f2b-b120-bd7f1469fc85 req-944c8641-3d42-4bbe-80ff-3ed14452df0a service nova] Releasing lock "refresh_cache-c23141ee-0cbb-4d1b-8390-c3073fe354f1" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 869.842343] env[61957]: DEBUG nova.objects.instance [None req-c7666870-1954-4260-b0a0-b151a9e3e160 tempest-AttachInterfacesV270Test-2032541732 tempest-AttachInterfacesV270Test-2032541732-project-member] Lazy-loading 'pci_requests' on Instance uuid 281b9644-0e7d-48b9-a7b6-45fd6102d558 {{(pid=61957) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 870.003302] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ca2fb2e2-e99f-422a-be65-983bd81b6e15 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 870.030987] env[61957]: DEBUG oslo_vmware.api [None req-902199c7-5b38-4a0e-a88f-80539210637b tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52b9ba82-0936-d5ac-e83c-03c85b134a27, 'name': SearchDatastore_Task, 'duration_secs': 0.019887} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.030987] env[61957]: DEBUG oslo_concurrency.lockutils [None req-902199c7-5b38-4a0e-a88f-80539210637b tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 870.030987] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-902199c7-5b38-4a0e-a88f-80539210637b tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore2] 526ff179-62a6-4763-ab25-797617c4ed57/526ff179-62a6-4763-ab25-797617c4ed57.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 870.030987] env[61957]: INFO nova.virt.vmwareapi.vm_util [None req-c13ea078-3782-4688-93ee-408e1ac828c0 tempest-ServerTagsTestJSON-969152625 tempest-ServerTagsTestJSON-969152625-project-member] Created folder: Instances in parent group-v274523. [ 870.031419] env[61957]: DEBUG oslo.service.loopingcall [None req-c13ea078-3782-4688-93ee-408e1ac828c0 tempest-ServerTagsTestJSON-969152625 tempest-ServerTagsTestJSON-969152625-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 870.031629] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f07a97a6-f547-4f07-8046-417056635068 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.034927] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8e04a098-25de-4d57-9e3c-ea44d234d57e] Creating VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 870.036438] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9c2f9bd7-6c97-43d4-a89d-f9ba28b2e285 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.057710] env[61957]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 870.057710] env[61957]: value = "task-1277574" [ 870.057710] env[61957]: _type = "Task" [ 870.057710] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.059153] env[61957]: DEBUG oslo_vmware.api [None req-902199c7-5b38-4a0e-a88f-80539210637b tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Waiting for the task: (returnval){ [ 870.059153] env[61957]: value = "task-1277573" [ 870.059153] env[61957]: _type = "Task" [ 870.059153] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.072246] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277574, 'name': CreateVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.076425] env[61957]: DEBUG oslo_vmware.api [None req-902199c7-5b38-4a0e-a88f-80539210637b tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Task: {'id': task-1277573, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.325477] env[61957]: DEBUG nova.network.neutron [req-c5327e90-2700-4ba7-8e61-d5ecac3bf62a req-82136c0f-b90e-4f96-a4d8-6cb77a3dd41b service nova] [instance: 8e04a098-25de-4d57-9e3c-ea44d234d57e] Updated VIF entry in instance network info cache for port 55344e12-1398-4752-a0f4-fdea985786bc. {{(pid=61957) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 870.325934] env[61957]: DEBUG nova.network.neutron [req-c5327e90-2700-4ba7-8e61-d5ecac3bf62a req-82136c0f-b90e-4f96-a4d8-6cb77a3dd41b service nova] [instance: 8e04a098-25de-4d57-9e3c-ea44d234d57e] Updating instance_info_cache with network_info: [{"id": "55344e12-1398-4752-a0f4-fdea985786bc", "address": "fa:16:3e:f8:bc:bc", "network": {"id": "2153fc60-b52d-4ed2-b36e-de231997cbe9", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-2097535952-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e8d71e68f10c4ee3a816f73cf9add0a5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4055505f-97ab-400b-969c-43d99b38fd48", "external-id": "nsx-vlan-transportzone-952", "segmentation_id": 952, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap55344e12-13", "ovs_interfaceid": "55344e12-1398-4752-a0f4-fdea985786bc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 870.347591] env[61957]: DEBUG nova.objects.base [None req-c7666870-1954-4260-b0a0-b151a9e3e160 tempest-AttachInterfacesV270Test-2032541732 tempest-AttachInterfacesV270Test-2032541732-project-member] Object Instance<281b9644-0e7d-48b9-a7b6-45fd6102d558> lazy-loaded attributes: flavor,pci_requests {{(pid=61957) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 870.347818] env[61957]: DEBUG nova.network.neutron [None req-c7666870-1954-4260-b0a0-b151a9e3e160 tempest-AttachInterfacesV270Test-2032541732 tempest-AttachInterfacesV270Test-2032541732-project-member] [instance: 281b9644-0e7d-48b9-a7b6-45fd6102d558] allocate_for_instance() {{(pid=61957) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 870.439048] env[61957]: DEBUG oslo_concurrency.lockutils [None req-c7666870-1954-4260-b0a0-b151a9e3e160 tempest-AttachInterfacesV270Test-2032541732 tempest-AttachInterfacesV270Test-2032541732-project-member] Lock "interface-281b9644-0e7d-48b9-a7b6-45fd6102d558-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 1.102s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 870.457013] env[61957]: DEBUG nova.compute.manager [None req-85c79741-463e-4bc0-9ed2-4682b88fcc00 tempest-ServersV294TestFqdnHostnames-1426486100 tempest-ServersV294TestFqdnHostnames-1426486100-project-member] [instance: 2e4a5344-600f-4b61-826e-c15f96b50af2] Start spawning the instance on the hypervisor. {{(pid=61957) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 870.489981] env[61957]: DEBUG nova.virt.hardware [None req-85c79741-463e-4bc0-9ed2-4682b88fcc00 tempest-ServersV294TestFqdnHostnames-1426486100 tempest-ServersV294TestFqdnHostnames-1426486100-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T17:22:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T17:21:50Z,direct_url=,disk_format='vmdk',id=11c76a2c-f705-470a-ba9d-4657858bab38,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='81d3096d0f094373913fc3dc8f5c3c6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T17:21:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 870.489981] env[61957]: DEBUG nova.virt.hardware [None req-85c79741-463e-4bc0-9ed2-4682b88fcc00 tempest-ServersV294TestFqdnHostnames-1426486100 tempest-ServersV294TestFqdnHostnames-1426486100-project-member] Flavor limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 870.489981] env[61957]: DEBUG nova.virt.hardware [None req-85c79741-463e-4bc0-9ed2-4682b88fcc00 tempest-ServersV294TestFqdnHostnames-1426486100 tempest-ServersV294TestFqdnHostnames-1426486100-project-member] Image limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 870.490413] env[61957]: DEBUG nova.virt.hardware [None req-85c79741-463e-4bc0-9ed2-4682b88fcc00 tempest-ServersV294TestFqdnHostnames-1426486100 tempest-ServersV294TestFqdnHostnames-1426486100-project-member] Flavor pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 870.490413] env[61957]: DEBUG nova.virt.hardware [None req-85c79741-463e-4bc0-9ed2-4682b88fcc00 tempest-ServersV294TestFqdnHostnames-1426486100 tempest-ServersV294TestFqdnHostnames-1426486100-project-member] Image pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 870.490413] env[61957]: DEBUG nova.virt.hardware [None req-85c79741-463e-4bc0-9ed2-4682b88fcc00 tempest-ServersV294TestFqdnHostnames-1426486100 tempest-ServersV294TestFqdnHostnames-1426486100-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 870.490814] env[61957]: DEBUG nova.virt.hardware [None req-85c79741-463e-4bc0-9ed2-4682b88fcc00 tempest-ServersV294TestFqdnHostnames-1426486100 tempest-ServersV294TestFqdnHostnames-1426486100-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 870.490949] env[61957]: DEBUG nova.virt.hardware [None req-85c79741-463e-4bc0-9ed2-4682b88fcc00 tempest-ServersV294TestFqdnHostnames-1426486100 tempest-ServersV294TestFqdnHostnames-1426486100-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 870.492127] env[61957]: DEBUG nova.virt.hardware [None req-85c79741-463e-4bc0-9ed2-4682b88fcc00 tempest-ServersV294TestFqdnHostnames-1426486100 tempest-ServersV294TestFqdnHostnames-1426486100-project-member] Got 1 possible topologies {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 870.492127] env[61957]: DEBUG nova.virt.hardware [None req-85c79741-463e-4bc0-9ed2-4682b88fcc00 tempest-ServersV294TestFqdnHostnames-1426486100 tempest-ServersV294TestFqdnHostnames-1426486100-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 870.492127] env[61957]: DEBUG nova.virt.hardware [None req-85c79741-463e-4bc0-9ed2-4682b88fcc00 tempest-ServersV294TestFqdnHostnames-1426486100 tempest-ServersV294TestFqdnHostnames-1426486100-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 870.492754] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cccb129-3f60-40ea-8a4f-44bcb9656f18 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.504382] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a89e13cc-73b9-47c2-a917-bca5796b0c88 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.574321] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277574, 'name': CreateVM_Task} progress is 25%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.579192] env[61957]: DEBUG oslo_vmware.api [None req-902199c7-5b38-4a0e-a88f-80539210637b tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Task: {'id': task-1277573, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.518632} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.579628] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-902199c7-5b38-4a0e-a88f-80539210637b tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore2] 526ff179-62a6-4763-ab25-797617c4ed57/526ff179-62a6-4763-ab25-797617c4ed57.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 870.579796] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-902199c7-5b38-4a0e-a88f-80539210637b tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] [instance: 526ff179-62a6-4763-ab25-797617c4ed57] Extending root virtual disk to 1048576 {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 870.580535] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c5c42a48-9209-4b59-bf42-7cb1ff27e9e8 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.590701] env[61957]: DEBUG oslo_vmware.api [None req-902199c7-5b38-4a0e-a88f-80539210637b tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Waiting for the task: (returnval){ [ 870.590701] env[61957]: value = "task-1277575" [ 870.590701] env[61957]: _type = "Task" [ 870.590701] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.599605] env[61957]: DEBUG oslo_vmware.api [None req-902199c7-5b38-4a0e-a88f-80539210637b tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Task: {'id': task-1277575, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.823453] env[61957]: DEBUG oslo_concurrency.lockutils [None req-bb69c8c2-41e0-4fa7-aea3-e810870c1532 tempest-ServersNegativeTestMultiTenantJSON-2135399494 tempest-ServersNegativeTestMultiTenantJSON-2135399494-project-member] Acquiring lock "19966b0a-53b7-48c5-849c-a9d00dc024f8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 870.823715] env[61957]: DEBUG oslo_concurrency.lockutils [None req-bb69c8c2-41e0-4fa7-aea3-e810870c1532 tempest-ServersNegativeTestMultiTenantJSON-2135399494 tempest-ServersNegativeTestMultiTenantJSON-2135399494-project-member] Lock "19966b0a-53b7-48c5-849c-a9d00dc024f8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 870.829624] env[61957]: DEBUG oslo_concurrency.lockutils [req-c5327e90-2700-4ba7-8e61-d5ecac3bf62a req-82136c0f-b90e-4f96-a4d8-6cb77a3dd41b service nova] Releasing lock "refresh_cache-8e04a098-25de-4d57-9e3c-ea44d234d57e" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 870.856390] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9c51685-4eae-4a54-8bd2-a3afb7de13d1 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.862084] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8244609b-8782-4ff6-825e-b28369cb8929 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.896891] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f385921f-423d-4193-9d0e-e752f5c15e01 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.905022] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-627a611d-bfff-4a36-a959-d16f83b99e71 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.919546] env[61957]: DEBUG nova.compute.provider_tree [None req-582b6bd7-3488-48cf-a2f8-f44952b10345 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 871.072274] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277574, 'name': CreateVM_Task, 'duration_secs': 0.633408} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.075341] env[61957]: DEBUG nova.network.neutron [None req-85c79741-463e-4bc0-9ed2-4682b88fcc00 tempest-ServersV294TestFqdnHostnames-1426486100 tempest-ServersV294TestFqdnHostnames-1426486100-project-member] [instance: 2e4a5344-600f-4b61-826e-c15f96b50af2] Successfully updated port: a1026b09-e6b6-47f4-b5b2-567a15abfa41 {{(pid=61957) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 871.079248] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8e04a098-25de-4d57-9e3c-ea44d234d57e] Created VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 871.079248] env[61957]: DEBUG oslo_concurrency.lockutils [None req-c13ea078-3782-4688-93ee-408e1ac828c0 tempest-ServerTagsTestJSON-969152625 tempest-ServerTagsTestJSON-969152625-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 871.079248] env[61957]: DEBUG oslo_concurrency.lockutils [None req-c13ea078-3782-4688-93ee-408e1ac828c0 tempest-ServerTagsTestJSON-969152625 tempest-ServerTagsTestJSON-969152625-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 871.079248] env[61957]: DEBUG oslo_concurrency.lockutils [None req-c13ea078-3782-4688-93ee-408e1ac828c0 tempest-ServerTagsTestJSON-969152625 tempest-ServerTagsTestJSON-969152625-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 871.079248] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-94850cf6-7963-4148-bda6-ab1688cffb35 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.084739] env[61957]: DEBUG oslo_vmware.api [None req-c13ea078-3782-4688-93ee-408e1ac828c0 tempest-ServerTagsTestJSON-969152625 tempest-ServerTagsTestJSON-969152625-project-member] Waiting for the task: (returnval){ [ 871.084739] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]5285d44f-b0ea-a94e-da35-94f99cad46bb" [ 871.084739] env[61957]: _type = "Task" [ 871.084739] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.100308] env[61957]: DEBUG oslo_vmware.api [None req-c13ea078-3782-4688-93ee-408e1ac828c0 tempest-ServerTagsTestJSON-969152625 tempest-ServerTagsTestJSON-969152625-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]5285d44f-b0ea-a94e-da35-94f99cad46bb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.104047] env[61957]: DEBUG oslo_vmware.api [None req-902199c7-5b38-4a0e-a88f-80539210637b tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Task: {'id': task-1277575, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.094682} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.104334] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-902199c7-5b38-4a0e-a88f-80539210637b tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] [instance: 526ff179-62a6-4763-ab25-797617c4ed57] Extended root virtual disk {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 871.105126] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a09872b3-76ec-4b2e-87ee-5d34f427fa6a {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.134148] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-902199c7-5b38-4a0e-a88f-80539210637b tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] [instance: 526ff179-62a6-4763-ab25-797617c4ed57] Reconfiguring VM instance instance-00000041 to attach disk [datastore2] 526ff179-62a6-4763-ab25-797617c4ed57/526ff179-62a6-4763-ab25-797617c4ed57.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 871.134497] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6235d733-a0b4-4bc2-b7ac-4f948a19e1a5 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.160838] env[61957]: DEBUG oslo_vmware.api [None req-902199c7-5b38-4a0e-a88f-80539210637b tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Waiting for the task: (returnval){ [ 871.160838] env[61957]: value = "task-1277576" [ 871.160838] env[61957]: _type = "Task" [ 871.160838] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.169014] env[61957]: DEBUG oslo_vmware.api [None req-902199c7-5b38-4a0e-a88f-80539210637b tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Task: {'id': task-1277576, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.423624] env[61957]: DEBUG nova.scheduler.client.report [None req-582b6bd7-3488-48cf-a2f8-f44952b10345 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 871.578014] env[61957]: DEBUG oslo_concurrency.lockutils [None req-85c79741-463e-4bc0-9ed2-4682b88fcc00 tempest-ServersV294TestFqdnHostnames-1426486100 tempest-ServersV294TestFqdnHostnames-1426486100-project-member] Acquiring lock "refresh_cache-2e4a5344-600f-4b61-826e-c15f96b50af2" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 871.578190] env[61957]: DEBUG oslo_concurrency.lockutils [None req-85c79741-463e-4bc0-9ed2-4682b88fcc00 tempest-ServersV294TestFqdnHostnames-1426486100 tempest-ServersV294TestFqdnHostnames-1426486100-project-member] Acquired lock "refresh_cache-2e4a5344-600f-4b61-826e-c15f96b50af2" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 871.578471] env[61957]: DEBUG nova.network.neutron [None req-85c79741-463e-4bc0-9ed2-4682b88fcc00 tempest-ServersV294TestFqdnHostnames-1426486100 tempest-ServersV294TestFqdnHostnames-1426486100-project-member] [instance: 2e4a5344-600f-4b61-826e-c15f96b50af2] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 871.596555] env[61957]: DEBUG oslo_vmware.api [None req-c13ea078-3782-4688-93ee-408e1ac828c0 tempest-ServerTagsTestJSON-969152625 tempest-ServerTagsTestJSON-969152625-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]5285d44f-b0ea-a94e-da35-94f99cad46bb, 'name': SearchDatastore_Task, 'duration_secs': 0.01605} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.596874] env[61957]: DEBUG oslo_concurrency.lockutils [None req-c13ea078-3782-4688-93ee-408e1ac828c0 tempest-ServerTagsTestJSON-969152625 tempest-ServerTagsTestJSON-969152625-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 871.597133] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-c13ea078-3782-4688-93ee-408e1ac828c0 tempest-ServerTagsTestJSON-969152625 tempest-ServerTagsTestJSON-969152625-project-member] [instance: 8e04a098-25de-4d57-9e3c-ea44d234d57e] Processing image 11c76a2c-f705-470a-ba9d-4657858bab38 {{(pid=61957) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 871.597370] env[61957]: DEBUG oslo_concurrency.lockutils [None req-c13ea078-3782-4688-93ee-408e1ac828c0 tempest-ServerTagsTestJSON-969152625 tempest-ServerTagsTestJSON-969152625-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 871.597540] env[61957]: DEBUG oslo_concurrency.lockutils [None req-c13ea078-3782-4688-93ee-408e1ac828c0 tempest-ServerTagsTestJSON-969152625 tempest-ServerTagsTestJSON-969152625-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 871.597726] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-c13ea078-3782-4688-93ee-408e1ac828c0 tempest-ServerTagsTestJSON-969152625 tempest-ServerTagsTestJSON-969152625-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 871.598025] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-bc14531d-7522-4202-8744-b8f561ef5dd2 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.608325] env[61957]: DEBUG nova.compute.manager [req-e1048e44-6f2e-43f6-bd12-bcb9ea1d063a req-f341c3b2-83c2-474d-b9a6-c61e93eda512 service nova] [instance: 2e4a5344-600f-4b61-826e-c15f96b50af2] Received event network-vif-plugged-a1026b09-e6b6-47f4-b5b2-567a15abfa41 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 871.608638] env[61957]: DEBUG oslo_concurrency.lockutils [req-e1048e44-6f2e-43f6-bd12-bcb9ea1d063a req-f341c3b2-83c2-474d-b9a6-c61e93eda512 service nova] Acquiring lock "2e4a5344-600f-4b61-826e-c15f96b50af2-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 871.608876] env[61957]: DEBUG oslo_concurrency.lockutils [req-e1048e44-6f2e-43f6-bd12-bcb9ea1d063a req-f341c3b2-83c2-474d-b9a6-c61e93eda512 service nova] Lock "2e4a5344-600f-4b61-826e-c15f96b50af2-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 871.609097] env[61957]: DEBUG oslo_concurrency.lockutils [req-e1048e44-6f2e-43f6-bd12-bcb9ea1d063a req-f341c3b2-83c2-474d-b9a6-c61e93eda512 service nova] Lock "2e4a5344-600f-4b61-826e-c15f96b50af2-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 871.609341] env[61957]: DEBUG nova.compute.manager [req-e1048e44-6f2e-43f6-bd12-bcb9ea1d063a req-f341c3b2-83c2-474d-b9a6-c61e93eda512 service nova] [instance: 2e4a5344-600f-4b61-826e-c15f96b50af2] No waiting events found dispatching network-vif-plugged-a1026b09-e6b6-47f4-b5b2-567a15abfa41 {{(pid=61957) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 871.609556] env[61957]: WARNING nova.compute.manager [req-e1048e44-6f2e-43f6-bd12-bcb9ea1d063a req-f341c3b2-83c2-474d-b9a6-c61e93eda512 service nova] [instance: 2e4a5344-600f-4b61-826e-c15f96b50af2] Received unexpected event network-vif-plugged-a1026b09-e6b6-47f4-b5b2-567a15abfa41 for instance with vm_state building and task_state spawning. [ 871.609745] env[61957]: DEBUG nova.compute.manager [req-e1048e44-6f2e-43f6-bd12-bcb9ea1d063a req-f341c3b2-83c2-474d-b9a6-c61e93eda512 service nova] [instance: 2e4a5344-600f-4b61-826e-c15f96b50af2] Received event network-changed-a1026b09-e6b6-47f4-b5b2-567a15abfa41 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 871.609955] env[61957]: DEBUG nova.compute.manager [req-e1048e44-6f2e-43f6-bd12-bcb9ea1d063a req-f341c3b2-83c2-474d-b9a6-c61e93eda512 service nova] [instance: 2e4a5344-600f-4b61-826e-c15f96b50af2] Refreshing instance network info cache due to event network-changed-a1026b09-e6b6-47f4-b5b2-567a15abfa41. {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 871.610217] env[61957]: DEBUG oslo_concurrency.lockutils [req-e1048e44-6f2e-43f6-bd12-bcb9ea1d063a req-f341c3b2-83c2-474d-b9a6-c61e93eda512 service nova] Acquiring lock "refresh_cache-2e4a5344-600f-4b61-826e-c15f96b50af2" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 871.611935] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-c13ea078-3782-4688-93ee-408e1ac828c0 tempest-ServerTagsTestJSON-969152625 tempest-ServerTagsTestJSON-969152625-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 871.612132] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-c13ea078-3782-4688-93ee-408e1ac828c0 tempest-ServerTagsTestJSON-969152625 tempest-ServerTagsTestJSON-969152625-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61957) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 871.612841] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d84c73e9-c4e1-4c87-884a-b362d97d6937 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.618830] env[61957]: DEBUG oslo_vmware.api [None req-c13ea078-3782-4688-93ee-408e1ac828c0 tempest-ServerTagsTestJSON-969152625 tempest-ServerTagsTestJSON-969152625-project-member] Waiting for the task: (returnval){ [ 871.618830] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]5285c9bc-c2a5-4f11-dc24-07717197dc38" [ 871.618830] env[61957]: _type = "Task" [ 871.618830] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.627103] env[61957]: DEBUG oslo_vmware.api [None req-c13ea078-3782-4688-93ee-408e1ac828c0 tempest-ServerTagsTestJSON-969152625 tempest-ServerTagsTestJSON-969152625-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]5285c9bc-c2a5-4f11-dc24-07717197dc38, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.673130] env[61957]: DEBUG oslo_vmware.api [None req-902199c7-5b38-4a0e-a88f-80539210637b tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Task: {'id': task-1277576, 'name': ReconfigVM_Task, 'duration_secs': 0.315952} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.673539] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-902199c7-5b38-4a0e-a88f-80539210637b tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] [instance: 526ff179-62a6-4763-ab25-797617c4ed57] Reconfigured VM instance instance-00000041 to attach disk [datastore2] 526ff179-62a6-4763-ab25-797617c4ed57/526ff179-62a6-4763-ab25-797617c4ed57.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 871.674292] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-33bc7e5b-021c-4558-9144-7ae1781a2703 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.681034] env[61957]: DEBUG oslo_vmware.api [None req-902199c7-5b38-4a0e-a88f-80539210637b tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Waiting for the task: (returnval){ [ 871.681034] env[61957]: value = "task-1277577" [ 871.681034] env[61957]: _type = "Task" [ 871.681034] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.689153] env[61957]: DEBUG oslo_vmware.api [None req-902199c7-5b38-4a0e-a88f-80539210637b tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Task: {'id': task-1277577, 'name': Rename_Task} progress is 5%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.928657] env[61957]: DEBUG oslo_concurrency.lockutils [None req-582b6bd7-3488-48cf-a2f8-f44952b10345 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.492s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 871.933020] env[61957]: DEBUG nova.compute.manager [None req-582b6bd7-3488-48cf-a2f8-f44952b10345 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] [instance: e40fa112-4648-428f-a403-b3e3b8319ea8] Start building networks asynchronously for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 871.934693] env[61957]: DEBUG oslo_concurrency.lockutils [None req-06bfc4e2-99e2-4510-98f7-f5b198d6231a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.195s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 871.936184] env[61957]: INFO nova.compute.claims [None req-06bfc4e2-99e2-4510-98f7-f5b198d6231a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 1a0327c2-5671-4970-9db7-c7cc912d8678] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 872.118911] env[61957]: DEBUG nova.network.neutron [None req-85c79741-463e-4bc0-9ed2-4682b88fcc00 tempest-ServersV294TestFqdnHostnames-1426486100 tempest-ServersV294TestFqdnHostnames-1426486100-project-member] [instance: 2e4a5344-600f-4b61-826e-c15f96b50af2] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 872.131250] env[61957]: DEBUG oslo_vmware.api [None req-c13ea078-3782-4688-93ee-408e1ac828c0 tempest-ServerTagsTestJSON-969152625 tempest-ServerTagsTestJSON-969152625-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]5285c9bc-c2a5-4f11-dc24-07717197dc38, 'name': SearchDatastore_Task, 'duration_secs': 0.019086} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.132279] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7ad15b72-eab3-4203-a9e5-8df309d7ef61 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.137848] env[61957]: DEBUG oslo_vmware.api [None req-c13ea078-3782-4688-93ee-408e1ac828c0 tempest-ServerTagsTestJSON-969152625 tempest-ServerTagsTestJSON-969152625-project-member] Waiting for the task: (returnval){ [ 872.137848] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52518cab-2a84-5355-e3da-affd57c37157" [ 872.137848] env[61957]: _type = "Task" [ 872.137848] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.147563] env[61957]: DEBUG oslo_vmware.api [None req-c13ea078-3782-4688-93ee-408e1ac828c0 tempest-ServerTagsTestJSON-969152625 tempest-ServerTagsTestJSON-969152625-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52518cab-2a84-5355-e3da-affd57c37157, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.191543] env[61957]: DEBUG oslo_vmware.api [None req-902199c7-5b38-4a0e-a88f-80539210637b tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Task: {'id': task-1277577, 'name': Rename_Task, 'duration_secs': 0.149578} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.191797] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-902199c7-5b38-4a0e-a88f-80539210637b tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] [instance: 526ff179-62a6-4763-ab25-797617c4ed57] Powering on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 872.192076] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2305d618-61cf-45e5-9662-00c3a73a30e2 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.198497] env[61957]: DEBUG oslo_vmware.api [None req-902199c7-5b38-4a0e-a88f-80539210637b tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Waiting for the task: (returnval){ [ 872.198497] env[61957]: value = "task-1277578" [ 872.198497] env[61957]: _type = "Task" [ 872.198497] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.206036] env[61957]: DEBUG oslo_vmware.api [None req-902199c7-5b38-4a0e-a88f-80539210637b tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Task: {'id': task-1277578, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.267740] env[61957]: DEBUG nova.network.neutron [None req-85c79741-463e-4bc0-9ed2-4682b88fcc00 tempest-ServersV294TestFqdnHostnames-1426486100 tempest-ServersV294TestFqdnHostnames-1426486100-project-member] [instance: 2e4a5344-600f-4b61-826e-c15f96b50af2] Updating instance_info_cache with network_info: [{"id": "a1026b09-e6b6-47f4-b5b2-567a15abfa41", "address": "fa:16:3e:38:85:22", "network": {"id": "7ca8d6d5-0a42-4dd6-b7bd-bba4b8618fca", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-636668201-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ecf709e5c49146238729af50f362b1ea", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e2e8b74b-aa27-4f31-9414-7bcf531e8642", "external-id": "nsx-vlan-transportzone-544", "segmentation_id": 544, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa1026b09-e6", "ovs_interfaceid": "a1026b09-e6b6-47f4-b5b2-567a15abfa41", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 872.418616] env[61957]: DEBUG oslo_concurrency.lockutils [None req-fd686081-caee-4453-9fa7-68efd0fac70e tempest-AttachInterfacesV270Test-2032541732 tempest-AttachInterfacesV270Test-2032541732-project-member] Acquiring lock "281b9644-0e7d-48b9-a7b6-45fd6102d558" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 872.418895] env[61957]: DEBUG oslo_concurrency.lockutils [None req-fd686081-caee-4453-9fa7-68efd0fac70e tempest-AttachInterfacesV270Test-2032541732 tempest-AttachInterfacesV270Test-2032541732-project-member] Lock "281b9644-0e7d-48b9-a7b6-45fd6102d558" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 872.419579] env[61957]: DEBUG oslo_concurrency.lockutils [None req-fd686081-caee-4453-9fa7-68efd0fac70e tempest-AttachInterfacesV270Test-2032541732 tempest-AttachInterfacesV270Test-2032541732-project-member] Acquiring lock "281b9644-0e7d-48b9-a7b6-45fd6102d558-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 872.419579] env[61957]: DEBUG oslo_concurrency.lockutils [None req-fd686081-caee-4453-9fa7-68efd0fac70e tempest-AttachInterfacesV270Test-2032541732 tempest-AttachInterfacesV270Test-2032541732-project-member] Lock "281b9644-0e7d-48b9-a7b6-45fd6102d558-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 872.419758] env[61957]: DEBUG oslo_concurrency.lockutils [None req-fd686081-caee-4453-9fa7-68efd0fac70e tempest-AttachInterfacesV270Test-2032541732 tempest-AttachInterfacesV270Test-2032541732-project-member] Lock "281b9644-0e7d-48b9-a7b6-45fd6102d558-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 872.422017] env[61957]: INFO nova.compute.manager [None req-fd686081-caee-4453-9fa7-68efd0fac70e tempest-AttachInterfacesV270Test-2032541732 tempest-AttachInterfacesV270Test-2032541732-project-member] [instance: 281b9644-0e7d-48b9-a7b6-45fd6102d558] Terminating instance [ 872.423954] env[61957]: DEBUG nova.compute.manager [None req-fd686081-caee-4453-9fa7-68efd0fac70e tempest-AttachInterfacesV270Test-2032541732 tempest-AttachInterfacesV270Test-2032541732-project-member] [instance: 281b9644-0e7d-48b9-a7b6-45fd6102d558] Start destroying the instance on the hypervisor. {{(pid=61957) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 872.424182] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-fd686081-caee-4453-9fa7-68efd0fac70e tempest-AttachInterfacesV270Test-2032541732 tempest-AttachInterfacesV270Test-2032541732-project-member] [instance: 281b9644-0e7d-48b9-a7b6-45fd6102d558] Destroying instance {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 872.425068] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2a5d67f-2a37-4f51-9964-7b55c0b717e5 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.433196] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd686081-caee-4453-9fa7-68efd0fac70e tempest-AttachInterfacesV270Test-2032541732 tempest-AttachInterfacesV270Test-2032541732-project-member] [instance: 281b9644-0e7d-48b9-a7b6-45fd6102d558] Powering off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 872.433309] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-523d8c77-cd25-465d-a7da-2c7f8298723c {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.438876] env[61957]: DEBUG oslo_vmware.api [None req-fd686081-caee-4453-9fa7-68efd0fac70e tempest-AttachInterfacesV270Test-2032541732 tempest-AttachInterfacesV270Test-2032541732-project-member] Waiting for the task: (returnval){ [ 872.438876] env[61957]: value = "task-1277579" [ 872.438876] env[61957]: _type = "Task" [ 872.438876] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.443060] env[61957]: DEBUG nova.compute.utils [None req-582b6bd7-3488-48cf-a2f8-f44952b10345 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Using /dev/sd instead of None {{(pid=61957) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 872.446996] env[61957]: DEBUG nova.compute.manager [None req-582b6bd7-3488-48cf-a2f8-f44952b10345 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] [instance: e40fa112-4648-428f-a403-b3e3b8319ea8] Allocating IP information in the background. {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 872.447223] env[61957]: DEBUG nova.network.neutron [None req-582b6bd7-3488-48cf-a2f8-f44952b10345 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] [instance: e40fa112-4648-428f-a403-b3e3b8319ea8] allocate_for_instance() {{(pid=61957) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 872.454796] env[61957]: DEBUG oslo_vmware.api [None req-fd686081-caee-4453-9fa7-68efd0fac70e tempest-AttachInterfacesV270Test-2032541732 tempest-AttachInterfacesV270Test-2032541732-project-member] Task: {'id': task-1277579, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.499349] env[61957]: DEBUG nova.policy [None req-582b6bd7-3488-48cf-a2f8-f44952b10345 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9e3ccf0e318349b38c7eff2c89440e54', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd849d922cfad41848a33e8d3bc83ea0f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61957) authorize /opt/stack/nova/nova/policy.py:203}} [ 872.649410] env[61957]: DEBUG oslo_vmware.api [None req-c13ea078-3782-4688-93ee-408e1ac828c0 tempest-ServerTagsTestJSON-969152625 tempest-ServerTagsTestJSON-969152625-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52518cab-2a84-5355-e3da-affd57c37157, 'name': SearchDatastore_Task, 'duration_secs': 0.014103} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.649694] env[61957]: DEBUG oslo_concurrency.lockutils [None req-c13ea078-3782-4688-93ee-408e1ac828c0 tempest-ServerTagsTestJSON-969152625 tempest-ServerTagsTestJSON-969152625-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 872.649962] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-c13ea078-3782-4688-93ee-408e1ac828c0 tempest-ServerTagsTestJSON-969152625 tempest-ServerTagsTestJSON-969152625-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore2] 8e04a098-25de-4d57-9e3c-ea44d234d57e/8e04a098-25de-4d57-9e3c-ea44d234d57e.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 872.650316] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ea2adcac-6e5d-4589-b279-6b9a5fe6995b {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.657040] env[61957]: DEBUG oslo_vmware.api [None req-c13ea078-3782-4688-93ee-408e1ac828c0 tempest-ServerTagsTestJSON-969152625 tempest-ServerTagsTestJSON-969152625-project-member] Waiting for the task: (returnval){ [ 872.657040] env[61957]: value = "task-1277580" [ 872.657040] env[61957]: _type = "Task" [ 872.657040] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.667175] env[61957]: DEBUG oslo_vmware.api [None req-c13ea078-3782-4688-93ee-408e1ac828c0 tempest-ServerTagsTestJSON-969152625 tempest-ServerTagsTestJSON-969152625-project-member] Task: {'id': task-1277580, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.712130] env[61957]: DEBUG oslo_vmware.api [None req-902199c7-5b38-4a0e-a88f-80539210637b tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Task: {'id': task-1277578, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.774162] env[61957]: DEBUG oslo_concurrency.lockutils [None req-85c79741-463e-4bc0-9ed2-4682b88fcc00 tempest-ServersV294TestFqdnHostnames-1426486100 tempest-ServersV294TestFqdnHostnames-1426486100-project-member] Releasing lock "refresh_cache-2e4a5344-600f-4b61-826e-c15f96b50af2" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 872.774162] env[61957]: DEBUG nova.compute.manager [None req-85c79741-463e-4bc0-9ed2-4682b88fcc00 tempest-ServersV294TestFqdnHostnames-1426486100 tempest-ServersV294TestFqdnHostnames-1426486100-project-member] [instance: 2e4a5344-600f-4b61-826e-c15f96b50af2] Instance network_info: |[{"id": "a1026b09-e6b6-47f4-b5b2-567a15abfa41", "address": "fa:16:3e:38:85:22", "network": {"id": "7ca8d6d5-0a42-4dd6-b7bd-bba4b8618fca", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-636668201-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ecf709e5c49146238729af50f362b1ea", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e2e8b74b-aa27-4f31-9414-7bcf531e8642", "external-id": "nsx-vlan-transportzone-544", "segmentation_id": 544, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa1026b09-e6", "ovs_interfaceid": "a1026b09-e6b6-47f4-b5b2-567a15abfa41", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 872.774162] env[61957]: DEBUG oslo_concurrency.lockutils [req-e1048e44-6f2e-43f6-bd12-bcb9ea1d063a req-f341c3b2-83c2-474d-b9a6-c61e93eda512 service nova] Acquired lock "refresh_cache-2e4a5344-600f-4b61-826e-c15f96b50af2" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 872.774162] env[61957]: DEBUG nova.network.neutron [req-e1048e44-6f2e-43f6-bd12-bcb9ea1d063a req-f341c3b2-83c2-474d-b9a6-c61e93eda512 service nova] [instance: 2e4a5344-600f-4b61-826e-c15f96b50af2] Refreshing network info cache for port a1026b09-e6b6-47f4-b5b2-567a15abfa41 {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 872.774162] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-85c79741-463e-4bc0-9ed2-4682b88fcc00 tempest-ServersV294TestFqdnHostnames-1426486100 tempest-ServersV294TestFqdnHostnames-1426486100-project-member] [instance: 2e4a5344-600f-4b61-826e-c15f96b50af2] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:38:85:22', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e2e8b74b-aa27-4f31-9414-7bcf531e8642', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a1026b09-e6b6-47f4-b5b2-567a15abfa41', 'vif_model': 'vmxnet3'}] {{(pid=61957) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 872.784079] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-85c79741-463e-4bc0-9ed2-4682b88fcc00 tempest-ServersV294TestFqdnHostnames-1426486100 tempest-ServersV294TestFqdnHostnames-1426486100-project-member] Creating folder: Project (ecf709e5c49146238729af50f362b1ea). Parent ref: group-v274445. {{(pid=61957) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 872.788754] env[61957]: DEBUG nova.network.neutron [None req-582b6bd7-3488-48cf-a2f8-f44952b10345 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] [instance: e40fa112-4648-428f-a403-b3e3b8319ea8] Successfully created port: c407d755-4caa-44d0-b9b2-8e59106cc23e {{(pid=61957) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 872.790762] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-28e26f72-cee6-43fb-bda6-dd23c2070ecf {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.805236] env[61957]: INFO nova.virt.vmwareapi.vm_util [None req-85c79741-463e-4bc0-9ed2-4682b88fcc00 tempest-ServersV294TestFqdnHostnames-1426486100 tempest-ServersV294TestFqdnHostnames-1426486100-project-member] Created folder: Project (ecf709e5c49146238729af50f362b1ea) in parent group-v274445. [ 872.805236] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-85c79741-463e-4bc0-9ed2-4682b88fcc00 tempest-ServersV294TestFqdnHostnames-1426486100 tempest-ServersV294TestFqdnHostnames-1426486100-project-member] Creating folder: Instances. Parent ref: group-v274526. {{(pid=61957) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 872.805236] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-becc7e19-6652-4022-bf9d-ea13f19ded07 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.813917] env[61957]: INFO nova.virt.vmwareapi.vm_util [None req-85c79741-463e-4bc0-9ed2-4682b88fcc00 tempest-ServersV294TestFqdnHostnames-1426486100 tempest-ServersV294TestFqdnHostnames-1426486100-project-member] Created folder: Instances in parent group-v274526. [ 872.815595] env[61957]: DEBUG oslo.service.loopingcall [None req-85c79741-463e-4bc0-9ed2-4682b88fcc00 tempest-ServersV294TestFqdnHostnames-1426486100 tempest-ServersV294TestFqdnHostnames-1426486100-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 872.815595] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2e4a5344-600f-4b61-826e-c15f96b50af2] Creating VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 872.815595] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c6149581-0f1a-46ce-ba2c-fd494f691f62 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.836164] env[61957]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 872.836164] env[61957]: value = "task-1277583" [ 872.836164] env[61957]: _type = "Task" [ 872.836164] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.849344] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277583, 'name': CreateVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.950166] env[61957]: DEBUG nova.compute.manager [None req-582b6bd7-3488-48cf-a2f8-f44952b10345 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] [instance: e40fa112-4648-428f-a403-b3e3b8319ea8] Start building block device mappings for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 872.958504] env[61957]: DEBUG oslo_vmware.api [None req-fd686081-caee-4453-9fa7-68efd0fac70e tempest-AttachInterfacesV270Test-2032541732 tempest-AttachInterfacesV270Test-2032541732-project-member] Task: {'id': task-1277579, 'name': PowerOffVM_Task, 'duration_secs': 0.194773} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.960160] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd686081-caee-4453-9fa7-68efd0fac70e tempest-AttachInterfacesV270Test-2032541732 tempest-AttachInterfacesV270Test-2032541732-project-member] [instance: 281b9644-0e7d-48b9-a7b6-45fd6102d558] Powered off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 872.960449] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-fd686081-caee-4453-9fa7-68efd0fac70e tempest-AttachInterfacesV270Test-2032541732 tempest-AttachInterfacesV270Test-2032541732-project-member] [instance: 281b9644-0e7d-48b9-a7b6-45fd6102d558] Unregistering the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 872.960796] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-51216f47-73fd-4d71-bca1-8185e2f36f8c {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.071476] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-fd686081-caee-4453-9fa7-68efd0fac70e tempest-AttachInterfacesV270Test-2032541732 tempest-AttachInterfacesV270Test-2032541732-project-member] [instance: 281b9644-0e7d-48b9-a7b6-45fd6102d558] Unregistered the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 873.071714] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-fd686081-caee-4453-9fa7-68efd0fac70e tempest-AttachInterfacesV270Test-2032541732 tempest-AttachInterfacesV270Test-2032541732-project-member] [instance: 281b9644-0e7d-48b9-a7b6-45fd6102d558] Deleting contents of the VM from datastore datastore2 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 873.072134] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-fd686081-caee-4453-9fa7-68efd0fac70e tempest-AttachInterfacesV270Test-2032541732 tempest-AttachInterfacesV270Test-2032541732-project-member] Deleting the datastore file [datastore2] 281b9644-0e7d-48b9-a7b6-45fd6102d558 {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 873.072444] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-90a100d2-45b7-41f1-9892-70c41e484ec0 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.081750] env[61957]: DEBUG oslo_vmware.api [None req-fd686081-caee-4453-9fa7-68efd0fac70e tempest-AttachInterfacesV270Test-2032541732 tempest-AttachInterfacesV270Test-2032541732-project-member] Waiting for the task: (returnval){ [ 873.081750] env[61957]: value = "task-1277585" [ 873.081750] env[61957]: _type = "Task" [ 873.081750] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.095766] env[61957]: DEBUG oslo_vmware.api [None req-fd686081-caee-4453-9fa7-68efd0fac70e tempest-AttachInterfacesV270Test-2032541732 tempest-AttachInterfacesV270Test-2032541732-project-member] Task: {'id': task-1277585, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.170483] env[61957]: DEBUG oslo_vmware.api [None req-c13ea078-3782-4688-93ee-408e1ac828c0 tempest-ServerTagsTestJSON-969152625 tempest-ServerTagsTestJSON-969152625-project-member] Task: {'id': task-1277580, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.457664} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.173305] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-c13ea078-3782-4688-93ee-408e1ac828c0 tempest-ServerTagsTestJSON-969152625 tempest-ServerTagsTestJSON-969152625-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore2] 8e04a098-25de-4d57-9e3c-ea44d234d57e/8e04a098-25de-4d57-9e3c-ea44d234d57e.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 873.173522] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-c13ea078-3782-4688-93ee-408e1ac828c0 tempest-ServerTagsTestJSON-969152625 tempest-ServerTagsTestJSON-969152625-project-member] [instance: 8e04a098-25de-4d57-9e3c-ea44d234d57e] Extending root virtual disk to 1048576 {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 873.174081] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0e255ac2-b778-4fd5-8889-04d266aa443f {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.182384] env[61957]: DEBUG oslo_vmware.api [None req-c13ea078-3782-4688-93ee-408e1ac828c0 tempest-ServerTagsTestJSON-969152625 tempest-ServerTagsTestJSON-969152625-project-member] Waiting for the task: (returnval){ [ 873.182384] env[61957]: value = "task-1277586" [ 873.182384] env[61957]: _type = "Task" [ 873.182384] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.193670] env[61957]: DEBUG oslo_vmware.api [None req-c13ea078-3782-4688-93ee-408e1ac828c0 tempest-ServerTagsTestJSON-969152625 tempest-ServerTagsTestJSON-969152625-project-member] Task: {'id': task-1277586, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.194665] env[61957]: DEBUG nova.network.neutron [req-e1048e44-6f2e-43f6-bd12-bcb9ea1d063a req-f341c3b2-83c2-474d-b9a6-c61e93eda512 service nova] [instance: 2e4a5344-600f-4b61-826e-c15f96b50af2] Updated VIF entry in instance network info cache for port a1026b09-e6b6-47f4-b5b2-567a15abfa41. {{(pid=61957) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 873.194958] env[61957]: DEBUG nova.network.neutron [req-e1048e44-6f2e-43f6-bd12-bcb9ea1d063a req-f341c3b2-83c2-474d-b9a6-c61e93eda512 service nova] [instance: 2e4a5344-600f-4b61-826e-c15f96b50af2] Updating instance_info_cache with network_info: [{"id": "a1026b09-e6b6-47f4-b5b2-567a15abfa41", "address": "fa:16:3e:38:85:22", "network": {"id": "7ca8d6d5-0a42-4dd6-b7bd-bba4b8618fca", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-636668201-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ecf709e5c49146238729af50f362b1ea", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e2e8b74b-aa27-4f31-9414-7bcf531e8642", "external-id": "nsx-vlan-transportzone-544", "segmentation_id": 544, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa1026b09-e6", "ovs_interfaceid": "a1026b09-e6b6-47f4-b5b2-567a15abfa41", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 873.209831] env[61957]: DEBUG oslo_vmware.api [None req-902199c7-5b38-4a0e-a88f-80539210637b tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Task: {'id': task-1277578, 'name': PowerOnVM_Task, 'duration_secs': 0.524228} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.213395] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-902199c7-5b38-4a0e-a88f-80539210637b tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] [instance: 526ff179-62a6-4763-ab25-797617c4ed57] Powered on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 873.213649] env[61957]: INFO nova.compute.manager [None req-902199c7-5b38-4a0e-a88f-80539210637b tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] [instance: 526ff179-62a6-4763-ab25-797617c4ed57] Took 8.95 seconds to spawn the instance on the hypervisor. [ 873.213726] env[61957]: DEBUG nova.compute.manager [None req-902199c7-5b38-4a0e-a88f-80539210637b tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] [instance: 526ff179-62a6-4763-ab25-797617c4ed57] Checking state {{(pid=61957) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 873.215418] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2168f624-981f-44fb-ab0c-1cc11da5ba7f {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.349761] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277583, 'name': CreateVM_Task, 'duration_secs': 0.448139} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.349761] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2e4a5344-600f-4b61-826e-c15f96b50af2] Created VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 873.350131] env[61957]: DEBUG oslo_concurrency.lockutils [None req-85c79741-463e-4bc0-9ed2-4682b88fcc00 tempest-ServersV294TestFqdnHostnames-1426486100 tempest-ServersV294TestFqdnHostnames-1426486100-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 873.350363] env[61957]: DEBUG oslo_concurrency.lockutils [None req-85c79741-463e-4bc0-9ed2-4682b88fcc00 tempest-ServersV294TestFqdnHostnames-1426486100 tempest-ServersV294TestFqdnHostnames-1426486100-project-member] Acquired lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 873.350685] env[61957]: DEBUG oslo_concurrency.lockutils [None req-85c79741-463e-4bc0-9ed2-4682b88fcc00 tempest-ServersV294TestFqdnHostnames-1426486100 tempest-ServersV294TestFqdnHostnames-1426486100-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 873.351453] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-418d2854-f4fa-4d06-b1a7-1ca323547295 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.359776] env[61957]: DEBUG oslo_vmware.api [None req-85c79741-463e-4bc0-9ed2-4682b88fcc00 tempest-ServersV294TestFqdnHostnames-1426486100 tempest-ServersV294TestFqdnHostnames-1426486100-project-member] Waiting for the task: (returnval){ [ 873.359776] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]524ec82f-2914-b491-2994-156e97d70496" [ 873.359776] env[61957]: _type = "Task" [ 873.359776] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.366708] env[61957]: DEBUG oslo_vmware.api [None req-85c79741-463e-4bc0-9ed2-4682b88fcc00 tempest-ServersV294TestFqdnHostnames-1426486100 tempest-ServersV294TestFqdnHostnames-1426486100-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]524ec82f-2914-b491-2994-156e97d70496, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.406214] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62b9d18a-a448-44b1-868c-71e4193f06ae {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.414268] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84a42a12-2f11-4c84-a90f-a401bb5d92b8 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.447238] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-255a0759-34f3-4249-aa39-30ccf15fa712 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.455157] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd7738ab-e803-429d-8dab-e7f5ce0e116d {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.474153] env[61957]: DEBUG nova.compute.provider_tree [None req-06bfc4e2-99e2-4510-98f7-f5b198d6231a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 873.592034] env[61957]: DEBUG oslo_vmware.api [None req-fd686081-caee-4453-9fa7-68efd0fac70e tempest-AttachInterfacesV270Test-2032541732 tempest-AttachInterfacesV270Test-2032541732-project-member] Task: {'id': task-1277585, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.248213} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.592433] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-fd686081-caee-4453-9fa7-68efd0fac70e tempest-AttachInterfacesV270Test-2032541732 tempest-AttachInterfacesV270Test-2032541732-project-member] Deleted the datastore file {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 873.592648] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-fd686081-caee-4453-9fa7-68efd0fac70e tempest-AttachInterfacesV270Test-2032541732 tempest-AttachInterfacesV270Test-2032541732-project-member] [instance: 281b9644-0e7d-48b9-a7b6-45fd6102d558] Deleted contents of the VM from datastore datastore2 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 873.592834] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-fd686081-caee-4453-9fa7-68efd0fac70e tempest-AttachInterfacesV270Test-2032541732 tempest-AttachInterfacesV270Test-2032541732-project-member] [instance: 281b9644-0e7d-48b9-a7b6-45fd6102d558] Instance destroyed {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 873.593133] env[61957]: INFO nova.compute.manager [None req-fd686081-caee-4453-9fa7-68efd0fac70e tempest-AttachInterfacesV270Test-2032541732 tempest-AttachInterfacesV270Test-2032541732-project-member] [instance: 281b9644-0e7d-48b9-a7b6-45fd6102d558] Took 1.17 seconds to destroy the instance on the hypervisor. [ 873.593272] env[61957]: DEBUG oslo.service.loopingcall [None req-fd686081-caee-4453-9fa7-68efd0fac70e tempest-AttachInterfacesV270Test-2032541732 tempest-AttachInterfacesV270Test-2032541732-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 873.593460] env[61957]: DEBUG nova.compute.manager [-] [instance: 281b9644-0e7d-48b9-a7b6-45fd6102d558] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 873.593556] env[61957]: DEBUG nova.network.neutron [-] [instance: 281b9644-0e7d-48b9-a7b6-45fd6102d558] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 873.694372] env[61957]: DEBUG oslo_vmware.api [None req-c13ea078-3782-4688-93ee-408e1ac828c0 tempest-ServerTagsTestJSON-969152625 tempest-ServerTagsTestJSON-969152625-project-member] Task: {'id': task-1277586, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.092963} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.694699] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-c13ea078-3782-4688-93ee-408e1ac828c0 tempest-ServerTagsTestJSON-969152625 tempest-ServerTagsTestJSON-969152625-project-member] [instance: 8e04a098-25de-4d57-9e3c-ea44d234d57e] Extended root virtual disk {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 873.695528] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ab0aa83-c84a-4ef7-9147-543edb778351 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.698340] env[61957]: DEBUG oslo_concurrency.lockutils [req-e1048e44-6f2e-43f6-bd12-bcb9ea1d063a req-f341c3b2-83c2-474d-b9a6-c61e93eda512 service nova] Releasing lock "refresh_cache-2e4a5344-600f-4b61-826e-c15f96b50af2" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 873.718666] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-c13ea078-3782-4688-93ee-408e1ac828c0 tempest-ServerTagsTestJSON-969152625 tempest-ServerTagsTestJSON-969152625-project-member] [instance: 8e04a098-25de-4d57-9e3c-ea44d234d57e] Reconfiguring VM instance instance-00000042 to attach disk [datastore2] 8e04a098-25de-4d57-9e3c-ea44d234d57e/8e04a098-25de-4d57-9e3c-ea44d234d57e.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 873.718995] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ac045ab2-13ec-4b4c-96a8-0fd0e39418c4 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.741243] env[61957]: INFO nova.compute.manager [None req-902199c7-5b38-4a0e-a88f-80539210637b tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] [instance: 526ff179-62a6-4763-ab25-797617c4ed57] Took 39.29 seconds to build instance. [ 873.744831] env[61957]: DEBUG oslo_vmware.api [None req-c13ea078-3782-4688-93ee-408e1ac828c0 tempest-ServerTagsTestJSON-969152625 tempest-ServerTagsTestJSON-969152625-project-member] Waiting for the task: (returnval){ [ 873.744831] env[61957]: value = "task-1277587" [ 873.744831] env[61957]: _type = "Task" [ 873.744831] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.754785] env[61957]: DEBUG oslo_vmware.api [None req-c13ea078-3782-4688-93ee-408e1ac828c0 tempest-ServerTagsTestJSON-969152625 tempest-ServerTagsTestJSON-969152625-project-member] Task: {'id': task-1277587, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.870724] env[61957]: DEBUG oslo_vmware.api [None req-85c79741-463e-4bc0-9ed2-4682b88fcc00 tempest-ServersV294TestFqdnHostnames-1426486100 tempest-ServersV294TestFqdnHostnames-1426486100-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]524ec82f-2914-b491-2994-156e97d70496, 'name': SearchDatastore_Task, 'duration_secs': 0.011317} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.871050] env[61957]: DEBUG oslo_concurrency.lockutils [None req-85c79741-463e-4bc0-9ed2-4682b88fcc00 tempest-ServersV294TestFqdnHostnames-1426486100 tempest-ServersV294TestFqdnHostnames-1426486100-project-member] Releasing lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 873.871672] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-85c79741-463e-4bc0-9ed2-4682b88fcc00 tempest-ServersV294TestFqdnHostnames-1426486100 tempest-ServersV294TestFqdnHostnames-1426486100-project-member] [instance: 2e4a5344-600f-4b61-826e-c15f96b50af2] Processing image 11c76a2c-f705-470a-ba9d-4657858bab38 {{(pid=61957) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 873.871672] env[61957]: DEBUG oslo_concurrency.lockutils [None req-85c79741-463e-4bc0-9ed2-4682b88fcc00 tempest-ServersV294TestFqdnHostnames-1426486100 tempest-ServersV294TestFqdnHostnames-1426486100-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 873.871876] env[61957]: DEBUG oslo_concurrency.lockutils [None req-85c79741-463e-4bc0-9ed2-4682b88fcc00 tempest-ServersV294TestFqdnHostnames-1426486100 tempest-ServersV294TestFqdnHostnames-1426486100-project-member] Acquired lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 873.871912] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-85c79741-463e-4bc0-9ed2-4682b88fcc00 tempest-ServersV294TestFqdnHostnames-1426486100 tempest-ServersV294TestFqdnHostnames-1426486100-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 873.872164] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-398bfa2b-76b7-422e-aada-68b74c34a795 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.888241] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-85c79741-463e-4bc0-9ed2-4682b88fcc00 tempest-ServersV294TestFqdnHostnames-1426486100 tempest-ServersV294TestFqdnHostnames-1426486100-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 873.888780] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-85c79741-463e-4bc0-9ed2-4682b88fcc00 tempest-ServersV294TestFqdnHostnames-1426486100 tempest-ServersV294TestFqdnHostnames-1426486100-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61957) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 873.889339] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ecd5ba0f-df03-4dd3-9aa0-2445725705da {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.894845] env[61957]: DEBUG oslo_vmware.api [None req-85c79741-463e-4bc0-9ed2-4682b88fcc00 tempest-ServersV294TestFqdnHostnames-1426486100 tempest-ServersV294TestFqdnHostnames-1426486100-project-member] Waiting for the task: (returnval){ [ 873.894845] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]5290e9df-4c06-1a8d-6f5a-84776cfc472c" [ 873.894845] env[61957]: _type = "Task" [ 873.894845] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.903674] env[61957]: DEBUG oslo_vmware.api [None req-85c79741-463e-4bc0-9ed2-4682b88fcc00 tempest-ServersV294TestFqdnHostnames-1426486100 tempest-ServersV294TestFqdnHostnames-1426486100-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]5290e9df-4c06-1a8d-6f5a-84776cfc472c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.908659] env[61957]: DEBUG nova.compute.manager [req-6a8ccb17-5346-494a-a98a-9a16c6d1ef3b req-0b526815-59d3-493b-8035-e496445adadb service nova] [instance: 281b9644-0e7d-48b9-a7b6-45fd6102d558] Received event network-vif-deleted-22b42651-5480-45d9-b363-9106dd6a7ce6 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 873.909063] env[61957]: INFO nova.compute.manager [req-6a8ccb17-5346-494a-a98a-9a16c6d1ef3b req-0b526815-59d3-493b-8035-e496445adadb service nova] [instance: 281b9644-0e7d-48b9-a7b6-45fd6102d558] Neutron deleted interface 22b42651-5480-45d9-b363-9106dd6a7ce6; detaching it from the instance and deleting it from the info cache [ 873.909063] env[61957]: DEBUG nova.network.neutron [req-6a8ccb17-5346-494a-a98a-9a16c6d1ef3b req-0b526815-59d3-493b-8035-e496445adadb service nova] [instance: 281b9644-0e7d-48b9-a7b6-45fd6102d558] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 873.965197] env[61957]: DEBUG nova.compute.manager [None req-582b6bd7-3488-48cf-a2f8-f44952b10345 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] [instance: e40fa112-4648-428f-a403-b3e3b8319ea8] Start spawning the instance on the hypervisor. {{(pid=61957) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 873.978947] env[61957]: DEBUG nova.scheduler.client.report [None req-06bfc4e2-99e2-4510-98f7-f5b198d6231a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 874.003426] env[61957]: DEBUG nova.virt.hardware [None req-582b6bd7-3488-48cf-a2f8-f44952b10345 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T17:22:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T17:21:50Z,direct_url=,disk_format='vmdk',id=11c76a2c-f705-470a-ba9d-4657858bab38,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='81d3096d0f094373913fc3dc8f5c3c6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T17:21:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 874.003426] env[61957]: DEBUG nova.virt.hardware [None req-582b6bd7-3488-48cf-a2f8-f44952b10345 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Flavor limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 874.003426] env[61957]: DEBUG nova.virt.hardware [None req-582b6bd7-3488-48cf-a2f8-f44952b10345 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Image limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 874.003426] env[61957]: DEBUG nova.virt.hardware [None req-582b6bd7-3488-48cf-a2f8-f44952b10345 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Flavor pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 874.003426] env[61957]: DEBUG nova.virt.hardware [None req-582b6bd7-3488-48cf-a2f8-f44952b10345 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Image pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 874.003426] env[61957]: DEBUG nova.virt.hardware [None req-582b6bd7-3488-48cf-a2f8-f44952b10345 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 874.004366] env[61957]: DEBUG nova.virt.hardware [None req-582b6bd7-3488-48cf-a2f8-f44952b10345 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 874.004366] env[61957]: DEBUG nova.virt.hardware [None req-582b6bd7-3488-48cf-a2f8-f44952b10345 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 874.004366] env[61957]: DEBUG nova.virt.hardware [None req-582b6bd7-3488-48cf-a2f8-f44952b10345 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Got 1 possible topologies {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 874.004366] env[61957]: DEBUG nova.virt.hardware [None req-582b6bd7-3488-48cf-a2f8-f44952b10345 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 874.004366] env[61957]: DEBUG nova.virt.hardware [None req-582b6bd7-3488-48cf-a2f8-f44952b10345 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 874.005376] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d66db97a-d187-437c-bda8-c148ef1ee3a1 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.014195] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d5d57da-1b1b-4a47-ba78-1923bbdebf76 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.243565] env[61957]: DEBUG oslo_concurrency.lockutils [None req-902199c7-5b38-4a0e-a88f-80539210637b tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Lock "526ff179-62a6-4763-ab25-797617c4ed57" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 83.607s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 874.256026] env[61957]: DEBUG oslo_vmware.api [None req-c13ea078-3782-4688-93ee-408e1ac828c0 tempest-ServerTagsTestJSON-969152625 tempest-ServerTagsTestJSON-969152625-project-member] Task: {'id': task-1277587, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.354254] env[61957]: DEBUG nova.network.neutron [-] [instance: 281b9644-0e7d-48b9-a7b6-45fd6102d558] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 874.406293] env[61957]: DEBUG oslo_vmware.api [None req-85c79741-463e-4bc0-9ed2-4682b88fcc00 tempest-ServersV294TestFqdnHostnames-1426486100 tempest-ServersV294TestFqdnHostnames-1426486100-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]5290e9df-4c06-1a8d-6f5a-84776cfc472c, 'name': SearchDatastore_Task, 'duration_secs': 0.026603} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.407259] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c17e42d3-0089-485c-8028-2f25c38c7e57 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.412994] env[61957]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0390fed8-dbf0-467b-bd09-6ab9591d6bce {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.414924] env[61957]: DEBUG oslo_vmware.api [None req-85c79741-463e-4bc0-9ed2-4682b88fcc00 tempest-ServersV294TestFqdnHostnames-1426486100 tempest-ServersV294TestFqdnHostnames-1426486100-project-member] Waiting for the task: (returnval){ [ 874.414924] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]5248ea6d-ded9-6b86-66b0-3c2ae4747f66" [ 874.414924] env[61957]: _type = "Task" [ 874.414924] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.425976] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e601b0b4-4cfc-43fe-9503-f5b0b66f5fea {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.444335] env[61957]: DEBUG oslo_vmware.api [None req-85c79741-463e-4bc0-9ed2-4682b88fcc00 tempest-ServersV294TestFqdnHostnames-1426486100 tempest-ServersV294TestFqdnHostnames-1426486100-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]5248ea6d-ded9-6b86-66b0-3c2ae4747f66, 'name': SearchDatastore_Task, 'duration_secs': 0.008651} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.445421] env[61957]: DEBUG oslo_concurrency.lockutils [None req-85c79741-463e-4bc0-9ed2-4682b88fcc00 tempest-ServersV294TestFqdnHostnames-1426486100 tempest-ServersV294TestFqdnHostnames-1426486100-project-member] Releasing lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 874.445621] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-85c79741-463e-4bc0-9ed2-4682b88fcc00 tempest-ServersV294TestFqdnHostnames-1426486100 tempest-ServersV294TestFqdnHostnames-1426486100-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore1] 2e4a5344-600f-4b61-826e-c15f96b50af2/2e4a5344-600f-4b61-826e-c15f96b50af2.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 874.446073] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4b25b6e3-8e7d-4abd-93c5-902946648645 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.453121] env[61957]: DEBUG oslo_vmware.api [None req-85c79741-463e-4bc0-9ed2-4682b88fcc00 tempest-ServersV294TestFqdnHostnames-1426486100 tempest-ServersV294TestFqdnHostnames-1426486100-project-member] Waiting for the task: (returnval){ [ 874.453121] env[61957]: value = "task-1277588" [ 874.453121] env[61957]: _type = "Task" [ 874.453121] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.464392] env[61957]: DEBUG nova.compute.manager [req-6a8ccb17-5346-494a-a98a-9a16c6d1ef3b req-0b526815-59d3-493b-8035-e496445adadb service nova] [instance: 281b9644-0e7d-48b9-a7b6-45fd6102d558] Detach interface failed, port_id=22b42651-5480-45d9-b363-9106dd6a7ce6, reason: Instance 281b9644-0e7d-48b9-a7b6-45fd6102d558 could not be found. {{(pid=61957) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 874.470129] env[61957]: DEBUG oslo_vmware.api [None req-85c79741-463e-4bc0-9ed2-4682b88fcc00 tempest-ServersV294TestFqdnHostnames-1426486100 tempest-ServersV294TestFqdnHostnames-1426486100-project-member] Task: {'id': task-1277588, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.485299] env[61957]: DEBUG oslo_concurrency.lockutils [None req-06bfc4e2-99e2-4510-98f7-f5b198d6231a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.550s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 874.485883] env[61957]: DEBUG nova.compute.manager [None req-06bfc4e2-99e2-4510-98f7-f5b198d6231a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 1a0327c2-5671-4970-9db7-c7cc912d8678] Start building networks asynchronously for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 874.488682] env[61957]: DEBUG oslo_concurrency.lockutils [None req-afaaf689-a0df-401a-b5c9-bce0445ce0ad tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 26.095s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 874.488924] env[61957]: DEBUG nova.objects.instance [None req-afaaf689-a0df-401a-b5c9-bce0445ce0ad tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Lazy-loading 'resources' on Instance uuid 17f00b20-9d3b-45e6-919d-6fab9999ec77 {{(pid=61957) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 874.688866] env[61957]: DEBUG oslo_concurrency.lockutils [None req-c05cb56e-24b0-42e3-b111-9cb18dbe73d6 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Acquiring lock "interface-4951132e-7247-4772-8f88-3664c6a7e61e-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 874.689187] env[61957]: DEBUG oslo_concurrency.lockutils [None req-c05cb56e-24b0-42e3-b111-9cb18dbe73d6 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Lock "interface-4951132e-7247-4772-8f88-3664c6a7e61e-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 874.690772] env[61957]: DEBUG nova.objects.instance [None req-c05cb56e-24b0-42e3-b111-9cb18dbe73d6 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Lazy-loading 'flavor' on Instance uuid 4951132e-7247-4772-8f88-3664c6a7e61e {{(pid=61957) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 874.715780] env[61957]: DEBUG nova.network.neutron [None req-582b6bd7-3488-48cf-a2f8-f44952b10345 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] [instance: e40fa112-4648-428f-a403-b3e3b8319ea8] Successfully updated port: c407d755-4caa-44d0-b9b2-8e59106cc23e {{(pid=61957) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 874.752895] env[61957]: DEBUG nova.compute.manager [None req-3c689a75-14e7-47a1-b918-cef3025704e3 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: dbeeb200-70b9-4cb4-b5a4-182389d21918] Starting instance... {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 874.761215] env[61957]: DEBUG oslo_vmware.api [None req-c13ea078-3782-4688-93ee-408e1ac828c0 tempest-ServerTagsTestJSON-969152625 tempest-ServerTagsTestJSON-969152625-project-member] Task: {'id': task-1277587, 'name': ReconfigVM_Task, 'duration_secs': 0.543325} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.761571] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-c13ea078-3782-4688-93ee-408e1ac828c0 tempest-ServerTagsTestJSON-969152625 tempest-ServerTagsTestJSON-969152625-project-member] [instance: 8e04a098-25de-4d57-9e3c-ea44d234d57e] Reconfigured VM instance instance-00000042 to attach disk [datastore2] 8e04a098-25de-4d57-9e3c-ea44d234d57e/8e04a098-25de-4d57-9e3c-ea44d234d57e.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 874.762243] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-af692f31-7abf-470a-95ee-ef9b5965f51a {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.769486] env[61957]: DEBUG oslo_vmware.api [None req-c13ea078-3782-4688-93ee-408e1ac828c0 tempest-ServerTagsTestJSON-969152625 tempest-ServerTagsTestJSON-969152625-project-member] Waiting for the task: (returnval){ [ 874.769486] env[61957]: value = "task-1277589" [ 874.769486] env[61957]: _type = "Task" [ 874.769486] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.779693] env[61957]: DEBUG oslo_vmware.api [None req-c13ea078-3782-4688-93ee-408e1ac828c0 tempest-ServerTagsTestJSON-969152625 tempest-ServerTagsTestJSON-969152625-project-member] Task: {'id': task-1277589, 'name': Rename_Task} progress is 5%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.859627] env[61957]: INFO nova.compute.manager [-] [instance: 281b9644-0e7d-48b9-a7b6-45fd6102d558] Took 1.26 seconds to deallocate network for instance. [ 874.963111] env[61957]: DEBUG oslo_vmware.api [None req-85c79741-463e-4bc0-9ed2-4682b88fcc00 tempest-ServersV294TestFqdnHostnames-1426486100 tempest-ServersV294TestFqdnHostnames-1426486100-project-member] Task: {'id': task-1277588, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.997805] env[61957]: DEBUG nova.compute.utils [None req-06bfc4e2-99e2-4510-98f7-f5b198d6231a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Using /dev/sd instead of None {{(pid=61957) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 874.997805] env[61957]: DEBUG nova.compute.manager [None req-06bfc4e2-99e2-4510-98f7-f5b198d6231a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 1a0327c2-5671-4970-9db7-c7cc912d8678] Allocating IP information in the background. {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 874.997805] env[61957]: DEBUG nova.network.neutron [None req-06bfc4e2-99e2-4510-98f7-f5b198d6231a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 1a0327c2-5671-4970-9db7-c7cc912d8678] allocate_for_instance() {{(pid=61957) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 875.039574] env[61957]: DEBUG nova.policy [None req-06bfc4e2-99e2-4510-98f7-f5b198d6231a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e2f13689624b483d903ce12ef290db8b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '65fd71bcd1bf41238a9cc3a5d6dd4924', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61957) authorize /opt/stack/nova/nova/policy.py:203}} [ 875.219848] env[61957]: DEBUG oslo_concurrency.lockutils [None req-582b6bd7-3488-48cf-a2f8-f44952b10345 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Acquiring lock "refresh_cache-e40fa112-4648-428f-a403-b3e3b8319ea8" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 875.220132] env[61957]: DEBUG oslo_concurrency.lockutils [None req-582b6bd7-3488-48cf-a2f8-f44952b10345 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Acquired lock "refresh_cache-e40fa112-4648-428f-a403-b3e3b8319ea8" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 875.220463] env[61957]: DEBUG nova.network.neutron [None req-582b6bd7-3488-48cf-a2f8-f44952b10345 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] [instance: e40fa112-4648-428f-a403-b3e3b8319ea8] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 875.284356] env[61957]: DEBUG oslo_concurrency.lockutils [None req-3c689a75-14e7-47a1-b918-cef3025704e3 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 875.288693] env[61957]: DEBUG oslo_vmware.api [None req-c13ea078-3782-4688-93ee-408e1ac828c0 tempest-ServerTagsTestJSON-969152625 tempest-ServerTagsTestJSON-969152625-project-member] Task: {'id': task-1277589, 'name': Rename_Task, 'duration_secs': 0.244132} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.293467] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-c13ea078-3782-4688-93ee-408e1ac828c0 tempest-ServerTagsTestJSON-969152625 tempest-ServerTagsTestJSON-969152625-project-member] [instance: 8e04a098-25de-4d57-9e3c-ea44d234d57e] Powering on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 875.293874] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-66a4a280-c193-46f7-86bf-7f19cde12685 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.300893] env[61957]: DEBUG oslo_vmware.api [None req-c13ea078-3782-4688-93ee-408e1ac828c0 tempest-ServerTagsTestJSON-969152625 tempest-ServerTagsTestJSON-969152625-project-member] Waiting for the task: (returnval){ [ 875.300893] env[61957]: value = "task-1277590" [ 875.300893] env[61957]: _type = "Task" [ 875.300893] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.310979] env[61957]: DEBUG oslo_vmware.api [None req-c13ea078-3782-4688-93ee-408e1ac828c0 tempest-ServerTagsTestJSON-969152625 tempest-ServerTagsTestJSON-969152625-project-member] Task: {'id': task-1277590, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.354102] env[61957]: DEBUG nova.objects.instance [None req-c05cb56e-24b0-42e3-b111-9cb18dbe73d6 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Lazy-loading 'pci_requests' on Instance uuid 4951132e-7247-4772-8f88-3664c6a7e61e {{(pid=61957) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 875.369307] env[61957]: DEBUG oslo_concurrency.lockutils [None req-fd686081-caee-4453-9fa7-68efd0fac70e tempest-AttachInterfacesV270Test-2032541732 tempest-AttachInterfacesV270Test-2032541732-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 875.382197] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65538013-8d0a-4ed2-ab4d-1f95f61f7f6f {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.394813] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05416664-d446-4810-bbb3-05fa5506b0cb {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.430580] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-861029cd-a295-4693-bcb4-b89d4ebaae71 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.435929] env[61957]: DEBUG nova.compute.manager [req-351d0206-918a-42bb-bb10-4c76f0c25e68 req-238dd5bc-6398-4c71-a106-b8ed86f10b12 service nova] [instance: 526ff179-62a6-4763-ab25-797617c4ed57] Received event network-changed-25abeea6-dd92-49e0-b014-d35dd1c0f8ee {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 875.435929] env[61957]: DEBUG nova.compute.manager [req-351d0206-918a-42bb-bb10-4c76f0c25e68 req-238dd5bc-6398-4c71-a106-b8ed86f10b12 service nova] [instance: 526ff179-62a6-4763-ab25-797617c4ed57] Refreshing instance network info cache due to event network-changed-25abeea6-dd92-49e0-b014-d35dd1c0f8ee. {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 875.435929] env[61957]: DEBUG oslo_concurrency.lockutils [req-351d0206-918a-42bb-bb10-4c76f0c25e68 req-238dd5bc-6398-4c71-a106-b8ed86f10b12 service nova] Acquiring lock "refresh_cache-526ff179-62a6-4763-ab25-797617c4ed57" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 875.435929] env[61957]: DEBUG oslo_concurrency.lockutils [req-351d0206-918a-42bb-bb10-4c76f0c25e68 req-238dd5bc-6398-4c71-a106-b8ed86f10b12 service nova] Acquired lock "refresh_cache-526ff179-62a6-4763-ab25-797617c4ed57" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 875.436556] env[61957]: DEBUG nova.network.neutron [req-351d0206-918a-42bb-bb10-4c76f0c25e68 req-238dd5bc-6398-4c71-a106-b8ed86f10b12 service nova] [instance: 526ff179-62a6-4763-ab25-797617c4ed57] Refreshing network info cache for port 25abeea6-dd92-49e0-b014-d35dd1c0f8ee {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 875.438995] env[61957]: DEBUG nova.network.neutron [None req-06bfc4e2-99e2-4510-98f7-f5b198d6231a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 1a0327c2-5671-4970-9db7-c7cc912d8678] Successfully created port: e84fe58e-f595-4b03-9e8a-88238d003168 {{(pid=61957) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 875.448154] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19b9f2bc-4641-49cc-9f86-8b68259009c3 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.466554] env[61957]: DEBUG nova.compute.provider_tree [None req-afaaf689-a0df-401a-b5c9-bce0445ce0ad tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 875.474409] env[61957]: DEBUG oslo_vmware.api [None req-85c79741-463e-4bc0-9ed2-4682b88fcc00 tempest-ServersV294TestFqdnHostnames-1426486100 tempest-ServersV294TestFqdnHostnames-1426486100-project-member] Task: {'id': task-1277588, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.5228} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.474738] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-85c79741-463e-4bc0-9ed2-4682b88fcc00 tempest-ServersV294TestFqdnHostnames-1426486100 tempest-ServersV294TestFqdnHostnames-1426486100-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore1] 2e4a5344-600f-4b61-826e-c15f96b50af2/2e4a5344-600f-4b61-826e-c15f96b50af2.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 875.475009] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-85c79741-463e-4bc0-9ed2-4682b88fcc00 tempest-ServersV294TestFqdnHostnames-1426486100 tempest-ServersV294TestFqdnHostnames-1426486100-project-member] [instance: 2e4a5344-600f-4b61-826e-c15f96b50af2] Extending root virtual disk to 1048576 {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 875.475329] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2687f566-3fd8-45c7-98c0-9622d5e14150 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.484421] env[61957]: DEBUG oslo_vmware.api [None req-85c79741-463e-4bc0-9ed2-4682b88fcc00 tempest-ServersV294TestFqdnHostnames-1426486100 tempest-ServersV294TestFqdnHostnames-1426486100-project-member] Waiting for the task: (returnval){ [ 875.484421] env[61957]: value = "task-1277591" [ 875.484421] env[61957]: _type = "Task" [ 875.484421] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.494448] env[61957]: DEBUG oslo_vmware.api [None req-85c79741-463e-4bc0-9ed2-4682b88fcc00 tempest-ServersV294TestFqdnHostnames-1426486100 tempest-ServersV294TestFqdnHostnames-1426486100-project-member] Task: {'id': task-1277591, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.503123] env[61957]: DEBUG nova.compute.manager [None req-06bfc4e2-99e2-4510-98f7-f5b198d6231a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 1a0327c2-5671-4970-9db7-c7cc912d8678] Start building block device mappings for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 875.769110] env[61957]: DEBUG nova.network.neutron [None req-582b6bd7-3488-48cf-a2f8-f44952b10345 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] [instance: e40fa112-4648-428f-a403-b3e3b8319ea8] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 875.811786] env[61957]: DEBUG oslo_vmware.api [None req-c13ea078-3782-4688-93ee-408e1ac828c0 tempest-ServerTagsTestJSON-969152625 tempest-ServerTagsTestJSON-969152625-project-member] Task: {'id': task-1277590, 'name': PowerOnVM_Task, 'duration_secs': 0.464405} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.814643] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-c13ea078-3782-4688-93ee-408e1ac828c0 tempest-ServerTagsTestJSON-969152625 tempest-ServerTagsTestJSON-969152625-project-member] [instance: 8e04a098-25de-4d57-9e3c-ea44d234d57e] Powered on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 875.814643] env[61957]: INFO nova.compute.manager [None req-c13ea078-3782-4688-93ee-408e1ac828c0 tempest-ServerTagsTestJSON-969152625 tempest-ServerTagsTestJSON-969152625-project-member] [instance: 8e04a098-25de-4d57-9e3c-ea44d234d57e] Took 9.07 seconds to spawn the instance on the hypervisor. [ 875.814823] env[61957]: DEBUG nova.compute.manager [None req-c13ea078-3782-4688-93ee-408e1ac828c0 tempest-ServerTagsTestJSON-969152625 tempest-ServerTagsTestJSON-969152625-project-member] [instance: 8e04a098-25de-4d57-9e3c-ea44d234d57e] Checking state {{(pid=61957) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 875.815785] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e58b7ab-cdbb-44cd-959e-002892371444 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.856427] env[61957]: DEBUG nova.objects.base [None req-c05cb56e-24b0-42e3-b111-9cb18dbe73d6 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Object Instance<4951132e-7247-4772-8f88-3664c6a7e61e> lazy-loaded attributes: flavor,pci_requests {{(pid=61957) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 875.856703] env[61957]: DEBUG nova.network.neutron [None req-c05cb56e-24b0-42e3-b111-9cb18dbe73d6 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 4951132e-7247-4772-8f88-3664c6a7e61e] allocate_for_instance() {{(pid=61957) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 875.923058] env[61957]: DEBUG nova.policy [None req-c05cb56e-24b0-42e3-b111-9cb18dbe73d6 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6fd1c19dc3a44212ada44445e0919106', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8713aa35bcb24b86ad0b58ca9fc991ba', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61957) authorize /opt/stack/nova/nova/policy.py:203}} [ 875.970151] env[61957]: DEBUG nova.scheduler.client.report [None req-afaaf689-a0df-401a-b5c9-bce0445ce0ad tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 875.991137] env[61957]: DEBUG nova.network.neutron [None req-582b6bd7-3488-48cf-a2f8-f44952b10345 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] [instance: e40fa112-4648-428f-a403-b3e3b8319ea8] Updating instance_info_cache with network_info: [{"id": "c407d755-4caa-44d0-b9b2-8e59106cc23e", "address": "fa:16:3e:db:0f:8c", "network": {"id": "41a1701c-d9b6-4414-9e92-2e999cc916e3", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-727135203-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d849d922cfad41848a33e8d3bc83ea0f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f1b507ed-cd2d-4c09-9d96-c47bde6a7774", "external-id": "nsx-vlan-transportzone-980", "segmentation_id": 980, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc407d755-4c", "ovs_interfaceid": "c407d755-4caa-44d0-b9b2-8e59106cc23e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 875.996090] env[61957]: DEBUG oslo_vmware.api [None req-85c79741-463e-4bc0-9ed2-4682b88fcc00 tempest-ServersV294TestFqdnHostnames-1426486100 tempest-ServersV294TestFqdnHostnames-1426486100-project-member] Task: {'id': task-1277591, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.08516} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.998901] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-85c79741-463e-4bc0-9ed2-4682b88fcc00 tempest-ServersV294TestFqdnHostnames-1426486100 tempest-ServersV294TestFqdnHostnames-1426486100-project-member] [instance: 2e4a5344-600f-4b61-826e-c15f96b50af2] Extended root virtual disk {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 876.000139] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3e3611f-baa9-4ee9-8f6d-639c8bd863df {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.027656] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-85c79741-463e-4bc0-9ed2-4682b88fcc00 tempest-ServersV294TestFqdnHostnames-1426486100 tempest-ServersV294TestFqdnHostnames-1426486100-project-member] [instance: 2e4a5344-600f-4b61-826e-c15f96b50af2] Reconfiguring VM instance instance-00000043 to attach disk [datastore1] 2e4a5344-600f-4b61-826e-c15f96b50af2/2e4a5344-600f-4b61-826e-c15f96b50af2.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 876.027936] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ba43fd79-7094-4f09-a9b8-95a22ea6496d {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.051997] env[61957]: DEBUG oslo_vmware.api [None req-85c79741-463e-4bc0-9ed2-4682b88fcc00 tempest-ServersV294TestFqdnHostnames-1426486100 tempest-ServersV294TestFqdnHostnames-1426486100-project-member] Waiting for the task: (returnval){ [ 876.051997] env[61957]: value = "task-1277592" [ 876.051997] env[61957]: _type = "Task" [ 876.051997] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.060320] env[61957]: DEBUG oslo_vmware.api [None req-85c79741-463e-4bc0-9ed2-4682b88fcc00 tempest-ServersV294TestFqdnHostnames-1426486100 tempest-ServersV294TestFqdnHostnames-1426486100-project-member] Task: {'id': task-1277592, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.200155] env[61957]: DEBUG nova.compute.manager [req-27ba2647-c9db-4ee4-9931-203f6f4b6680 req-6ed7594f-e60f-4d5d-b80a-9e1197f4976f service nova] [instance: e40fa112-4648-428f-a403-b3e3b8319ea8] Received event network-vif-plugged-c407d755-4caa-44d0-b9b2-8e59106cc23e {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 876.200422] env[61957]: DEBUG oslo_concurrency.lockutils [req-27ba2647-c9db-4ee4-9931-203f6f4b6680 req-6ed7594f-e60f-4d5d-b80a-9e1197f4976f service nova] Acquiring lock "e40fa112-4648-428f-a403-b3e3b8319ea8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 876.200688] env[61957]: DEBUG oslo_concurrency.lockutils [req-27ba2647-c9db-4ee4-9931-203f6f4b6680 req-6ed7594f-e60f-4d5d-b80a-9e1197f4976f service nova] Lock "e40fa112-4648-428f-a403-b3e3b8319ea8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 876.200887] env[61957]: DEBUG oslo_concurrency.lockutils [req-27ba2647-c9db-4ee4-9931-203f6f4b6680 req-6ed7594f-e60f-4d5d-b80a-9e1197f4976f service nova] Lock "e40fa112-4648-428f-a403-b3e3b8319ea8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 876.201219] env[61957]: DEBUG nova.compute.manager [req-27ba2647-c9db-4ee4-9931-203f6f4b6680 req-6ed7594f-e60f-4d5d-b80a-9e1197f4976f service nova] [instance: e40fa112-4648-428f-a403-b3e3b8319ea8] No waiting events found dispatching network-vif-plugged-c407d755-4caa-44d0-b9b2-8e59106cc23e {{(pid=61957) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 876.201709] env[61957]: WARNING nova.compute.manager [req-27ba2647-c9db-4ee4-9931-203f6f4b6680 req-6ed7594f-e60f-4d5d-b80a-9e1197f4976f service nova] [instance: e40fa112-4648-428f-a403-b3e3b8319ea8] Received unexpected event network-vif-plugged-c407d755-4caa-44d0-b9b2-8e59106cc23e for instance with vm_state building and task_state spawning. [ 876.201709] env[61957]: DEBUG nova.compute.manager [req-27ba2647-c9db-4ee4-9931-203f6f4b6680 req-6ed7594f-e60f-4d5d-b80a-9e1197f4976f service nova] [instance: e40fa112-4648-428f-a403-b3e3b8319ea8] Received event network-changed-c407d755-4caa-44d0-b9b2-8e59106cc23e {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 876.201942] env[61957]: DEBUG nova.compute.manager [req-27ba2647-c9db-4ee4-9931-203f6f4b6680 req-6ed7594f-e60f-4d5d-b80a-9e1197f4976f service nova] [instance: e40fa112-4648-428f-a403-b3e3b8319ea8] Refreshing instance network info cache due to event network-changed-c407d755-4caa-44d0-b9b2-8e59106cc23e. {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 876.202109] env[61957]: DEBUG oslo_concurrency.lockutils [req-27ba2647-c9db-4ee4-9931-203f6f4b6680 req-6ed7594f-e60f-4d5d-b80a-9e1197f4976f service nova] Acquiring lock "refresh_cache-e40fa112-4648-428f-a403-b3e3b8319ea8" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 876.337257] env[61957]: INFO nova.compute.manager [None req-c13ea078-3782-4688-93ee-408e1ac828c0 tempest-ServerTagsTestJSON-969152625 tempest-ServerTagsTestJSON-969152625-project-member] [instance: 8e04a098-25de-4d57-9e3c-ea44d234d57e] Took 38.15 seconds to build instance. [ 876.347760] env[61957]: DEBUG nova.network.neutron [req-351d0206-918a-42bb-bb10-4c76f0c25e68 req-238dd5bc-6398-4c71-a106-b8ed86f10b12 service nova] [instance: 526ff179-62a6-4763-ab25-797617c4ed57] Updated VIF entry in instance network info cache for port 25abeea6-dd92-49e0-b014-d35dd1c0f8ee. {{(pid=61957) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 876.348370] env[61957]: DEBUG nova.network.neutron [req-351d0206-918a-42bb-bb10-4c76f0c25e68 req-238dd5bc-6398-4c71-a106-b8ed86f10b12 service nova] [instance: 526ff179-62a6-4763-ab25-797617c4ed57] Updating instance_info_cache with network_info: [{"id": "25abeea6-dd92-49e0-b014-d35dd1c0f8ee", "address": "fa:16:3e:0a:44:84", "network": {"id": "8e60ab13-6dd3-4cda-9cdf-974ad8bebe1f", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1404238496-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.173", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ea955bd4d84548ff8e6a0c21542d6ee7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9f208df-1fb5-4403-9796-7fd19e4bfb85", "external-id": "cl2-zone-400", "segmentation_id": 400, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap25abeea6-dd", "ovs_interfaceid": "25abeea6-dd92-49e0-b014-d35dd1c0f8ee", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 876.398998] env[61957]: DEBUG nova.network.neutron [None req-c05cb56e-24b0-42e3-b111-9cb18dbe73d6 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 4951132e-7247-4772-8f88-3664c6a7e61e] Successfully created port: a093d1c9-1f94-4000-8f87-f76e9db98ec8 {{(pid=61957) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 876.475890] env[61957]: DEBUG oslo_concurrency.lockutils [None req-afaaf689-a0df-401a-b5c9-bce0445ce0ad tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.987s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 876.478976] env[61957]: DEBUG oslo_concurrency.lockutils [None req-0d2ed604-6dce-44bf-b238-71a34c40662a tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 24.475s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 876.478976] env[61957]: DEBUG nova.objects.instance [None req-0d2ed604-6dce-44bf-b238-71a34c40662a tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] Lazy-loading 'resources' on Instance uuid 37a9e1c1-9f6c-4047-9ff5-e141d0ca383e {{(pid=61957) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 876.496537] env[61957]: DEBUG oslo_concurrency.lockutils [None req-582b6bd7-3488-48cf-a2f8-f44952b10345 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Releasing lock "refresh_cache-e40fa112-4648-428f-a403-b3e3b8319ea8" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 876.496685] env[61957]: DEBUG nova.compute.manager [None req-582b6bd7-3488-48cf-a2f8-f44952b10345 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] [instance: e40fa112-4648-428f-a403-b3e3b8319ea8] Instance network_info: |[{"id": "c407d755-4caa-44d0-b9b2-8e59106cc23e", "address": "fa:16:3e:db:0f:8c", "network": {"id": "41a1701c-d9b6-4414-9e92-2e999cc916e3", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-727135203-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d849d922cfad41848a33e8d3bc83ea0f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f1b507ed-cd2d-4c09-9d96-c47bde6a7774", "external-id": "nsx-vlan-transportzone-980", "segmentation_id": 980, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc407d755-4c", "ovs_interfaceid": "c407d755-4caa-44d0-b9b2-8e59106cc23e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 876.497840] env[61957]: DEBUG oslo_concurrency.lockutils [req-27ba2647-c9db-4ee4-9931-203f6f4b6680 req-6ed7594f-e60f-4d5d-b80a-9e1197f4976f service nova] Acquired lock "refresh_cache-e40fa112-4648-428f-a403-b3e3b8319ea8" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 876.497840] env[61957]: DEBUG nova.network.neutron [req-27ba2647-c9db-4ee4-9931-203f6f4b6680 req-6ed7594f-e60f-4d5d-b80a-9e1197f4976f service nova] [instance: e40fa112-4648-428f-a403-b3e3b8319ea8] Refreshing network info cache for port c407d755-4caa-44d0-b9b2-8e59106cc23e {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 876.498468] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-582b6bd7-3488-48cf-a2f8-f44952b10345 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] [instance: e40fa112-4648-428f-a403-b3e3b8319ea8] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:db:0f:8c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f1b507ed-cd2d-4c09-9d96-c47bde6a7774', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c407d755-4caa-44d0-b9b2-8e59106cc23e', 'vif_model': 'vmxnet3'}] {{(pid=61957) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 876.506280] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-582b6bd7-3488-48cf-a2f8-f44952b10345 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Creating folder: Project (d849d922cfad41848a33e8d3bc83ea0f). Parent ref: group-v274445. {{(pid=61957) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 876.508340] env[61957]: INFO nova.scheduler.client.report [None req-afaaf689-a0df-401a-b5c9-bce0445ce0ad tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Deleted allocations for instance 17f00b20-9d3b-45e6-919d-6fab9999ec77 [ 876.509254] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d44c08b9-bade-4736-ae15-da132f1665b4 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.523258] env[61957]: INFO nova.virt.vmwareapi.vm_util [None req-582b6bd7-3488-48cf-a2f8-f44952b10345 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Created folder: Project (d849d922cfad41848a33e8d3bc83ea0f) in parent group-v274445. [ 876.523473] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-582b6bd7-3488-48cf-a2f8-f44952b10345 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Creating folder: Instances. Parent ref: group-v274529. {{(pid=61957) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 876.523701] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2dffb506-6fc7-4c53-853f-c1479f2ee4c1 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.528951] env[61957]: DEBUG nova.compute.manager [None req-06bfc4e2-99e2-4510-98f7-f5b198d6231a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 1a0327c2-5671-4970-9db7-c7cc912d8678] Start spawning the instance on the hypervisor. {{(pid=61957) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 876.533162] env[61957]: INFO nova.virt.vmwareapi.vm_util [None req-582b6bd7-3488-48cf-a2f8-f44952b10345 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Created folder: Instances in parent group-v274529. [ 876.533409] env[61957]: DEBUG oslo.service.loopingcall [None req-582b6bd7-3488-48cf-a2f8-f44952b10345 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 876.533599] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e40fa112-4648-428f-a403-b3e3b8319ea8] Creating VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 876.533815] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e7cbf8af-1a87-47a3-9e9f-09f9a8fd516f {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.553713] env[61957]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 876.553713] env[61957]: value = "task-1277595" [ 876.553713] env[61957]: _type = "Task" [ 876.553713] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.565748] env[61957]: DEBUG nova.virt.hardware [None req-06bfc4e2-99e2-4510-98f7-f5b198d6231a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T17:22:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T17:21:50Z,direct_url=,disk_format='vmdk',id=11c76a2c-f705-470a-ba9d-4657858bab38,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='81d3096d0f094373913fc3dc8f5c3c6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T17:21:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 876.566094] env[61957]: DEBUG nova.virt.hardware [None req-06bfc4e2-99e2-4510-98f7-f5b198d6231a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Flavor limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 876.566333] env[61957]: DEBUG nova.virt.hardware [None req-06bfc4e2-99e2-4510-98f7-f5b198d6231a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Image limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 876.566436] env[61957]: DEBUG nova.virt.hardware [None req-06bfc4e2-99e2-4510-98f7-f5b198d6231a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Flavor pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 876.566627] env[61957]: DEBUG nova.virt.hardware [None req-06bfc4e2-99e2-4510-98f7-f5b198d6231a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Image pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 876.566847] env[61957]: DEBUG nova.virt.hardware [None req-06bfc4e2-99e2-4510-98f7-f5b198d6231a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 876.567209] env[61957]: DEBUG nova.virt.hardware [None req-06bfc4e2-99e2-4510-98f7-f5b198d6231a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 876.567377] env[61957]: DEBUG nova.virt.hardware [None req-06bfc4e2-99e2-4510-98f7-f5b198d6231a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 876.567607] env[61957]: DEBUG nova.virt.hardware [None req-06bfc4e2-99e2-4510-98f7-f5b198d6231a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Got 1 possible topologies {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 876.567805] env[61957]: DEBUG nova.virt.hardware [None req-06bfc4e2-99e2-4510-98f7-f5b198d6231a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 876.567990] env[61957]: DEBUG nova.virt.hardware [None req-06bfc4e2-99e2-4510-98f7-f5b198d6231a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 876.568865] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf94a04c-34d6-4e46-a424-5b3c95d5ede4 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.577926] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277595, 'name': CreateVM_Task} progress is 5%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.578381] env[61957]: DEBUG oslo_vmware.api [None req-85c79741-463e-4bc0-9ed2-4682b88fcc00 tempest-ServersV294TestFqdnHostnames-1426486100 tempest-ServersV294TestFqdnHostnames-1426486100-project-member] Task: {'id': task-1277592, 'name': ReconfigVM_Task, 'duration_secs': 0.386278} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.579011] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-85c79741-463e-4bc0-9ed2-4682b88fcc00 tempest-ServersV294TestFqdnHostnames-1426486100 tempest-ServersV294TestFqdnHostnames-1426486100-project-member] [instance: 2e4a5344-600f-4b61-826e-c15f96b50af2] Reconfigured VM instance instance-00000043 to attach disk [datastore1] 2e4a5344-600f-4b61-826e-c15f96b50af2/2e4a5344-600f-4b61-826e-c15f96b50af2.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 876.579621] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-36cf30cf-cc3f-4c6f-99ca-a0f3f1aee40c {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.585068] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-031fa3b1-f637-4e1e-bd10-5a496fe70692 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.590758] env[61957]: DEBUG oslo_vmware.api [None req-85c79741-463e-4bc0-9ed2-4682b88fcc00 tempest-ServersV294TestFqdnHostnames-1426486100 tempest-ServersV294TestFqdnHostnames-1426486100-project-member] Waiting for the task: (returnval){ [ 876.590758] env[61957]: value = "task-1277596" [ 876.590758] env[61957]: _type = "Task" [ 876.590758] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.607407] env[61957]: DEBUG oslo_vmware.api [None req-85c79741-463e-4bc0-9ed2-4682b88fcc00 tempest-ServersV294TestFqdnHostnames-1426486100 tempest-ServersV294TestFqdnHostnames-1426486100-project-member] Task: {'id': task-1277596, 'name': Rename_Task} progress is 5%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.840730] env[61957]: DEBUG oslo_concurrency.lockutils [None req-c13ea078-3782-4688-93ee-408e1ac828c0 tempest-ServerTagsTestJSON-969152625 tempest-ServerTagsTestJSON-969152625-project-member] Lock "8e04a098-25de-4d57-9e3c-ea44d234d57e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 69.065s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 876.850859] env[61957]: DEBUG oslo_concurrency.lockutils [req-351d0206-918a-42bb-bb10-4c76f0c25e68 req-238dd5bc-6398-4c71-a106-b8ed86f10b12 service nova] Releasing lock "refresh_cache-526ff179-62a6-4763-ab25-797617c4ed57" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 877.024823] env[61957]: DEBUG oslo_concurrency.lockutils [None req-afaaf689-a0df-401a-b5c9-bce0445ce0ad tempest-ServerShowV254Test-1330700790 tempest-ServerShowV254Test-1330700790-project-member] Lock "17f00b20-9d3b-45e6-919d-6fab9999ec77" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 32.353s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 877.074122] env[61957]: DEBUG nova.network.neutron [None req-06bfc4e2-99e2-4510-98f7-f5b198d6231a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 1a0327c2-5671-4970-9db7-c7cc912d8678] Successfully updated port: e84fe58e-f595-4b03-9e8a-88238d003168 {{(pid=61957) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 877.095078] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277595, 'name': CreateVM_Task, 'duration_secs': 0.38282} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.100049] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e40fa112-4648-428f-a403-b3e3b8319ea8] Created VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 877.105419] env[61957]: DEBUG oslo_concurrency.lockutils [None req-582b6bd7-3488-48cf-a2f8-f44952b10345 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 877.105637] env[61957]: DEBUG oslo_concurrency.lockutils [None req-582b6bd7-3488-48cf-a2f8-f44952b10345 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 877.105992] env[61957]: DEBUG oslo_concurrency.lockutils [None req-582b6bd7-3488-48cf-a2f8-f44952b10345 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 877.107252] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cbe7be6e-9d2e-4c36-972c-e48cf2745df1 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.113489] env[61957]: DEBUG oslo_vmware.api [None req-85c79741-463e-4bc0-9ed2-4682b88fcc00 tempest-ServersV294TestFqdnHostnames-1426486100 tempest-ServersV294TestFqdnHostnames-1426486100-project-member] Task: {'id': task-1277596, 'name': Rename_Task, 'duration_secs': 0.189481} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.114139] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-85c79741-463e-4bc0-9ed2-4682b88fcc00 tempest-ServersV294TestFqdnHostnames-1426486100 tempest-ServersV294TestFqdnHostnames-1426486100-project-member] [instance: 2e4a5344-600f-4b61-826e-c15f96b50af2] Powering on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 877.114471] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4ddd380c-552a-47b7-a764-462ff6d273b9 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.117473] env[61957]: DEBUG oslo_vmware.api [None req-582b6bd7-3488-48cf-a2f8-f44952b10345 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Waiting for the task: (returnval){ [ 877.117473] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52a2fd54-1848-a90f-d93b-2471a7ff54c7" [ 877.117473] env[61957]: _type = "Task" [ 877.117473] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.129226] env[61957]: DEBUG oslo_vmware.api [None req-85c79741-463e-4bc0-9ed2-4682b88fcc00 tempest-ServersV294TestFqdnHostnames-1426486100 tempest-ServersV294TestFqdnHostnames-1426486100-project-member] Waiting for the task: (returnval){ [ 877.129226] env[61957]: value = "task-1277597" [ 877.129226] env[61957]: _type = "Task" [ 877.129226] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.137774] env[61957]: DEBUG oslo_vmware.api [None req-582b6bd7-3488-48cf-a2f8-f44952b10345 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52a2fd54-1848-a90f-d93b-2471a7ff54c7, 'name': SearchDatastore_Task, 'duration_secs': 0.014721} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.138621] env[61957]: DEBUG oslo_concurrency.lockutils [None req-582b6bd7-3488-48cf-a2f8-f44952b10345 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 877.138947] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-582b6bd7-3488-48cf-a2f8-f44952b10345 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] [instance: e40fa112-4648-428f-a403-b3e3b8319ea8] Processing image 11c76a2c-f705-470a-ba9d-4657858bab38 {{(pid=61957) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 877.139290] env[61957]: DEBUG oslo_concurrency.lockutils [None req-582b6bd7-3488-48cf-a2f8-f44952b10345 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 877.139527] env[61957]: DEBUG oslo_concurrency.lockutils [None req-582b6bd7-3488-48cf-a2f8-f44952b10345 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 877.139781] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-582b6bd7-3488-48cf-a2f8-f44952b10345 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 877.145831] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-cdd69142-18af-45a8-ad8a-94c925c85cbc {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.150670] env[61957]: DEBUG oslo_vmware.api [None req-85c79741-463e-4bc0-9ed2-4682b88fcc00 tempest-ServersV294TestFqdnHostnames-1426486100 tempest-ServersV294TestFqdnHostnames-1426486100-project-member] Task: {'id': task-1277597, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.166524] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-582b6bd7-3488-48cf-a2f8-f44952b10345 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 877.166524] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-582b6bd7-3488-48cf-a2f8-f44952b10345 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61957) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 877.166940] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-963d4bfa-f04b-4efa-aab1-44e06333a6a2 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.174711] env[61957]: DEBUG oslo_vmware.api [None req-582b6bd7-3488-48cf-a2f8-f44952b10345 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Waiting for the task: (returnval){ [ 877.174711] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52749164-281f-1a60-b837-26fc9d17a6c9" [ 877.174711] env[61957]: _type = "Task" [ 877.174711] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.193903] env[61957]: DEBUG oslo_vmware.api [None req-582b6bd7-3488-48cf-a2f8-f44952b10345 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52749164-281f-1a60-b837-26fc9d17a6c9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.325411] env[61957]: DEBUG nova.network.neutron [req-27ba2647-c9db-4ee4-9931-203f6f4b6680 req-6ed7594f-e60f-4d5d-b80a-9e1197f4976f service nova] [instance: e40fa112-4648-428f-a403-b3e3b8319ea8] Updated VIF entry in instance network info cache for port c407d755-4caa-44d0-b9b2-8e59106cc23e. {{(pid=61957) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 877.325786] env[61957]: DEBUG nova.network.neutron [req-27ba2647-c9db-4ee4-9931-203f6f4b6680 req-6ed7594f-e60f-4d5d-b80a-9e1197f4976f service nova] [instance: e40fa112-4648-428f-a403-b3e3b8319ea8] Updating instance_info_cache with network_info: [{"id": "c407d755-4caa-44d0-b9b2-8e59106cc23e", "address": "fa:16:3e:db:0f:8c", "network": {"id": "41a1701c-d9b6-4414-9e92-2e999cc916e3", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-727135203-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d849d922cfad41848a33e8d3bc83ea0f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f1b507ed-cd2d-4c09-9d96-c47bde6a7774", "external-id": "nsx-vlan-transportzone-980", "segmentation_id": 980, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc407d755-4c", "ovs_interfaceid": "c407d755-4caa-44d0-b9b2-8e59106cc23e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 877.343048] env[61957]: DEBUG nova.compute.manager [None req-a7399dbc-0983-4a91-82d1-9c1b173d4f53 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 39b8b1df-efdd-4c22-9f31-85c85be4f0eb] Starting instance... {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 877.400115] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c841d852-f89b-4505-b491-672a50357662 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.408440] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c126070-ad16-4245-a8f6-96af33c8883d {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.441969] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90d25aff-c991-4a59-ba94-7e36b94e4ef0 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.451844] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04f1a4b1-c630-4a9a-9494-79f1a3100c26 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.466359] env[61957]: DEBUG nova.compute.provider_tree [None req-0d2ed604-6dce-44bf-b238-71a34c40662a tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 877.477649] env[61957]: DEBUG oslo_vmware.rw_handles [None req-22255302-8924-4370-98de-304a0c0dd7ba tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5210ccde-3f01-e031-c858-b8a459135f72/disk-0.vmdk. {{(pid=61957) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 877.478693] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8036c0d-e12d-435e-9106-e1fc835f7166 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.487185] env[61957]: DEBUG oslo_vmware.rw_handles [None req-22255302-8924-4370-98de-304a0c0dd7ba tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5210ccde-3f01-e031-c858-b8a459135f72/disk-0.vmdk is in state: ready. {{(pid=61957) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 877.487260] env[61957]: ERROR oslo_vmware.rw_handles [None req-22255302-8924-4370-98de-304a0c0dd7ba tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5210ccde-3f01-e031-c858-b8a459135f72/disk-0.vmdk due to incomplete transfer. [ 877.487705] env[61957]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-5589d14f-cc1c-41a8-9220-cfcbcbb3cbfa {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.496116] env[61957]: DEBUG oslo_vmware.rw_handles [None req-22255302-8924-4370-98de-304a0c0dd7ba tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5210ccde-3f01-e031-c858-b8a459135f72/disk-0.vmdk. {{(pid=61957) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 877.496392] env[61957]: DEBUG nova.virt.vmwareapi.images [None req-22255302-8924-4370-98de-304a0c0dd7ba tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: 0adae8e1-8c2f-4110-805b-1f286debc833] Uploaded image 64157260-64c5-4eb6-9410-9be776e1d28d to the Glance image server {{(pid=61957) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 877.498992] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-22255302-8924-4370-98de-304a0c0dd7ba tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: 0adae8e1-8c2f-4110-805b-1f286debc833] Destroying the VM {{(pid=61957) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 877.498992] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-555f3d71-385b-46af-a781-aec14897d03e {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.505550] env[61957]: DEBUG oslo_vmware.api [None req-22255302-8924-4370-98de-304a0c0dd7ba tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Waiting for the task: (returnval){ [ 877.505550] env[61957]: value = "task-1277598" [ 877.505550] env[61957]: _type = "Task" [ 877.505550] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.514305] env[61957]: DEBUG oslo_vmware.api [None req-22255302-8924-4370-98de-304a0c0dd7ba tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': task-1277598, 'name': Destroy_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.585768] env[61957]: DEBUG oslo_concurrency.lockutils [None req-06bfc4e2-99e2-4510-98f7-f5b198d6231a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Acquiring lock "refresh_cache-1a0327c2-5671-4970-9db7-c7cc912d8678" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 877.585768] env[61957]: DEBUG oslo_concurrency.lockutils [None req-06bfc4e2-99e2-4510-98f7-f5b198d6231a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Acquired lock "refresh_cache-1a0327c2-5671-4970-9db7-c7cc912d8678" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 877.585768] env[61957]: DEBUG nova.network.neutron [None req-06bfc4e2-99e2-4510-98f7-f5b198d6231a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 1a0327c2-5671-4970-9db7-c7cc912d8678] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 877.641989] env[61957]: DEBUG oslo_vmware.api [None req-85c79741-463e-4bc0-9ed2-4682b88fcc00 tempest-ServersV294TestFqdnHostnames-1426486100 tempest-ServersV294TestFqdnHostnames-1426486100-project-member] Task: {'id': task-1277597, 'name': PowerOnVM_Task} progress is 94%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.688859] env[61957]: DEBUG oslo_vmware.api [None req-582b6bd7-3488-48cf-a2f8-f44952b10345 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52749164-281f-1a60-b837-26fc9d17a6c9, 'name': SearchDatastore_Task, 'duration_secs': 0.021602} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.690027] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cf359d22-f04d-4685-b520-3851d1b26429 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.696394] env[61957]: DEBUG oslo_vmware.api [None req-582b6bd7-3488-48cf-a2f8-f44952b10345 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Waiting for the task: (returnval){ [ 877.696394] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]528623ed-91c2-0f04-d4be-16747dcb2a70" [ 877.696394] env[61957]: _type = "Task" [ 877.696394] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.705886] env[61957]: DEBUG oslo_vmware.api [None req-582b6bd7-3488-48cf-a2f8-f44952b10345 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]528623ed-91c2-0f04-d4be-16747dcb2a70, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.830728] env[61957]: DEBUG oslo_concurrency.lockutils [req-27ba2647-c9db-4ee4-9931-203f6f4b6680 req-6ed7594f-e60f-4d5d-b80a-9e1197f4976f service nova] Releasing lock "refresh_cache-e40fa112-4648-428f-a403-b3e3b8319ea8" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 877.870078] env[61957]: DEBUG oslo_concurrency.lockutils [None req-a7399dbc-0983-4a91-82d1-9c1b173d4f53 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 877.938742] env[61957]: DEBUG nova.network.neutron [None req-c05cb56e-24b0-42e3-b111-9cb18dbe73d6 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 4951132e-7247-4772-8f88-3664c6a7e61e] Successfully updated port: a093d1c9-1f94-4000-8f87-f76e9db98ec8 {{(pid=61957) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 877.970831] env[61957]: DEBUG nova.scheduler.client.report [None req-0d2ed604-6dce-44bf-b238-71a34c40662a tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 878.016113] env[61957]: DEBUG oslo_vmware.api [None req-22255302-8924-4370-98de-304a0c0dd7ba tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': task-1277598, 'name': Destroy_Task} progress is 33%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.116799] env[61957]: DEBUG nova.network.neutron [None req-06bfc4e2-99e2-4510-98f7-f5b198d6231a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 1a0327c2-5671-4970-9db7-c7cc912d8678] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 878.143690] env[61957]: DEBUG oslo_vmware.api [None req-85c79741-463e-4bc0-9ed2-4682b88fcc00 tempest-ServersV294TestFqdnHostnames-1426486100 tempest-ServersV294TestFqdnHostnames-1426486100-project-member] Task: {'id': task-1277597, 'name': PowerOnVM_Task, 'duration_secs': 0.65824} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.143995] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-85c79741-463e-4bc0-9ed2-4682b88fcc00 tempest-ServersV294TestFqdnHostnames-1426486100 tempest-ServersV294TestFqdnHostnames-1426486100-project-member] [instance: 2e4a5344-600f-4b61-826e-c15f96b50af2] Powered on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 878.144220] env[61957]: INFO nova.compute.manager [None req-85c79741-463e-4bc0-9ed2-4682b88fcc00 tempest-ServersV294TestFqdnHostnames-1426486100 tempest-ServersV294TestFqdnHostnames-1426486100-project-member] [instance: 2e4a5344-600f-4b61-826e-c15f96b50af2] Took 7.69 seconds to spawn the instance on the hypervisor. [ 878.144405] env[61957]: DEBUG nova.compute.manager [None req-85c79741-463e-4bc0-9ed2-4682b88fcc00 tempest-ServersV294TestFqdnHostnames-1426486100 tempest-ServersV294TestFqdnHostnames-1426486100-project-member] [instance: 2e4a5344-600f-4b61-826e-c15f96b50af2] Checking state {{(pid=61957) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 878.145212] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d71cb3b1-7330-43fc-ac49-00a4c7610218 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.207407] env[61957]: DEBUG oslo_vmware.api [None req-582b6bd7-3488-48cf-a2f8-f44952b10345 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]528623ed-91c2-0f04-d4be-16747dcb2a70, 'name': SearchDatastore_Task, 'duration_secs': 0.029499} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.207736] env[61957]: DEBUG oslo_concurrency.lockutils [None req-582b6bd7-3488-48cf-a2f8-f44952b10345 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 878.208007] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-582b6bd7-3488-48cf-a2f8-f44952b10345 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore2] e40fa112-4648-428f-a403-b3e3b8319ea8/e40fa112-4648-428f-a403-b3e3b8319ea8.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 878.208356] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4288848a-7c00-4475-9267-f79415380872 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.216616] env[61957]: DEBUG oslo_vmware.api [None req-582b6bd7-3488-48cf-a2f8-f44952b10345 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Waiting for the task: (returnval){ [ 878.216616] env[61957]: value = "task-1277599" [ 878.216616] env[61957]: _type = "Task" [ 878.216616] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.227956] env[61957]: DEBUG oslo_vmware.api [None req-582b6bd7-3488-48cf-a2f8-f44952b10345 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Task: {'id': task-1277599, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.265733] env[61957]: DEBUG nova.network.neutron [None req-06bfc4e2-99e2-4510-98f7-f5b198d6231a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 1a0327c2-5671-4970-9db7-c7cc912d8678] Updating instance_info_cache with network_info: [{"id": "e84fe58e-f595-4b03-9e8a-88238d003168", "address": "fa:16:3e:7e:5a:a2", "network": {"id": "4f0e57be-8906-48b5-865f-59ff12250a8b", "bridge": "br-int", "label": "tempest-ServersTestJSON-1916182687-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "65fd71bcd1bf41238a9cc3a5d6dd4924", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "939c05b6-8f31-4f3a-95ac-6297e0bd243e", "external-id": "nsx-vlan-transportzone-825", "segmentation_id": 825, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape84fe58e-f5", "ovs_interfaceid": "e84fe58e-f595-4b03-9e8a-88238d003168", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 878.360423] env[61957]: DEBUG nova.compute.manager [req-165f61e0-ba07-48fb-8904-1f16d6fde2f1 req-5cbc9a15-be79-403f-b2b9-bb3f92864d30 service nova] [instance: 1a0327c2-5671-4970-9db7-c7cc912d8678] Received event network-vif-plugged-e84fe58e-f595-4b03-9e8a-88238d003168 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 878.360691] env[61957]: DEBUG oslo_concurrency.lockutils [req-165f61e0-ba07-48fb-8904-1f16d6fde2f1 req-5cbc9a15-be79-403f-b2b9-bb3f92864d30 service nova] Acquiring lock "1a0327c2-5671-4970-9db7-c7cc912d8678-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 878.360923] env[61957]: DEBUG oslo_concurrency.lockutils [req-165f61e0-ba07-48fb-8904-1f16d6fde2f1 req-5cbc9a15-be79-403f-b2b9-bb3f92864d30 service nova] Lock "1a0327c2-5671-4970-9db7-c7cc912d8678-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 878.361173] env[61957]: DEBUG oslo_concurrency.lockutils [req-165f61e0-ba07-48fb-8904-1f16d6fde2f1 req-5cbc9a15-be79-403f-b2b9-bb3f92864d30 service nova] Lock "1a0327c2-5671-4970-9db7-c7cc912d8678-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 878.361408] env[61957]: DEBUG nova.compute.manager [req-165f61e0-ba07-48fb-8904-1f16d6fde2f1 req-5cbc9a15-be79-403f-b2b9-bb3f92864d30 service nova] [instance: 1a0327c2-5671-4970-9db7-c7cc912d8678] No waiting events found dispatching network-vif-plugged-e84fe58e-f595-4b03-9e8a-88238d003168 {{(pid=61957) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 878.361717] env[61957]: WARNING nova.compute.manager [req-165f61e0-ba07-48fb-8904-1f16d6fde2f1 req-5cbc9a15-be79-403f-b2b9-bb3f92864d30 service nova] [instance: 1a0327c2-5671-4970-9db7-c7cc912d8678] Received unexpected event network-vif-plugged-e84fe58e-f595-4b03-9e8a-88238d003168 for instance with vm_state building and task_state spawning. [ 878.361897] env[61957]: DEBUG nova.compute.manager [req-165f61e0-ba07-48fb-8904-1f16d6fde2f1 req-5cbc9a15-be79-403f-b2b9-bb3f92864d30 service nova] [instance: 1a0327c2-5671-4970-9db7-c7cc912d8678] Received event network-changed-e84fe58e-f595-4b03-9e8a-88238d003168 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 878.362204] env[61957]: DEBUG nova.compute.manager [req-165f61e0-ba07-48fb-8904-1f16d6fde2f1 req-5cbc9a15-be79-403f-b2b9-bb3f92864d30 service nova] [instance: 1a0327c2-5671-4970-9db7-c7cc912d8678] Refreshing instance network info cache due to event network-changed-e84fe58e-f595-4b03-9e8a-88238d003168. {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 878.362390] env[61957]: DEBUG oslo_concurrency.lockutils [req-165f61e0-ba07-48fb-8904-1f16d6fde2f1 req-5cbc9a15-be79-403f-b2b9-bb3f92864d30 service nova] Acquiring lock "refresh_cache-1a0327c2-5671-4970-9db7-c7cc912d8678" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 878.442855] env[61957]: DEBUG oslo_concurrency.lockutils [None req-c05cb56e-24b0-42e3-b111-9cb18dbe73d6 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Acquiring lock "refresh_cache-4951132e-7247-4772-8f88-3664c6a7e61e" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 878.443225] env[61957]: DEBUG oslo_concurrency.lockutils [None req-c05cb56e-24b0-42e3-b111-9cb18dbe73d6 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Acquired lock "refresh_cache-4951132e-7247-4772-8f88-3664c6a7e61e" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 878.443414] env[61957]: DEBUG nova.network.neutron [None req-c05cb56e-24b0-42e3-b111-9cb18dbe73d6 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 4951132e-7247-4772-8f88-3664c6a7e61e] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 878.475462] env[61957]: DEBUG oslo_concurrency.lockutils [None req-0d2ed604-6dce-44bf-b238-71a34c40662a tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.997s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 878.478011] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d096c609-3f99-4aff-8b69-219b9552f608 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 22.304s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 878.478403] env[61957]: DEBUG nova.objects.instance [None req-d096c609-3f99-4aff-8b69-219b9552f608 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: 37ce46a1-8f9a-4d15-bd81-e40845a0e48a] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61957) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 878.502234] env[61957]: INFO nova.scheduler.client.report [None req-0d2ed604-6dce-44bf-b238-71a34c40662a tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] Deleted allocations for instance 37a9e1c1-9f6c-4047-9ff5-e141d0ca383e [ 878.520309] env[61957]: DEBUG oslo_vmware.api [None req-22255302-8924-4370-98de-304a0c0dd7ba tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': task-1277598, 'name': Destroy_Task, 'duration_secs': 0.899781} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.521517] env[61957]: INFO nova.virt.vmwareapi.vm_util [None req-22255302-8924-4370-98de-304a0c0dd7ba tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: 0adae8e1-8c2f-4110-805b-1f286debc833] Destroyed the VM [ 878.522078] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-22255302-8924-4370-98de-304a0c0dd7ba tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: 0adae8e1-8c2f-4110-805b-1f286debc833] Deleting Snapshot of the VM instance {{(pid=61957) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 878.522789] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-643f7ad6-69d3-482f-b321-6aabde66b872 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.532344] env[61957]: DEBUG oslo_vmware.api [None req-22255302-8924-4370-98de-304a0c0dd7ba tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Waiting for the task: (returnval){ [ 878.532344] env[61957]: value = "task-1277600" [ 878.532344] env[61957]: _type = "Task" [ 878.532344] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.545586] env[61957]: DEBUG oslo_vmware.api [None req-22255302-8924-4370-98de-304a0c0dd7ba tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': task-1277600, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.666038] env[61957]: INFO nova.compute.manager [None req-85c79741-463e-4bc0-9ed2-4682b88fcc00 tempest-ServersV294TestFqdnHostnames-1426486100 tempest-ServersV294TestFqdnHostnames-1426486100-project-member] [instance: 2e4a5344-600f-4b61-826e-c15f96b50af2] Took 35.33 seconds to build instance. [ 878.731138] env[61957]: DEBUG oslo_vmware.api [None req-582b6bd7-3488-48cf-a2f8-f44952b10345 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Task: {'id': task-1277599, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.768946] env[61957]: DEBUG oslo_concurrency.lockutils [None req-06bfc4e2-99e2-4510-98f7-f5b198d6231a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Releasing lock "refresh_cache-1a0327c2-5671-4970-9db7-c7cc912d8678" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 878.769315] env[61957]: DEBUG nova.compute.manager [None req-06bfc4e2-99e2-4510-98f7-f5b198d6231a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 1a0327c2-5671-4970-9db7-c7cc912d8678] Instance network_info: |[{"id": "e84fe58e-f595-4b03-9e8a-88238d003168", "address": "fa:16:3e:7e:5a:a2", "network": {"id": "4f0e57be-8906-48b5-865f-59ff12250a8b", "bridge": "br-int", "label": "tempest-ServersTestJSON-1916182687-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "65fd71bcd1bf41238a9cc3a5d6dd4924", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "939c05b6-8f31-4f3a-95ac-6297e0bd243e", "external-id": "nsx-vlan-transportzone-825", "segmentation_id": 825, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape84fe58e-f5", "ovs_interfaceid": "e84fe58e-f595-4b03-9e8a-88238d003168", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 878.769680] env[61957]: DEBUG oslo_concurrency.lockutils [req-165f61e0-ba07-48fb-8904-1f16d6fde2f1 req-5cbc9a15-be79-403f-b2b9-bb3f92864d30 service nova] Acquired lock "refresh_cache-1a0327c2-5671-4970-9db7-c7cc912d8678" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 878.769892] env[61957]: DEBUG nova.network.neutron [req-165f61e0-ba07-48fb-8904-1f16d6fde2f1 req-5cbc9a15-be79-403f-b2b9-bb3f92864d30 service nova] [instance: 1a0327c2-5671-4970-9db7-c7cc912d8678] Refreshing network info cache for port e84fe58e-f595-4b03-9e8a-88238d003168 {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 878.771513] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-06bfc4e2-99e2-4510-98f7-f5b198d6231a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 1a0327c2-5671-4970-9db7-c7cc912d8678] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:7e:5a:a2', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '939c05b6-8f31-4f3a-95ac-6297e0bd243e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e84fe58e-f595-4b03-9e8a-88238d003168', 'vif_model': 'vmxnet3'}] {{(pid=61957) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 878.779752] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-06bfc4e2-99e2-4510-98f7-f5b198d6231a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Creating folder: Project (65fd71bcd1bf41238a9cc3a5d6dd4924). Parent ref: group-v274445. {{(pid=61957) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 878.780990] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b4e8b9c8-3d4c-413f-871d-7f9461726034 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.795678] env[61957]: INFO nova.virt.vmwareapi.vm_util [None req-06bfc4e2-99e2-4510-98f7-f5b198d6231a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Created folder: Project (65fd71bcd1bf41238a9cc3a5d6dd4924) in parent group-v274445. [ 878.795966] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-06bfc4e2-99e2-4510-98f7-f5b198d6231a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Creating folder: Instances. Parent ref: group-v274532. {{(pid=61957) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 878.796296] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-19741693-6602-4869-818c-571f40547630 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.808921] env[61957]: INFO nova.virt.vmwareapi.vm_util [None req-06bfc4e2-99e2-4510-98f7-f5b198d6231a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Created folder: Instances in parent group-v274532. [ 878.809574] env[61957]: DEBUG oslo.service.loopingcall [None req-06bfc4e2-99e2-4510-98f7-f5b198d6231a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 878.810149] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1a0327c2-5671-4970-9db7-c7cc912d8678] Creating VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 878.810149] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6dc20f4d-402c-41ee-b6de-e285b0aabe1e {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.832527] env[61957]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 878.832527] env[61957]: value = "task-1277603" [ 878.832527] env[61957]: _type = "Task" [ 878.832527] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.841733] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277603, 'name': CreateVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.924711] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ba6997f6-9064-4c0a-859c-de0e4f8ea2ee tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Acquiring lock "0adae8e1-8c2f-4110-805b-1f286debc833" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 878.925074] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ba6997f6-9064-4c0a-859c-de0e4f8ea2ee tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Lock "0adae8e1-8c2f-4110-805b-1f286debc833" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 878.925380] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ba6997f6-9064-4c0a-859c-de0e4f8ea2ee tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Acquiring lock "0adae8e1-8c2f-4110-805b-1f286debc833-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 878.925618] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ba6997f6-9064-4c0a-859c-de0e4f8ea2ee tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Lock "0adae8e1-8c2f-4110-805b-1f286debc833-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 878.925830] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ba6997f6-9064-4c0a-859c-de0e4f8ea2ee tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Lock "0adae8e1-8c2f-4110-805b-1f286debc833-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 878.928011] env[61957]: INFO nova.compute.manager [None req-ba6997f6-9064-4c0a-859c-de0e4f8ea2ee tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: 0adae8e1-8c2f-4110-805b-1f286debc833] Terminating instance [ 878.929961] env[61957]: DEBUG nova.compute.manager [None req-ba6997f6-9064-4c0a-859c-de0e4f8ea2ee tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: 0adae8e1-8c2f-4110-805b-1f286debc833] Start destroying the instance on the hypervisor. {{(pid=61957) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 878.930186] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-ba6997f6-9064-4c0a-859c-de0e4f8ea2ee tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: 0adae8e1-8c2f-4110-805b-1f286debc833] Destroying instance {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 878.931259] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd343c3d-67f1-46e8-8207-bd40bbe6233c {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.941245] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-ba6997f6-9064-4c0a-859c-de0e4f8ea2ee tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: 0adae8e1-8c2f-4110-805b-1f286debc833] Unregistering the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 878.941502] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5c5a221c-37d0-4b7d-84da-b9a53da57a15 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.977906] env[61957]: WARNING nova.network.neutron [None req-c05cb56e-24b0-42e3-b111-9cb18dbe73d6 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 4951132e-7247-4772-8f88-3664c6a7e61e] 364cabcb-046d-4f91-b1ee-5dd3adfae6d7 already exists in list: networks containing: ['364cabcb-046d-4f91-b1ee-5dd3adfae6d7']. ignoring it [ 879.011801] env[61957]: DEBUG oslo_concurrency.lockutils [None req-0d2ed604-6dce-44bf-b238-71a34c40662a tempest-VolumesAdminNegativeTest-1401768390 tempest-VolumesAdminNegativeTest-1401768390-project-member] Lock "37a9e1c1-9f6c-4047-9ff5-e141d0ca383e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 30.018s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 879.045666] env[61957]: DEBUG oslo_vmware.api [None req-22255302-8924-4370-98de-304a0c0dd7ba tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': task-1277600, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.170217] env[61957]: DEBUG oslo_concurrency.lockutils [None req-85c79741-463e-4bc0-9ed2-4682b88fcc00 tempest-ServersV294TestFqdnHostnames-1426486100 tempest-ServersV294TestFqdnHostnames-1426486100-project-member] Lock "2e4a5344-600f-4b61-826e-c15f96b50af2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 64.960s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 879.228378] env[61957]: DEBUG oslo_vmware.api [None req-582b6bd7-3488-48cf-a2f8-f44952b10345 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Task: {'id': task-1277599, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.574927} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.228773] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-582b6bd7-3488-48cf-a2f8-f44952b10345 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore2] e40fa112-4648-428f-a403-b3e3b8319ea8/e40fa112-4648-428f-a403-b3e3b8319ea8.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 879.228915] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-582b6bd7-3488-48cf-a2f8-f44952b10345 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] [instance: e40fa112-4648-428f-a403-b3e3b8319ea8] Extending root virtual disk to 1048576 {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 879.229240] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-65438e9b-cf0b-481e-be8f-41240a7b3b64 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.239869] env[61957]: DEBUG oslo_vmware.api [None req-582b6bd7-3488-48cf-a2f8-f44952b10345 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Waiting for the task: (returnval){ [ 879.239869] env[61957]: value = "task-1277605" [ 879.239869] env[61957]: _type = "Task" [ 879.239869] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.248377] env[61957]: DEBUG oslo_vmware.api [None req-582b6bd7-3488-48cf-a2f8-f44952b10345 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Task: {'id': task-1277605, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.310628] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-ba6997f6-9064-4c0a-859c-de0e4f8ea2ee tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: 0adae8e1-8c2f-4110-805b-1f286debc833] Unregistered the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 879.310866] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-ba6997f6-9064-4c0a-859c-de0e4f8ea2ee tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: 0adae8e1-8c2f-4110-805b-1f286debc833] Deleting contents of the VM from datastore datastore2 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 879.311069] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-ba6997f6-9064-4c0a-859c-de0e4f8ea2ee tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Deleting the datastore file [datastore2] 0adae8e1-8c2f-4110-805b-1f286debc833 {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 879.311383] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c482c4cf-ce4b-4d14-a1d1-c58b27ea184c {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.318925] env[61957]: DEBUG oslo_vmware.api [None req-ba6997f6-9064-4c0a-859c-de0e4f8ea2ee tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Waiting for the task: (returnval){ [ 879.318925] env[61957]: value = "task-1277606" [ 879.318925] env[61957]: _type = "Task" [ 879.318925] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.329407] env[61957]: DEBUG oslo_vmware.api [None req-ba6997f6-9064-4c0a-859c-de0e4f8ea2ee tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': task-1277606, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.332845] env[61957]: DEBUG nova.network.neutron [None req-c05cb56e-24b0-42e3-b111-9cb18dbe73d6 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 4951132e-7247-4772-8f88-3664c6a7e61e] Updating instance_info_cache with network_info: [{"id": "3016c8eb-2ad3-40b1-8d88-eaec447de9ea", "address": "fa:16:3e:59:19:f2", "network": {"id": "364cabcb-046d-4f91-b1ee-5dd3adfae6d7", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1432396967-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.193", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8713aa35bcb24b86ad0b58ca9fc991ba", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b7a73c01-1bb9-4612-a1a7-16d71b732e81", "external-id": "nsx-vlan-transportzone-711", "segmentation_id": 711, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3016c8eb-2a", "ovs_interfaceid": "3016c8eb-2ad3-40b1-8d88-eaec447de9ea", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "a093d1c9-1f94-4000-8f87-f76e9db98ec8", "address": "fa:16:3e:84:fe:9d", "network": {"id": "364cabcb-046d-4f91-b1ee-5dd3adfae6d7", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1432396967-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8713aa35bcb24b86ad0b58ca9fc991ba", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b7a73c01-1bb9-4612-a1a7-16d71b732e81", "external-id": "nsx-vlan-transportzone-711", "segmentation_id": 711, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa093d1c9-1f", "ovs_interfaceid": "a093d1c9-1f94-4000-8f87-f76e9db98ec8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 879.346149] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277603, 'name': CreateVM_Task, 'duration_secs': 0.44145} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.346934] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1a0327c2-5671-4970-9db7-c7cc912d8678] Created VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 879.347956] env[61957]: DEBUG oslo_concurrency.lockutils [None req-06bfc4e2-99e2-4510-98f7-f5b198d6231a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 879.348147] env[61957]: DEBUG oslo_concurrency.lockutils [None req-06bfc4e2-99e2-4510-98f7-f5b198d6231a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 879.348484] env[61957]: DEBUG oslo_concurrency.lockutils [None req-06bfc4e2-99e2-4510-98f7-f5b198d6231a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 879.348989] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-19fba9bf-1f34-4438-b569-f19896dcb90d {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.355258] env[61957]: DEBUG oslo_vmware.api [None req-06bfc4e2-99e2-4510-98f7-f5b198d6231a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Waiting for the task: (returnval){ [ 879.355258] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]5256d54a-c519-a443-b9e8-220f4f04e007" [ 879.355258] env[61957]: _type = "Task" [ 879.355258] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.363760] env[61957]: DEBUG oslo_vmware.api [None req-06bfc4e2-99e2-4510-98f7-f5b198d6231a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]5256d54a-c519-a443-b9e8-220f4f04e007, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.488415] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d096c609-3f99-4aff-8b69-219b9552f608 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.010s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 879.490127] env[61957]: DEBUG oslo_concurrency.lockutils [None req-bafb915b-a249-4473-bf3a-4c41863ba98f tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.190s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 879.492033] env[61957]: INFO nova.compute.claims [None req-bafb915b-a249-4473-bf3a-4c41863ba98f tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] [instance: 28ab7d23-487f-4ae1-8fe4-58db55b59918] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 879.533929] env[61957]: DEBUG nova.network.neutron [req-165f61e0-ba07-48fb-8904-1f16d6fde2f1 req-5cbc9a15-be79-403f-b2b9-bb3f92864d30 service nova] [instance: 1a0327c2-5671-4970-9db7-c7cc912d8678] Updated VIF entry in instance network info cache for port e84fe58e-f595-4b03-9e8a-88238d003168. {{(pid=61957) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 879.534415] env[61957]: DEBUG nova.network.neutron [req-165f61e0-ba07-48fb-8904-1f16d6fde2f1 req-5cbc9a15-be79-403f-b2b9-bb3f92864d30 service nova] [instance: 1a0327c2-5671-4970-9db7-c7cc912d8678] Updating instance_info_cache with network_info: [{"id": "e84fe58e-f595-4b03-9e8a-88238d003168", "address": "fa:16:3e:7e:5a:a2", "network": {"id": "4f0e57be-8906-48b5-865f-59ff12250a8b", "bridge": "br-int", "label": "tempest-ServersTestJSON-1916182687-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "65fd71bcd1bf41238a9cc3a5d6dd4924", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "939c05b6-8f31-4f3a-95ac-6297e0bd243e", "external-id": "nsx-vlan-transportzone-825", "segmentation_id": 825, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape84fe58e-f5", "ovs_interfaceid": "e84fe58e-f595-4b03-9e8a-88238d003168", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 879.547381] env[61957]: DEBUG oslo_vmware.api [None req-22255302-8924-4370-98de-304a0c0dd7ba tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': task-1277600, 'name': RemoveSnapshot_Task, 'duration_secs': 0.695807} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.547665] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-22255302-8924-4370-98de-304a0c0dd7ba tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: 0adae8e1-8c2f-4110-805b-1f286debc833] Deleted Snapshot of the VM instance {{(pid=61957) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 879.547903] env[61957]: INFO nova.compute.manager [None req-22255302-8924-4370-98de-304a0c0dd7ba tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: 0adae8e1-8c2f-4110-805b-1f286debc833] Took 15.53 seconds to snapshot the instance on the hypervisor. [ 879.672953] env[61957]: DEBUG nova.compute.manager [None req-bb69c8c2-41e0-4fa7-aea3-e810870c1532 tempest-ServersNegativeTestMultiTenantJSON-2135399494 tempest-ServersNegativeTestMultiTenantJSON-2135399494-project-member] [instance: 19966b0a-53b7-48c5-849c-a9d00dc024f8] Starting instance... {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 879.748380] env[61957]: DEBUG oslo_vmware.api [None req-582b6bd7-3488-48cf-a2f8-f44952b10345 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Task: {'id': task-1277605, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.084119} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.748693] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-582b6bd7-3488-48cf-a2f8-f44952b10345 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] [instance: e40fa112-4648-428f-a403-b3e3b8319ea8] Extended root virtual disk {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 879.749512] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ad72fb5-aea9-49da-a0b2-e34c5f8ae269 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.774866] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-582b6bd7-3488-48cf-a2f8-f44952b10345 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] [instance: e40fa112-4648-428f-a403-b3e3b8319ea8] Reconfiguring VM instance instance-00000044 to attach disk [datastore2] e40fa112-4648-428f-a403-b3e3b8319ea8/e40fa112-4648-428f-a403-b3e3b8319ea8.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 879.775549] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-98a824cd-28df-444a-a19e-3a597a1aa88a {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.796825] env[61957]: DEBUG oslo_concurrency.lockutils [None req-5a91d738-be3a-4667-949d-376fc04acd42 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Acquiring lock "d4674fbc-47b1-42d1-aaba-e86d46c51e8f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 879.797031] env[61957]: DEBUG oslo_concurrency.lockutils [None req-5a91d738-be3a-4667-949d-376fc04acd42 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Lock "d4674fbc-47b1-42d1-aaba-e86d46c51e8f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 879.798477] env[61957]: DEBUG oslo_vmware.api [None req-582b6bd7-3488-48cf-a2f8-f44952b10345 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Waiting for the task: (returnval){ [ 879.798477] env[61957]: value = "task-1277607" [ 879.798477] env[61957]: _type = "Task" [ 879.798477] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.807417] env[61957]: DEBUG oslo_vmware.api [None req-582b6bd7-3488-48cf-a2f8-f44952b10345 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Task: {'id': task-1277607, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.829038] env[61957]: DEBUG oslo_vmware.api [None req-ba6997f6-9064-4c0a-859c-de0e4f8ea2ee tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': task-1277606, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.838854] env[61957]: DEBUG oslo_concurrency.lockutils [None req-c05cb56e-24b0-42e3-b111-9cb18dbe73d6 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Releasing lock "refresh_cache-4951132e-7247-4772-8f88-3664c6a7e61e" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 879.839505] env[61957]: DEBUG oslo_concurrency.lockutils [None req-c05cb56e-24b0-42e3-b111-9cb18dbe73d6 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Acquiring lock "4951132e-7247-4772-8f88-3664c6a7e61e" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 879.839671] env[61957]: DEBUG oslo_concurrency.lockutils [None req-c05cb56e-24b0-42e3-b111-9cb18dbe73d6 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Acquired lock "4951132e-7247-4772-8f88-3664c6a7e61e" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 879.840781] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ada1e67-7351-4891-9f3e-93866436492b {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.858729] env[61957]: DEBUG nova.virt.hardware [None req-c05cb56e-24b0-42e3-b111-9cb18dbe73d6 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T17:22:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 879.859036] env[61957]: DEBUG nova.virt.hardware [None req-c05cb56e-24b0-42e3-b111-9cb18dbe73d6 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Flavor limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 879.859160] env[61957]: DEBUG nova.virt.hardware [None req-c05cb56e-24b0-42e3-b111-9cb18dbe73d6 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Image limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 879.859348] env[61957]: DEBUG nova.virt.hardware [None req-c05cb56e-24b0-42e3-b111-9cb18dbe73d6 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Flavor pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 879.859500] env[61957]: DEBUG nova.virt.hardware [None req-c05cb56e-24b0-42e3-b111-9cb18dbe73d6 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Image pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 879.859651] env[61957]: DEBUG nova.virt.hardware [None req-c05cb56e-24b0-42e3-b111-9cb18dbe73d6 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 879.859852] env[61957]: DEBUG nova.virt.hardware [None req-c05cb56e-24b0-42e3-b111-9cb18dbe73d6 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 879.860026] env[61957]: DEBUG nova.virt.hardware [None req-c05cb56e-24b0-42e3-b111-9cb18dbe73d6 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 879.860204] env[61957]: DEBUG nova.virt.hardware [None req-c05cb56e-24b0-42e3-b111-9cb18dbe73d6 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Got 1 possible topologies {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 879.860401] env[61957]: DEBUG nova.virt.hardware [None req-c05cb56e-24b0-42e3-b111-9cb18dbe73d6 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 879.860592] env[61957]: DEBUG nova.virt.hardware [None req-c05cb56e-24b0-42e3-b111-9cb18dbe73d6 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 879.866749] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-c05cb56e-24b0-42e3-b111-9cb18dbe73d6 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 4951132e-7247-4772-8f88-3664c6a7e61e] Reconfiguring VM to attach interface {{(pid=61957) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1929}} [ 879.870390] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1d25e341-60aa-44e5-b4ca-ba2d4351a42f {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.888434] env[61957]: DEBUG oslo_vmware.api [None req-06bfc4e2-99e2-4510-98f7-f5b198d6231a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]5256d54a-c519-a443-b9e8-220f4f04e007, 'name': SearchDatastore_Task, 'duration_secs': 0.039683} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.889750] env[61957]: DEBUG oslo_concurrency.lockutils [None req-06bfc4e2-99e2-4510-98f7-f5b198d6231a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 879.889994] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-06bfc4e2-99e2-4510-98f7-f5b198d6231a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 1a0327c2-5671-4970-9db7-c7cc912d8678] Processing image 11c76a2c-f705-470a-ba9d-4657858bab38 {{(pid=61957) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 879.890253] env[61957]: DEBUG oslo_concurrency.lockutils [None req-06bfc4e2-99e2-4510-98f7-f5b198d6231a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 879.890437] env[61957]: DEBUG oslo_concurrency.lockutils [None req-06bfc4e2-99e2-4510-98f7-f5b198d6231a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 879.890639] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-06bfc4e2-99e2-4510-98f7-f5b198d6231a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 879.890971] env[61957]: DEBUG oslo_vmware.api [None req-c05cb56e-24b0-42e3-b111-9cb18dbe73d6 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Waiting for the task: (returnval){ [ 879.890971] env[61957]: value = "task-1277608" [ 879.890971] env[61957]: _type = "Task" [ 879.890971] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.891178] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1af69a2a-9819-4378-83a6-44c8883c3cc1 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.901700] env[61957]: DEBUG oslo_vmware.api [None req-c05cb56e-24b0-42e3-b111-9cb18dbe73d6 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': task-1277608, 'name': ReconfigVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.906030] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-06bfc4e2-99e2-4510-98f7-f5b198d6231a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 879.906242] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-06bfc4e2-99e2-4510-98f7-f5b198d6231a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61957) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 879.907015] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9a5297c7-d5cc-4af8-b429-3ab389b093ec {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.914497] env[61957]: DEBUG oslo_vmware.api [None req-06bfc4e2-99e2-4510-98f7-f5b198d6231a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Waiting for the task: (returnval){ [ 879.914497] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52270d0c-1461-5511-fdc9-c3cb9df4fe1d" [ 879.914497] env[61957]: _type = "Task" [ 879.914497] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.923358] env[61957]: DEBUG oslo_vmware.api [None req-06bfc4e2-99e2-4510-98f7-f5b198d6231a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52270d0c-1461-5511-fdc9-c3cb9df4fe1d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.040940] env[61957]: DEBUG oslo_concurrency.lockutils [req-165f61e0-ba07-48fb-8904-1f16d6fde2f1 req-5cbc9a15-be79-403f-b2b9-bb3f92864d30 service nova] Releasing lock "refresh_cache-1a0327c2-5671-4970-9db7-c7cc912d8678" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 880.041296] env[61957]: DEBUG nova.compute.manager [req-165f61e0-ba07-48fb-8904-1f16d6fde2f1 req-5cbc9a15-be79-403f-b2b9-bb3f92864d30 service nova] [instance: 4951132e-7247-4772-8f88-3664c6a7e61e] Received event network-vif-plugged-a093d1c9-1f94-4000-8f87-f76e9db98ec8 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 880.041600] env[61957]: DEBUG oslo_concurrency.lockutils [req-165f61e0-ba07-48fb-8904-1f16d6fde2f1 req-5cbc9a15-be79-403f-b2b9-bb3f92864d30 service nova] Acquiring lock "4951132e-7247-4772-8f88-3664c6a7e61e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 880.041917] env[61957]: DEBUG oslo_concurrency.lockutils [req-165f61e0-ba07-48fb-8904-1f16d6fde2f1 req-5cbc9a15-be79-403f-b2b9-bb3f92864d30 service nova] Lock "4951132e-7247-4772-8f88-3664c6a7e61e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 880.042199] env[61957]: DEBUG oslo_concurrency.lockutils [req-165f61e0-ba07-48fb-8904-1f16d6fde2f1 req-5cbc9a15-be79-403f-b2b9-bb3f92864d30 service nova] Lock "4951132e-7247-4772-8f88-3664c6a7e61e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 880.042473] env[61957]: DEBUG nova.compute.manager [req-165f61e0-ba07-48fb-8904-1f16d6fde2f1 req-5cbc9a15-be79-403f-b2b9-bb3f92864d30 service nova] [instance: 4951132e-7247-4772-8f88-3664c6a7e61e] No waiting events found dispatching network-vif-plugged-a093d1c9-1f94-4000-8f87-f76e9db98ec8 {{(pid=61957) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 880.042746] env[61957]: WARNING nova.compute.manager [req-165f61e0-ba07-48fb-8904-1f16d6fde2f1 req-5cbc9a15-be79-403f-b2b9-bb3f92864d30 service nova] [instance: 4951132e-7247-4772-8f88-3664c6a7e61e] Received unexpected event network-vif-plugged-a093d1c9-1f94-4000-8f87-f76e9db98ec8 for instance with vm_state active and task_state None. [ 880.043026] env[61957]: DEBUG nova.compute.manager [req-165f61e0-ba07-48fb-8904-1f16d6fde2f1 req-5cbc9a15-be79-403f-b2b9-bb3f92864d30 service nova] [instance: 4951132e-7247-4772-8f88-3664c6a7e61e] Received event network-changed-a093d1c9-1f94-4000-8f87-f76e9db98ec8 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 880.043511] env[61957]: DEBUG nova.compute.manager [req-165f61e0-ba07-48fb-8904-1f16d6fde2f1 req-5cbc9a15-be79-403f-b2b9-bb3f92864d30 service nova] [instance: 4951132e-7247-4772-8f88-3664c6a7e61e] Refreshing instance network info cache due to event network-changed-a093d1c9-1f94-4000-8f87-f76e9db98ec8. {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 880.044136] env[61957]: DEBUG oslo_concurrency.lockutils [req-165f61e0-ba07-48fb-8904-1f16d6fde2f1 req-5cbc9a15-be79-403f-b2b9-bb3f92864d30 service nova] Acquiring lock "refresh_cache-4951132e-7247-4772-8f88-3664c6a7e61e" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 880.044505] env[61957]: DEBUG oslo_concurrency.lockutils [req-165f61e0-ba07-48fb-8904-1f16d6fde2f1 req-5cbc9a15-be79-403f-b2b9-bb3f92864d30 service nova] Acquired lock "refresh_cache-4951132e-7247-4772-8f88-3664c6a7e61e" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 880.044799] env[61957]: DEBUG nova.network.neutron [req-165f61e0-ba07-48fb-8904-1f16d6fde2f1 req-5cbc9a15-be79-403f-b2b9-bb3f92864d30 service nova] [instance: 4951132e-7247-4772-8f88-3664c6a7e61e] Refreshing network info cache for port a093d1c9-1f94-4000-8f87-f76e9db98ec8 {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 880.051684] env[61957]: DEBUG nova.compute.manager [None req-22255302-8924-4370-98de-304a0c0dd7ba tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: 0adae8e1-8c2f-4110-805b-1f286debc833] Instance disappeared during snapshot {{(pid=61957) _snapshot_instance /opt/stack/nova/nova/compute/manager.py:4494}} [ 880.069774] env[61957]: DEBUG nova.compute.manager [None req-22255302-8924-4370-98de-304a0c0dd7ba tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Image not found during clean up 64157260-64c5-4eb6-9410-9be776e1d28d {{(pid=61957) _snapshot_instance /opt/stack/nova/nova/compute/manager.py:4500}} [ 880.202203] env[61957]: DEBUG oslo_concurrency.lockutils [None req-bb69c8c2-41e0-4fa7-aea3-e810870c1532 tempest-ServersNegativeTestMultiTenantJSON-2135399494 tempest-ServersNegativeTestMultiTenantJSON-2135399494-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 880.310858] env[61957]: DEBUG oslo_vmware.api [None req-582b6bd7-3488-48cf-a2f8-f44952b10345 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Task: {'id': task-1277607, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.331678] env[61957]: DEBUG oslo_vmware.api [None req-ba6997f6-9064-4c0a-859c-de0e4f8ea2ee tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': task-1277606, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.856402} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.331678] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-ba6997f6-9064-4c0a-859c-de0e4f8ea2ee tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Deleted the datastore file {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 880.331678] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-ba6997f6-9064-4c0a-859c-de0e4f8ea2ee tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: 0adae8e1-8c2f-4110-805b-1f286debc833] Deleted contents of the VM from datastore datastore2 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 880.331678] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-ba6997f6-9064-4c0a-859c-de0e4f8ea2ee tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: 0adae8e1-8c2f-4110-805b-1f286debc833] Instance destroyed {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 880.332190] env[61957]: INFO nova.compute.manager [None req-ba6997f6-9064-4c0a-859c-de0e4f8ea2ee tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: 0adae8e1-8c2f-4110-805b-1f286debc833] Took 1.40 seconds to destroy the instance on the hypervisor. [ 880.332190] env[61957]: DEBUG oslo.service.loopingcall [None req-ba6997f6-9064-4c0a-859c-de0e4f8ea2ee tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 880.332257] env[61957]: DEBUG nova.compute.manager [-] [instance: 0adae8e1-8c2f-4110-805b-1f286debc833] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 880.332353] env[61957]: DEBUG nova.network.neutron [-] [instance: 0adae8e1-8c2f-4110-805b-1f286debc833] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 880.406455] env[61957]: DEBUG oslo_vmware.api [None req-c05cb56e-24b0-42e3-b111-9cb18dbe73d6 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': task-1277608, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.427026] env[61957]: DEBUG oslo_vmware.api [None req-06bfc4e2-99e2-4510-98f7-f5b198d6231a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52270d0c-1461-5511-fdc9-c3cb9df4fe1d, 'name': SearchDatastore_Task, 'duration_secs': 0.017141} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.427955] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-85ca6487-1943-4885-ac9f-c9c35bb810f0 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.434990] env[61957]: DEBUG oslo_vmware.api [None req-06bfc4e2-99e2-4510-98f7-f5b198d6231a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Waiting for the task: (returnval){ [ 880.434990] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52105cbb-b6c0-2857-8fd3-9e7475be2fd7" [ 880.434990] env[61957]: _type = "Task" [ 880.434990] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.444678] env[61957]: DEBUG oslo_vmware.api [None req-06bfc4e2-99e2-4510-98f7-f5b198d6231a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52105cbb-b6c0-2857-8fd3-9e7475be2fd7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.589135] env[61957]: DEBUG nova.compute.manager [req-20daf719-c4bf-4201-9521-33cd6fc30ff1 req-efe8b81c-69fe-486b-b6bc-a17184b6d258 service nova] [instance: 2e4a5344-600f-4b61-826e-c15f96b50af2] Received event network-changed-a1026b09-e6b6-47f4-b5b2-567a15abfa41 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 880.589547] env[61957]: DEBUG nova.compute.manager [req-20daf719-c4bf-4201-9521-33cd6fc30ff1 req-efe8b81c-69fe-486b-b6bc-a17184b6d258 service nova] [instance: 2e4a5344-600f-4b61-826e-c15f96b50af2] Refreshing instance network info cache due to event network-changed-a1026b09-e6b6-47f4-b5b2-567a15abfa41. {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 880.593020] env[61957]: DEBUG oslo_concurrency.lockutils [req-20daf719-c4bf-4201-9521-33cd6fc30ff1 req-efe8b81c-69fe-486b-b6bc-a17184b6d258 service nova] Acquiring lock "refresh_cache-2e4a5344-600f-4b61-826e-c15f96b50af2" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 880.593020] env[61957]: DEBUG oslo_concurrency.lockutils [req-20daf719-c4bf-4201-9521-33cd6fc30ff1 req-efe8b81c-69fe-486b-b6bc-a17184b6d258 service nova] Acquired lock "refresh_cache-2e4a5344-600f-4b61-826e-c15f96b50af2" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 880.593020] env[61957]: DEBUG nova.network.neutron [req-20daf719-c4bf-4201-9521-33cd6fc30ff1 req-efe8b81c-69fe-486b-b6bc-a17184b6d258 service nova] [instance: 2e4a5344-600f-4b61-826e-c15f96b50af2] Refreshing network info cache for port a1026b09-e6b6-47f4-b5b2-567a15abfa41 {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 880.813776] env[61957]: DEBUG oslo_vmware.api [None req-582b6bd7-3488-48cf-a2f8-f44952b10345 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Task: {'id': task-1277607, 'name': ReconfigVM_Task, 'duration_secs': 0.936394} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.814599] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-582b6bd7-3488-48cf-a2f8-f44952b10345 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] [instance: e40fa112-4648-428f-a403-b3e3b8319ea8] Reconfigured VM instance instance-00000044 to attach disk [datastore2] e40fa112-4648-428f-a403-b3e3b8319ea8/e40fa112-4648-428f-a403-b3e3b8319ea8.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 880.815628] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3c7d6ac8-4ce3-45e4-97ce-d0320d1e0a46 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.830660] env[61957]: DEBUG oslo_vmware.api [None req-582b6bd7-3488-48cf-a2f8-f44952b10345 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Waiting for the task: (returnval){ [ 880.830660] env[61957]: value = "task-1277609" [ 880.830660] env[61957]: _type = "Task" [ 880.830660] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.847175] env[61957]: DEBUG oslo_vmware.api [None req-582b6bd7-3488-48cf-a2f8-f44952b10345 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Task: {'id': task-1277609, 'name': Rename_Task} progress is 5%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.909194] env[61957]: DEBUG oslo_vmware.api [None req-c05cb56e-24b0-42e3-b111-9cb18dbe73d6 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': task-1277608, 'name': ReconfigVM_Task, 'duration_secs': 0.923232} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.915775] env[61957]: DEBUG oslo_concurrency.lockutils [None req-c05cb56e-24b0-42e3-b111-9cb18dbe73d6 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Releasing lock "4951132e-7247-4772-8f88-3664c6a7e61e" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 880.916057] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-c05cb56e-24b0-42e3-b111-9cb18dbe73d6 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 4951132e-7247-4772-8f88-3664c6a7e61e] Reconfigured VM to attach interface {{(pid=61957) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1943}} [ 880.948513] env[61957]: DEBUG oslo_vmware.api [None req-06bfc4e2-99e2-4510-98f7-f5b198d6231a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52105cbb-b6c0-2857-8fd3-9e7475be2fd7, 'name': SearchDatastore_Task, 'duration_secs': 0.014226} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.949154] env[61957]: DEBUG oslo_concurrency.lockutils [None req-06bfc4e2-99e2-4510-98f7-f5b198d6231a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 880.949154] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-06bfc4e2-99e2-4510-98f7-f5b198d6231a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore2] 1a0327c2-5671-4970-9db7-c7cc912d8678/1a0327c2-5671-4970-9db7-c7cc912d8678.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 880.949419] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-45c4c3a8-41d0-45e0-ba9e-e923d8a01b69 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.958122] env[61957]: DEBUG oslo_vmware.api [None req-06bfc4e2-99e2-4510-98f7-f5b198d6231a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Waiting for the task: (returnval){ [ 880.958122] env[61957]: value = "task-1277610" [ 880.958122] env[61957]: _type = "Task" [ 880.958122] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.968398] env[61957]: DEBUG oslo_vmware.api [None req-06bfc4e2-99e2-4510-98f7-f5b198d6231a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': task-1277610, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.970030] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9247b13d-081a-4750-8103-1b02d1ffb311 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.978492] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ec9480d-35a3-46a6-be0f-dfc6d713216c {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.015378] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6d6c745-8b0c-454e-a176-b60f00d6a687 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.025271] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25c30454-8a64-46ff-8d41-e8b6d99b875b {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.040477] env[61957]: DEBUG nova.compute.provider_tree [None req-bafb915b-a249-4473-bf3a-4c41863ba98f tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Updating inventory in ProviderTree for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 881.090929] env[61957]: DEBUG nova.network.neutron [req-165f61e0-ba07-48fb-8904-1f16d6fde2f1 req-5cbc9a15-be79-403f-b2b9-bb3f92864d30 service nova] [instance: 4951132e-7247-4772-8f88-3664c6a7e61e] Updated VIF entry in instance network info cache for port a093d1c9-1f94-4000-8f87-f76e9db98ec8. {{(pid=61957) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 881.091052] env[61957]: DEBUG nova.network.neutron [req-165f61e0-ba07-48fb-8904-1f16d6fde2f1 req-5cbc9a15-be79-403f-b2b9-bb3f92864d30 service nova] [instance: 4951132e-7247-4772-8f88-3664c6a7e61e] Updating instance_info_cache with network_info: [{"id": "3016c8eb-2ad3-40b1-8d88-eaec447de9ea", "address": "fa:16:3e:59:19:f2", "network": {"id": "364cabcb-046d-4f91-b1ee-5dd3adfae6d7", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1432396967-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.193", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8713aa35bcb24b86ad0b58ca9fc991ba", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b7a73c01-1bb9-4612-a1a7-16d71b732e81", "external-id": "nsx-vlan-transportzone-711", "segmentation_id": 711, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3016c8eb-2a", "ovs_interfaceid": "3016c8eb-2ad3-40b1-8d88-eaec447de9ea", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "a093d1c9-1f94-4000-8f87-f76e9db98ec8", "address": "fa:16:3e:84:fe:9d", "network": {"id": "364cabcb-046d-4f91-b1ee-5dd3adfae6d7", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1432396967-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8713aa35bcb24b86ad0b58ca9fc991ba", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b7a73c01-1bb9-4612-a1a7-16d71b732e81", "external-id": "nsx-vlan-transportzone-711", "segmentation_id": 711, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa093d1c9-1f", "ovs_interfaceid": "a093d1c9-1f94-4000-8f87-f76e9db98ec8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 881.341169] env[61957]: DEBUG oslo_vmware.api [None req-582b6bd7-3488-48cf-a2f8-f44952b10345 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Task: {'id': task-1277609, 'name': Rename_Task, 'duration_secs': 0.269487} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.342321] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-582b6bd7-3488-48cf-a2f8-f44952b10345 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] [instance: e40fa112-4648-428f-a403-b3e3b8319ea8] Powering on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 881.342321] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9d1f94ca-c58d-4271-92f4-8367d08b0d57 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.349492] env[61957]: DEBUG oslo_vmware.api [None req-582b6bd7-3488-48cf-a2f8-f44952b10345 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Waiting for the task: (returnval){ [ 881.349492] env[61957]: value = "task-1277611" [ 881.349492] env[61957]: _type = "Task" [ 881.349492] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.354723] env[61957]: DEBUG nova.compute.manager [req-ea817a6d-82b4-414b-9ff8-64689f642162 req-fc2ad563-f7d4-45cf-bc2b-14cb62de94be service nova] [instance: 0adae8e1-8c2f-4110-805b-1f286debc833] Received event network-vif-deleted-06b59d3b-7dab-4254-9ccd-cbfcce2fce87 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 881.354928] env[61957]: INFO nova.compute.manager [req-ea817a6d-82b4-414b-9ff8-64689f642162 req-fc2ad563-f7d4-45cf-bc2b-14cb62de94be service nova] [instance: 0adae8e1-8c2f-4110-805b-1f286debc833] Neutron deleted interface 06b59d3b-7dab-4254-9ccd-cbfcce2fce87; detaching it from the instance and deleting it from the info cache [ 881.355120] env[61957]: DEBUG nova.network.neutron [req-ea817a6d-82b4-414b-9ff8-64689f642162 req-fc2ad563-f7d4-45cf-bc2b-14cb62de94be service nova] [instance: 0adae8e1-8c2f-4110-805b-1f286debc833] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 881.363237] env[61957]: DEBUG oslo_vmware.api [None req-582b6bd7-3488-48cf-a2f8-f44952b10345 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Task: {'id': task-1277611, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.419815] env[61957]: DEBUG nova.network.neutron [-] [instance: 0adae8e1-8c2f-4110-805b-1f286debc833] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 881.422200] env[61957]: DEBUG oslo_concurrency.lockutils [None req-c05cb56e-24b0-42e3-b111-9cb18dbe73d6 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Lock "interface-4951132e-7247-4772-8f88-3664c6a7e61e-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 6.733s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 881.470800] env[61957]: DEBUG oslo_vmware.api [None req-06bfc4e2-99e2-4510-98f7-f5b198d6231a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': task-1277610, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.570669] env[61957]: ERROR nova.scheduler.client.report [None req-bafb915b-a249-4473-bf3a-4c41863ba98f tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] [req-8881c99f-c159-4184-9b90-ad45b20cf22a] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 0ceb6c9e-61c6-496d-8579-9d32627e96da. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-8881c99f-c159-4184-9b90-ad45b20cf22a"}]} [ 881.590619] env[61957]: DEBUG nova.scheduler.client.report [None req-bafb915b-a249-4473-bf3a-4c41863ba98f tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Refreshing inventories for resource provider 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 881.593195] env[61957]: DEBUG oslo_concurrency.lockutils [req-165f61e0-ba07-48fb-8904-1f16d6fde2f1 req-5cbc9a15-be79-403f-b2b9-bb3f92864d30 service nova] Releasing lock "refresh_cache-4951132e-7247-4772-8f88-3664c6a7e61e" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 881.617477] env[61957]: DEBUG nova.scheduler.client.report [None req-bafb915b-a249-4473-bf3a-4c41863ba98f tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Updating ProviderTree inventory for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 881.618399] env[61957]: DEBUG nova.compute.provider_tree [None req-bafb915b-a249-4473-bf3a-4c41863ba98f tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Updating inventory in ProviderTree for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 881.633096] env[61957]: DEBUG nova.scheduler.client.report [None req-bafb915b-a249-4473-bf3a-4c41863ba98f tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Refreshing aggregate associations for resource provider 0ceb6c9e-61c6-496d-8579-9d32627e96da, aggregates: None {{(pid=61957) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 881.657476] env[61957]: DEBUG nova.scheduler.client.report [None req-bafb915b-a249-4473-bf3a-4c41863ba98f tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Refreshing trait associations for resource provider 0ceb6c9e-61c6-496d-8579-9d32627e96da, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO,HW_ARCH_X86_64,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=61957) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 881.662558] env[61957]: DEBUG nova.network.neutron [req-20daf719-c4bf-4201-9521-33cd6fc30ff1 req-efe8b81c-69fe-486b-b6bc-a17184b6d258 service nova] [instance: 2e4a5344-600f-4b61-826e-c15f96b50af2] Updated VIF entry in instance network info cache for port a1026b09-e6b6-47f4-b5b2-567a15abfa41. {{(pid=61957) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 881.663322] env[61957]: DEBUG nova.network.neutron [req-20daf719-c4bf-4201-9521-33cd6fc30ff1 req-efe8b81c-69fe-486b-b6bc-a17184b6d258 service nova] [instance: 2e4a5344-600f-4b61-826e-c15f96b50af2] Updating instance_info_cache with network_info: [{"id": "a1026b09-e6b6-47f4-b5b2-567a15abfa41", "address": "fa:16:3e:38:85:22", "network": {"id": "7ca8d6d5-0a42-4dd6-b7bd-bba4b8618fca", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-636668201-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.161", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ecf709e5c49146238729af50f362b1ea", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e2e8b74b-aa27-4f31-9414-7bcf531e8642", "external-id": "nsx-vlan-transportzone-544", "segmentation_id": 544, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa1026b09-e6", "ovs_interfaceid": "a1026b09-e6b6-47f4-b5b2-567a15abfa41", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 881.861886] env[61957]: DEBUG oslo_vmware.api [None req-582b6bd7-3488-48cf-a2f8-f44952b10345 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Task: {'id': task-1277611, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.864996] env[61957]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-314f7138-64db-4fb0-99ce-45a2f760ff09 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.875857] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38d7ec5a-62ea-4cff-adde-134a4945f089 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.916591] env[61957]: DEBUG nova.compute.manager [req-ea817a6d-82b4-414b-9ff8-64689f642162 req-fc2ad563-f7d4-45cf-bc2b-14cb62de94be service nova] [instance: 0adae8e1-8c2f-4110-805b-1f286debc833] Detach interface failed, port_id=06b59d3b-7dab-4254-9ccd-cbfcce2fce87, reason: Instance 0adae8e1-8c2f-4110-805b-1f286debc833 could not be found. {{(pid=61957) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 881.924291] env[61957]: INFO nova.compute.manager [-] [instance: 0adae8e1-8c2f-4110-805b-1f286debc833] Took 1.59 seconds to deallocate network for instance. [ 881.973632] env[61957]: DEBUG oslo_vmware.api [None req-06bfc4e2-99e2-4510-98f7-f5b198d6231a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': task-1277610, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.802123} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.973632] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-06bfc4e2-99e2-4510-98f7-f5b198d6231a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore2] 1a0327c2-5671-4970-9db7-c7cc912d8678/1a0327c2-5671-4970-9db7-c7cc912d8678.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 881.973632] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-06bfc4e2-99e2-4510-98f7-f5b198d6231a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 1a0327c2-5671-4970-9db7-c7cc912d8678] Extending root virtual disk to 1048576 {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 881.974258] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2df45ad8-b3e4-4722-b7ba-6b7a3c33cb81 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.981768] env[61957]: DEBUG oslo_vmware.api [None req-06bfc4e2-99e2-4510-98f7-f5b198d6231a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Waiting for the task: (returnval){ [ 881.981768] env[61957]: value = "task-1277612" [ 881.981768] env[61957]: _type = "Task" [ 881.981768] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.991299] env[61957]: DEBUG oslo_vmware.api [None req-06bfc4e2-99e2-4510-98f7-f5b198d6231a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': task-1277612, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.064694] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d0476d9-2c6d-4aaf-8f86-3d865ef96df9 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.074306] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d66fc7c5-55e2-4eec-9bfd-444016c5224a {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.107604] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae660980-924d-4ede-bafa-a51ea042cae4 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.117125] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db592678-8a33-4b36-a890-f82572065640 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.131960] env[61957]: DEBUG nova.compute.provider_tree [None req-bafb915b-a249-4473-bf3a-4c41863ba98f tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 882.168654] env[61957]: DEBUG oslo_concurrency.lockutils [req-20daf719-c4bf-4201-9521-33cd6fc30ff1 req-efe8b81c-69fe-486b-b6bc-a17184b6d258 service nova] Releasing lock "refresh_cache-2e4a5344-600f-4b61-826e-c15f96b50af2" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 882.361122] env[61957]: DEBUG oslo_vmware.api [None req-582b6bd7-3488-48cf-a2f8-f44952b10345 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Task: {'id': task-1277611, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.437536] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ba6997f6-9064-4c0a-859c-de0e4f8ea2ee tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 882.450423] env[61957]: DEBUG oslo_concurrency.lockutils [None req-9513c200-94de-43f4-aaa5-418ecbddf25f tempest-ServerTagsTestJSON-969152625 tempest-ServerTagsTestJSON-969152625-project-member] Acquiring lock "8e04a098-25de-4d57-9e3c-ea44d234d57e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 882.450423] env[61957]: DEBUG oslo_concurrency.lockutils [None req-9513c200-94de-43f4-aaa5-418ecbddf25f tempest-ServerTagsTestJSON-969152625 tempest-ServerTagsTestJSON-969152625-project-member] Lock "8e04a098-25de-4d57-9e3c-ea44d234d57e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 882.451238] env[61957]: DEBUG oslo_concurrency.lockutils [None req-9513c200-94de-43f4-aaa5-418ecbddf25f tempest-ServerTagsTestJSON-969152625 tempest-ServerTagsTestJSON-969152625-project-member] Acquiring lock "8e04a098-25de-4d57-9e3c-ea44d234d57e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 882.451238] env[61957]: DEBUG oslo_concurrency.lockutils [None req-9513c200-94de-43f4-aaa5-418ecbddf25f tempest-ServerTagsTestJSON-969152625 tempest-ServerTagsTestJSON-969152625-project-member] Lock "8e04a098-25de-4d57-9e3c-ea44d234d57e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 882.451238] env[61957]: DEBUG oslo_concurrency.lockutils [None req-9513c200-94de-43f4-aaa5-418ecbddf25f tempest-ServerTagsTestJSON-969152625 tempest-ServerTagsTestJSON-969152625-project-member] Lock "8e04a098-25de-4d57-9e3c-ea44d234d57e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 882.454103] env[61957]: INFO nova.compute.manager [None req-9513c200-94de-43f4-aaa5-418ecbddf25f tempest-ServerTagsTestJSON-969152625 tempest-ServerTagsTestJSON-969152625-project-member] [instance: 8e04a098-25de-4d57-9e3c-ea44d234d57e] Terminating instance [ 882.456669] env[61957]: DEBUG nova.compute.manager [None req-9513c200-94de-43f4-aaa5-418ecbddf25f tempest-ServerTagsTestJSON-969152625 tempest-ServerTagsTestJSON-969152625-project-member] [instance: 8e04a098-25de-4d57-9e3c-ea44d234d57e] Start destroying the instance on the hypervisor. {{(pid=61957) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 882.457059] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-9513c200-94de-43f4-aaa5-418ecbddf25f tempest-ServerTagsTestJSON-969152625 tempest-ServerTagsTestJSON-969152625-project-member] [instance: 8e04a098-25de-4d57-9e3c-ea44d234d57e] Destroying instance {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 882.457961] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb3a1b12-8d16-4ccb-a9a4-f1e721a62e0a {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.467045] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-9513c200-94de-43f4-aaa5-418ecbddf25f tempest-ServerTagsTestJSON-969152625 tempest-ServerTagsTestJSON-969152625-project-member] [instance: 8e04a098-25de-4d57-9e3c-ea44d234d57e] Powering off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 882.468022] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-dea5457d-6f42-41be-a4c5-e5f7280ff1c7 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.476861] env[61957]: DEBUG oslo_vmware.api [None req-9513c200-94de-43f4-aaa5-418ecbddf25f tempest-ServerTagsTestJSON-969152625 tempest-ServerTagsTestJSON-969152625-project-member] Waiting for the task: (returnval){ [ 882.476861] env[61957]: value = "task-1277613" [ 882.476861] env[61957]: _type = "Task" [ 882.476861] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.488572] env[61957]: DEBUG oslo_vmware.api [None req-9513c200-94de-43f4-aaa5-418ecbddf25f tempest-ServerTagsTestJSON-969152625 tempest-ServerTagsTestJSON-969152625-project-member] Task: {'id': task-1277613, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.497132] env[61957]: DEBUG oslo_vmware.api [None req-06bfc4e2-99e2-4510-98f7-f5b198d6231a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': task-1277612, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.24381} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.497693] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-06bfc4e2-99e2-4510-98f7-f5b198d6231a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 1a0327c2-5671-4970-9db7-c7cc912d8678] Extended root virtual disk {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 882.499087] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc63b478-21ea-4d16-a22e-5fbb4e637b95 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.527361] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-06bfc4e2-99e2-4510-98f7-f5b198d6231a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 1a0327c2-5671-4970-9db7-c7cc912d8678] Reconfiguring VM instance instance-00000045 to attach disk [datastore2] 1a0327c2-5671-4970-9db7-c7cc912d8678/1a0327c2-5671-4970-9db7-c7cc912d8678.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 882.527877] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0724d4ac-3f82-43cf-b4f1-376ed46474b3 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.554807] env[61957]: DEBUG oslo_vmware.api [None req-06bfc4e2-99e2-4510-98f7-f5b198d6231a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Waiting for the task: (returnval){ [ 882.554807] env[61957]: value = "task-1277614" [ 882.554807] env[61957]: _type = "Task" [ 882.554807] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.569208] env[61957]: DEBUG oslo_vmware.api [None req-06bfc4e2-99e2-4510-98f7-f5b198d6231a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': task-1277614, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.635490] env[61957]: DEBUG nova.scheduler.client.report [None req-bafb915b-a249-4473-bf3a-4c41863ba98f tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 882.643214] env[61957]: DEBUG oslo_concurrency.lockutils [None req-a3e467c9-b0de-4881-ac92-2329baed7e90 tempest-ServersTestJSON-1093633677 tempest-ServersTestJSON-1093633677-project-member] Acquiring lock "ead6aae1-36b5-4f57-9129-3bb02cf103ce" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 882.643456] env[61957]: DEBUG oslo_concurrency.lockutils [None req-a3e467c9-b0de-4881-ac92-2329baed7e90 tempest-ServersTestJSON-1093633677 tempest-ServersTestJSON-1093633677-project-member] Lock "ead6aae1-36b5-4f57-9129-3bb02cf103ce" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 882.644093] env[61957]: DEBUG oslo_concurrency.lockutils [None req-a3e467c9-b0de-4881-ac92-2329baed7e90 tempest-ServersTestJSON-1093633677 tempest-ServersTestJSON-1093633677-project-member] Acquiring lock "ead6aae1-36b5-4f57-9129-3bb02cf103ce-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 882.644093] env[61957]: DEBUG oslo_concurrency.lockutils [None req-a3e467c9-b0de-4881-ac92-2329baed7e90 tempest-ServersTestJSON-1093633677 tempest-ServersTestJSON-1093633677-project-member] Lock "ead6aae1-36b5-4f57-9129-3bb02cf103ce-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 882.644093] env[61957]: DEBUG oslo_concurrency.lockutils [None req-a3e467c9-b0de-4881-ac92-2329baed7e90 tempest-ServersTestJSON-1093633677 tempest-ServersTestJSON-1093633677-project-member] Lock "ead6aae1-36b5-4f57-9129-3bb02cf103ce-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 882.646324] env[61957]: INFO nova.compute.manager [None req-a3e467c9-b0de-4881-ac92-2329baed7e90 tempest-ServersTestJSON-1093633677 tempest-ServersTestJSON-1093633677-project-member] [instance: ead6aae1-36b5-4f57-9129-3bb02cf103ce] Terminating instance [ 882.648320] env[61957]: DEBUG nova.compute.manager [None req-a3e467c9-b0de-4881-ac92-2329baed7e90 tempest-ServersTestJSON-1093633677 tempest-ServersTestJSON-1093633677-project-member] [instance: ead6aae1-36b5-4f57-9129-3bb02cf103ce] Start destroying the instance on the hypervisor. {{(pid=61957) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 882.648518] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-a3e467c9-b0de-4881-ac92-2329baed7e90 tempest-ServersTestJSON-1093633677 tempest-ServersTestJSON-1093633677-project-member] [instance: ead6aae1-36b5-4f57-9129-3bb02cf103ce] Destroying instance {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 882.649374] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-578f69b1-83ab-4a24-921e-e2f7cfdd98dc {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.659053] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-a3e467c9-b0de-4881-ac92-2329baed7e90 tempest-ServersTestJSON-1093633677 tempest-ServersTestJSON-1093633677-project-member] [instance: ead6aae1-36b5-4f57-9129-3bb02cf103ce] Powering off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 882.659318] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f126af73-7c44-484f-8649-6f441dd8fce7 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.668349] env[61957]: DEBUG oslo_vmware.api [None req-a3e467c9-b0de-4881-ac92-2329baed7e90 tempest-ServersTestJSON-1093633677 tempest-ServersTestJSON-1093633677-project-member] Waiting for the task: (returnval){ [ 882.668349] env[61957]: value = "task-1277615" [ 882.668349] env[61957]: _type = "Task" [ 882.668349] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.683918] env[61957]: DEBUG oslo_vmware.api [None req-a3e467c9-b0de-4881-ac92-2329baed7e90 tempest-ServersTestJSON-1093633677 tempest-ServersTestJSON-1093633677-project-member] Task: {'id': task-1277615, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.861934] env[61957]: DEBUG oslo_vmware.api [None req-582b6bd7-3488-48cf-a2f8-f44952b10345 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Task: {'id': task-1277611, 'name': PowerOnVM_Task, 'duration_secs': 1.257898} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.862296] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-582b6bd7-3488-48cf-a2f8-f44952b10345 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] [instance: e40fa112-4648-428f-a403-b3e3b8319ea8] Powered on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 882.862622] env[61957]: INFO nova.compute.manager [None req-582b6bd7-3488-48cf-a2f8-f44952b10345 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] [instance: e40fa112-4648-428f-a403-b3e3b8319ea8] Took 8.90 seconds to spawn the instance on the hypervisor. [ 882.862849] env[61957]: DEBUG nova.compute.manager [None req-582b6bd7-3488-48cf-a2f8-f44952b10345 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] [instance: e40fa112-4648-428f-a403-b3e3b8319ea8] Checking state {{(pid=61957) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 882.863704] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c67d2c2-fb11-4192-a6ac-13a918482a06 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.987122] env[61957]: DEBUG oslo_vmware.api [None req-9513c200-94de-43f4-aaa5-418ecbddf25f tempest-ServerTagsTestJSON-969152625 tempest-ServerTagsTestJSON-969152625-project-member] Task: {'id': task-1277613, 'name': PowerOffVM_Task, 'duration_secs': 0.30284} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.987420] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-9513c200-94de-43f4-aaa5-418ecbddf25f tempest-ServerTagsTestJSON-969152625 tempest-ServerTagsTestJSON-969152625-project-member] [instance: 8e04a098-25de-4d57-9e3c-ea44d234d57e] Powered off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 882.987593] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-9513c200-94de-43f4-aaa5-418ecbddf25f tempest-ServerTagsTestJSON-969152625 tempest-ServerTagsTestJSON-969152625-project-member] [instance: 8e04a098-25de-4d57-9e3c-ea44d234d57e] Unregistering the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 882.987858] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-27464b6f-9f75-432c-bf8c-eb199432ea25 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.060686] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-9513c200-94de-43f4-aaa5-418ecbddf25f tempest-ServerTagsTestJSON-969152625 tempest-ServerTagsTestJSON-969152625-project-member] [instance: 8e04a098-25de-4d57-9e3c-ea44d234d57e] Unregistered the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 883.060805] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-9513c200-94de-43f4-aaa5-418ecbddf25f tempest-ServerTagsTestJSON-969152625 tempest-ServerTagsTestJSON-969152625-project-member] [instance: 8e04a098-25de-4d57-9e3c-ea44d234d57e] Deleting contents of the VM from datastore datastore2 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 883.061029] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-9513c200-94de-43f4-aaa5-418ecbddf25f tempest-ServerTagsTestJSON-969152625 tempest-ServerTagsTestJSON-969152625-project-member] Deleting the datastore file [datastore2] 8e04a098-25de-4d57-9e3c-ea44d234d57e {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 883.061254] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-62dc1e27-a2bd-4571-9846-94f8aed86b30 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.069468] env[61957]: DEBUG oslo_vmware.api [None req-06bfc4e2-99e2-4510-98f7-f5b198d6231a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': task-1277614, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.071378] env[61957]: DEBUG oslo_vmware.api [None req-9513c200-94de-43f4-aaa5-418ecbddf25f tempest-ServerTagsTestJSON-969152625 tempest-ServerTagsTestJSON-969152625-project-member] Waiting for the task: (returnval){ [ 883.071378] env[61957]: value = "task-1277617" [ 883.071378] env[61957]: _type = "Task" [ 883.071378] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.081359] env[61957]: DEBUG oslo_vmware.api [None req-9513c200-94de-43f4-aaa5-418ecbddf25f tempest-ServerTagsTestJSON-969152625 tempest-ServerTagsTestJSON-969152625-project-member] Task: {'id': task-1277617, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.147465] env[61957]: DEBUG oslo_concurrency.lockutils [None req-bafb915b-a249-4473-bf3a-4c41863ba98f tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.657s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 883.148485] env[61957]: DEBUG nova.compute.manager [None req-bafb915b-a249-4473-bf3a-4c41863ba98f tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] [instance: 28ab7d23-487f-4ae1-8fe4-58db55b59918] Start building networks asynchronously for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 883.152525] env[61957]: DEBUG oslo_concurrency.lockutils [None req-a0f6bb40-5b1e-4b57-987f-3c85bacc19c6 tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 22.020s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 883.152797] env[61957]: DEBUG nova.objects.instance [None req-a0f6bb40-5b1e-4b57-987f-3c85bacc19c6 tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Lazy-loading 'resources' on Instance uuid e0249936-d616-4ffb-8f77-d8107633c42a {{(pid=61957) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 883.182296] env[61957]: DEBUG oslo_vmware.api [None req-a3e467c9-b0de-4881-ac92-2329baed7e90 tempest-ServersTestJSON-1093633677 tempest-ServersTestJSON-1093633677-project-member] Task: {'id': task-1277615, 'name': PowerOffVM_Task} progress is 100%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.318609] env[61957]: DEBUG oslo_concurrency.lockutils [None req-4f69433c-53e0-4b6d-8d33-8d933f713309 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Acquiring lock "interface-4951132e-7247-4772-8f88-3664c6a7e61e-a093d1c9-1f94-4000-8f87-f76e9db98ec8" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 883.318609] env[61957]: DEBUG oslo_concurrency.lockutils [None req-4f69433c-53e0-4b6d-8d33-8d933f713309 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Lock "interface-4951132e-7247-4772-8f88-3664c6a7e61e-a093d1c9-1f94-4000-8f87-f76e9db98ec8" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.001s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 883.382756] env[61957]: INFO nova.compute.manager [None req-582b6bd7-3488-48cf-a2f8-f44952b10345 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] [instance: e40fa112-4648-428f-a403-b3e3b8319ea8] Took 39.03 seconds to build instance. [ 883.567035] env[61957]: DEBUG oslo_vmware.api [None req-06bfc4e2-99e2-4510-98f7-f5b198d6231a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': task-1277614, 'name': ReconfigVM_Task, 'duration_secs': 0.926111} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.567343] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-06bfc4e2-99e2-4510-98f7-f5b198d6231a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 1a0327c2-5671-4970-9db7-c7cc912d8678] Reconfigured VM instance instance-00000045 to attach disk [datastore2] 1a0327c2-5671-4970-9db7-c7cc912d8678/1a0327c2-5671-4970-9db7-c7cc912d8678.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 883.568038] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f1e8039a-60bf-47dc-adb4-d1793164ec17 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.577688] env[61957]: DEBUG oslo_vmware.api [None req-06bfc4e2-99e2-4510-98f7-f5b198d6231a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Waiting for the task: (returnval){ [ 883.577688] env[61957]: value = "task-1277618" [ 883.577688] env[61957]: _type = "Task" [ 883.577688] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.583627] env[61957]: DEBUG oslo_vmware.api [None req-9513c200-94de-43f4-aaa5-418ecbddf25f tempest-ServerTagsTestJSON-969152625 tempest-ServerTagsTestJSON-969152625-project-member] Task: {'id': task-1277617, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.281281} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.584322] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-9513c200-94de-43f4-aaa5-418ecbddf25f tempest-ServerTagsTestJSON-969152625 tempest-ServerTagsTestJSON-969152625-project-member] Deleted the datastore file {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 883.584630] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-9513c200-94de-43f4-aaa5-418ecbddf25f tempest-ServerTagsTestJSON-969152625 tempest-ServerTagsTestJSON-969152625-project-member] [instance: 8e04a098-25de-4d57-9e3c-ea44d234d57e] Deleted contents of the VM from datastore datastore2 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 883.584923] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-9513c200-94de-43f4-aaa5-418ecbddf25f tempest-ServerTagsTestJSON-969152625 tempest-ServerTagsTestJSON-969152625-project-member] [instance: 8e04a098-25de-4d57-9e3c-ea44d234d57e] Instance destroyed {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 883.585164] env[61957]: INFO nova.compute.manager [None req-9513c200-94de-43f4-aaa5-418ecbddf25f tempest-ServerTagsTestJSON-969152625 tempest-ServerTagsTestJSON-969152625-project-member] [instance: 8e04a098-25de-4d57-9e3c-ea44d234d57e] Took 1.13 seconds to destroy the instance on the hypervisor. [ 883.585446] env[61957]: DEBUG oslo.service.loopingcall [None req-9513c200-94de-43f4-aaa5-418ecbddf25f tempest-ServerTagsTestJSON-969152625 tempest-ServerTagsTestJSON-969152625-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 883.585698] env[61957]: DEBUG nova.compute.manager [-] [instance: 8e04a098-25de-4d57-9e3c-ea44d234d57e] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 883.585850] env[61957]: DEBUG nova.network.neutron [-] [instance: 8e04a098-25de-4d57-9e3c-ea44d234d57e] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 883.590811] env[61957]: DEBUG oslo_vmware.api [None req-06bfc4e2-99e2-4510-98f7-f5b198d6231a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': task-1277618, 'name': Rename_Task} progress is 5%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.658386] env[61957]: DEBUG nova.compute.utils [None req-bafb915b-a249-4473-bf3a-4c41863ba98f tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Using /dev/sd instead of None {{(pid=61957) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 883.660684] env[61957]: DEBUG nova.compute.manager [None req-bafb915b-a249-4473-bf3a-4c41863ba98f tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] [instance: 28ab7d23-487f-4ae1-8fe4-58db55b59918] Allocating IP information in the background. {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 883.660684] env[61957]: DEBUG nova.network.neutron [None req-bafb915b-a249-4473-bf3a-4c41863ba98f tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] [instance: 28ab7d23-487f-4ae1-8fe4-58db55b59918] allocate_for_instance() {{(pid=61957) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 883.685021] env[61957]: DEBUG oslo_vmware.api [None req-a3e467c9-b0de-4881-ac92-2329baed7e90 tempest-ServersTestJSON-1093633677 tempest-ServersTestJSON-1093633677-project-member] Task: {'id': task-1277615, 'name': PowerOffVM_Task, 'duration_secs': 0.612928} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.688141] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-a3e467c9-b0de-4881-ac92-2329baed7e90 tempest-ServersTestJSON-1093633677 tempest-ServersTestJSON-1093633677-project-member] [instance: ead6aae1-36b5-4f57-9129-3bb02cf103ce] Powered off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 883.688376] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-a3e467c9-b0de-4881-ac92-2329baed7e90 tempest-ServersTestJSON-1093633677 tempest-ServersTestJSON-1093633677-project-member] [instance: ead6aae1-36b5-4f57-9129-3bb02cf103ce] Unregistering the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 883.689611] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-263ab795-bf0c-4bdd-bec5-a84cd1922b73 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.726089] env[61957]: DEBUG nova.policy [None req-bafb915b-a249-4473-bf3a-4c41863ba98f tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9e3ccf0e318349b38c7eff2c89440e54', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd849d922cfad41848a33e8d3bc83ea0f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61957) authorize /opt/stack/nova/nova/policy.py:203}} [ 883.821806] env[61957]: DEBUG oslo_concurrency.lockutils [None req-4f69433c-53e0-4b6d-8d33-8d933f713309 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Acquiring lock "4951132e-7247-4772-8f88-3664c6a7e61e" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 883.822057] env[61957]: DEBUG oslo_concurrency.lockutils [None req-4f69433c-53e0-4b6d-8d33-8d933f713309 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Acquired lock "4951132e-7247-4772-8f88-3664c6a7e61e" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 883.823507] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc60f556-8ece-423a-84fb-c7a976fd1e2e {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.847805] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1010ccb0-f4cc-4e1f-b0f7-a11489161f54 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.876438] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-4f69433c-53e0-4b6d-8d33-8d933f713309 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 4951132e-7247-4772-8f88-3664c6a7e61e] Reconfiguring VM to detach interface {{(pid=61957) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1974}} [ 883.879915] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3b11efed-e0b2-4785-8d1d-27bd53b97feb {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.893765] env[61957]: DEBUG oslo_concurrency.lockutils [None req-582b6bd7-3488-48cf-a2f8-f44952b10345 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Lock "e40fa112-4648-428f-a403-b3e3b8319ea8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 65.803s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 883.901541] env[61957]: DEBUG oslo_vmware.api [None req-4f69433c-53e0-4b6d-8d33-8d933f713309 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Waiting for the task: (returnval){ [ 883.901541] env[61957]: value = "task-1277620" [ 883.901541] env[61957]: _type = "Task" [ 883.901541] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.920288] env[61957]: DEBUG oslo_vmware.api [None req-4f69433c-53e0-4b6d-8d33-8d933f713309 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': task-1277620, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.003958] env[61957]: DEBUG nova.compute.manager [req-cbf3b15f-372b-4ba8-9cd4-be0b9c41ee6e req-ce5bf8c0-347c-444e-90b3-912bb47a242a service nova] [instance: 8e04a098-25de-4d57-9e3c-ea44d234d57e] Received event network-vif-deleted-55344e12-1398-4752-a0f4-fdea985786bc {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 884.004222] env[61957]: INFO nova.compute.manager [req-cbf3b15f-372b-4ba8-9cd4-be0b9c41ee6e req-ce5bf8c0-347c-444e-90b3-912bb47a242a service nova] [instance: 8e04a098-25de-4d57-9e3c-ea44d234d57e] Neutron deleted interface 55344e12-1398-4752-a0f4-fdea985786bc; detaching it from the instance and deleting it from the info cache [ 884.004376] env[61957]: DEBUG nova.network.neutron [req-cbf3b15f-372b-4ba8-9cd4-be0b9c41ee6e req-ce5bf8c0-347c-444e-90b3-912bb47a242a service nova] [instance: 8e04a098-25de-4d57-9e3c-ea44d234d57e] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 884.087823] env[61957]: DEBUG oslo_vmware.api [None req-06bfc4e2-99e2-4510-98f7-f5b198d6231a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': task-1277618, 'name': Rename_Task} progress is 14%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.099087] env[61957]: DEBUG nova.network.neutron [None req-bafb915b-a249-4473-bf3a-4c41863ba98f tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] [instance: 28ab7d23-487f-4ae1-8fe4-58db55b59918] Successfully created port: 1018a669-af2d-42cb-9558-0043b6d775b2 {{(pid=61957) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 884.103527] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ab9d633-ff2a-48a9-95ac-c5931659a832 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.112094] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-806f7bd7-af87-4a91-add4-08088611ecbf {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.151142] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-047ee284-ca6a-480d-85a5-7a47df3e80c2 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.155640] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-a3e467c9-b0de-4881-ac92-2329baed7e90 tempest-ServersTestJSON-1093633677 tempest-ServersTestJSON-1093633677-project-member] [instance: ead6aae1-36b5-4f57-9129-3bb02cf103ce] Unregistered the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 884.155898] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-a3e467c9-b0de-4881-ac92-2329baed7e90 tempest-ServersTestJSON-1093633677 tempest-ServersTestJSON-1093633677-project-member] [instance: ead6aae1-36b5-4f57-9129-3bb02cf103ce] Deleting contents of the VM from datastore datastore1 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 884.156099] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-a3e467c9-b0de-4881-ac92-2329baed7e90 tempest-ServersTestJSON-1093633677 tempest-ServersTestJSON-1093633677-project-member] Deleting the datastore file [datastore1] ead6aae1-36b5-4f57-9129-3bb02cf103ce {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 884.156770] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-47870845-4cb4-4eeb-8d9e-f13589173377 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.162395] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03858c55-d3e5-4eea-9bda-84913dd0056a {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.168140] env[61957]: DEBUG nova.compute.manager [None req-bafb915b-a249-4473-bf3a-4c41863ba98f tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] [instance: 28ab7d23-487f-4ae1-8fe4-58db55b59918] Start building block device mappings for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 884.171148] env[61957]: DEBUG oslo_vmware.api [None req-a3e467c9-b0de-4881-ac92-2329baed7e90 tempest-ServersTestJSON-1093633677 tempest-ServersTestJSON-1093633677-project-member] Waiting for the task: (returnval){ [ 884.171148] env[61957]: value = "task-1277621" [ 884.171148] env[61957]: _type = "Task" [ 884.171148] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.185131] env[61957]: DEBUG nova.compute.provider_tree [None req-a0f6bb40-5b1e-4b57-987f-3c85bacc19c6 tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 884.190651] env[61957]: DEBUG oslo_vmware.api [None req-a3e467c9-b0de-4881-ac92-2329baed7e90 tempest-ServersTestJSON-1093633677 tempest-ServersTestJSON-1093633677-project-member] Task: {'id': task-1277621, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.396531] env[61957]: DEBUG nova.compute.manager [None req-5a91d738-be3a-4667-949d-376fc04acd42 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: d4674fbc-47b1-42d1-aaba-e86d46c51e8f] Starting instance... {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 884.412473] env[61957]: DEBUG oslo_vmware.api [None req-4f69433c-53e0-4b6d-8d33-8d933f713309 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': task-1277620, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.448622] env[61957]: DEBUG nova.network.neutron [-] [instance: 8e04a098-25de-4d57-9e3c-ea44d234d57e] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 884.507346] env[61957]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e51b205e-0def-4ccf-aaf1-72a0be6e2989 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.517614] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f912975f-cf27-49e6-9c58-6eedd30f4016 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.546897] env[61957]: DEBUG nova.compute.manager [req-cbf3b15f-372b-4ba8-9cd4-be0b9c41ee6e req-ce5bf8c0-347c-444e-90b3-912bb47a242a service nova] [instance: 8e04a098-25de-4d57-9e3c-ea44d234d57e] Detach interface failed, port_id=55344e12-1398-4752-a0f4-fdea985786bc, reason: Instance 8e04a098-25de-4d57-9e3c-ea44d234d57e could not be found. {{(pid=61957) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 884.587788] env[61957]: DEBUG oslo_vmware.api [None req-06bfc4e2-99e2-4510-98f7-f5b198d6231a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': task-1277618, 'name': Rename_Task} progress is 14%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.687727] env[61957]: DEBUG nova.scheduler.client.report [None req-a0f6bb40-5b1e-4b57-987f-3c85bacc19c6 tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 884.690978] env[61957]: DEBUG oslo_vmware.api [None req-a3e467c9-b0de-4881-ac92-2329baed7e90 tempest-ServersTestJSON-1093633677 tempest-ServersTestJSON-1093633677-project-member] Task: {'id': task-1277621, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.259488} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.692601] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-a3e467c9-b0de-4881-ac92-2329baed7e90 tempest-ServersTestJSON-1093633677 tempest-ServersTestJSON-1093633677-project-member] Deleted the datastore file {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 884.694150] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-a3e467c9-b0de-4881-ac92-2329baed7e90 tempest-ServersTestJSON-1093633677 tempest-ServersTestJSON-1093633677-project-member] [instance: ead6aae1-36b5-4f57-9129-3bb02cf103ce] Deleted contents of the VM from datastore datastore1 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 884.694150] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-a3e467c9-b0de-4881-ac92-2329baed7e90 tempest-ServersTestJSON-1093633677 tempest-ServersTestJSON-1093633677-project-member] [instance: ead6aae1-36b5-4f57-9129-3bb02cf103ce] Instance destroyed {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 884.694150] env[61957]: INFO nova.compute.manager [None req-a3e467c9-b0de-4881-ac92-2329baed7e90 tempest-ServersTestJSON-1093633677 tempest-ServersTestJSON-1093633677-project-member] [instance: ead6aae1-36b5-4f57-9129-3bb02cf103ce] Took 2.04 seconds to destroy the instance on the hypervisor. [ 884.694150] env[61957]: DEBUG oslo.service.loopingcall [None req-a3e467c9-b0de-4881-ac92-2329baed7e90 tempest-ServersTestJSON-1093633677 tempest-ServersTestJSON-1093633677-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 884.694150] env[61957]: DEBUG nova.compute.manager [-] [instance: ead6aae1-36b5-4f57-9129-3bb02cf103ce] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 884.694150] env[61957]: DEBUG nova.network.neutron [-] [instance: ead6aae1-36b5-4f57-9129-3bb02cf103ce] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 884.919817] env[61957]: DEBUG oslo_vmware.api [None req-4f69433c-53e0-4b6d-8d33-8d933f713309 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': task-1277620, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.928710] env[61957]: DEBUG oslo_concurrency.lockutils [None req-5a91d738-be3a-4667-949d-376fc04acd42 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 884.951504] env[61957]: INFO nova.compute.manager [-] [instance: 8e04a098-25de-4d57-9e3c-ea44d234d57e] Took 1.37 seconds to deallocate network for instance. [ 885.094377] env[61957]: DEBUG oslo_vmware.api [None req-06bfc4e2-99e2-4510-98f7-f5b198d6231a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': task-1277618, 'name': Rename_Task, 'duration_secs': 1.227564} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.095670] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-06bfc4e2-99e2-4510-98f7-f5b198d6231a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 1a0327c2-5671-4970-9db7-c7cc912d8678] Powering on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 885.096032] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c07c2136-acdf-4653-91ca-4fbaa25fda8a {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.104829] env[61957]: DEBUG oslo_vmware.api [None req-06bfc4e2-99e2-4510-98f7-f5b198d6231a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Waiting for the task: (returnval){ [ 885.104829] env[61957]: value = "task-1277622" [ 885.104829] env[61957]: _type = "Task" [ 885.104829] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.117066] env[61957]: DEBUG oslo_vmware.api [None req-06bfc4e2-99e2-4510-98f7-f5b198d6231a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': task-1277622, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.181873] env[61957]: DEBUG nova.compute.manager [None req-bafb915b-a249-4473-bf3a-4c41863ba98f tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] [instance: 28ab7d23-487f-4ae1-8fe4-58db55b59918] Start spawning the instance on the hypervisor. {{(pid=61957) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 885.194560] env[61957]: DEBUG oslo_concurrency.lockutils [None req-a0f6bb40-5b1e-4b57-987f-3c85bacc19c6 tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.042s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 885.196830] env[61957]: DEBUG oslo_concurrency.lockutils [None req-95a860a5-ca00-4c71-b827-6e3af6a25042 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 23.896s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 885.197079] env[61957]: DEBUG nova.objects.instance [None req-95a860a5-ca00-4c71-b827-6e3af6a25042 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Lazy-loading 'resources' on Instance uuid 37ce46a1-8f9a-4d15-bd81-e40845a0e48a {{(pid=61957) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 885.215561] env[61957]: DEBUG nova.virt.hardware [None req-bafb915b-a249-4473-bf3a-4c41863ba98f tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T17:22:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T17:21:50Z,direct_url=,disk_format='vmdk',id=11c76a2c-f705-470a-ba9d-4657858bab38,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='81d3096d0f094373913fc3dc8f5c3c6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T17:21:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 885.215924] env[61957]: DEBUG nova.virt.hardware [None req-bafb915b-a249-4473-bf3a-4c41863ba98f tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Flavor limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 885.216196] env[61957]: DEBUG nova.virt.hardware [None req-bafb915b-a249-4473-bf3a-4c41863ba98f tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Image limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 885.216489] env[61957]: DEBUG nova.virt.hardware [None req-bafb915b-a249-4473-bf3a-4c41863ba98f tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Flavor pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 885.216715] env[61957]: DEBUG nova.virt.hardware [None req-bafb915b-a249-4473-bf3a-4c41863ba98f tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Image pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 885.217025] env[61957]: DEBUG nova.virt.hardware [None req-bafb915b-a249-4473-bf3a-4c41863ba98f tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 885.217332] env[61957]: DEBUG nova.virt.hardware [None req-bafb915b-a249-4473-bf3a-4c41863ba98f tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 885.217512] env[61957]: DEBUG nova.virt.hardware [None req-bafb915b-a249-4473-bf3a-4c41863ba98f tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 885.217744] env[61957]: DEBUG nova.virt.hardware [None req-bafb915b-a249-4473-bf3a-4c41863ba98f tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Got 1 possible topologies {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 885.219049] env[61957]: DEBUG nova.virt.hardware [None req-bafb915b-a249-4473-bf3a-4c41863ba98f tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 885.219049] env[61957]: DEBUG nova.virt.hardware [None req-bafb915b-a249-4473-bf3a-4c41863ba98f tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 885.219663] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37275c5d-2973-41fb-8f66-ef3a4af960c8 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.224302] env[61957]: INFO nova.scheduler.client.report [None req-a0f6bb40-5b1e-4b57-987f-3c85bacc19c6 tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Deleted allocations for instance e0249936-d616-4ffb-8f77-d8107633c42a [ 885.234278] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9b51df8-f4c5-44fe-8616-031713acfc2f {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.413798] env[61957]: DEBUG oslo_vmware.api [None req-4f69433c-53e0-4b6d-8d33-8d933f713309 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': task-1277620, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.458721] env[61957]: DEBUG oslo_concurrency.lockutils [None req-9513c200-94de-43f4-aaa5-418ecbddf25f tempest-ServerTagsTestJSON-969152625 tempest-ServerTagsTestJSON-969152625-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 885.617552] env[61957]: DEBUG oslo_vmware.api [None req-06bfc4e2-99e2-4510-98f7-f5b198d6231a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': task-1277622, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.732022] env[61957]: DEBUG nova.network.neutron [None req-bafb915b-a249-4473-bf3a-4c41863ba98f tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] [instance: 28ab7d23-487f-4ae1-8fe4-58db55b59918] Successfully updated port: 1018a669-af2d-42cb-9558-0043b6d775b2 {{(pid=61957) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 885.738029] env[61957]: DEBUG nova.network.neutron [-] [instance: ead6aae1-36b5-4f57-9129-3bb02cf103ce] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 885.738029] env[61957]: DEBUG oslo_concurrency.lockutils [None req-a0f6bb40-5b1e-4b57-987f-3c85bacc19c6 tempest-ImagesOneServerTestJSON-872155793 tempest-ImagesOneServerTestJSON-872155793-project-member] Lock "e0249936-d616-4ffb-8f77-d8107633c42a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 27.992s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 885.916023] env[61957]: DEBUG oslo_vmware.api [None req-4f69433c-53e0-4b6d-8d33-8d933f713309 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': task-1277620, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.023578] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d51ea73-597c-48d5-aac7-a512f390725b {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.031503] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9c2c434-7af6-44ff-96e3-bcaac2c50dc8 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.037040] env[61957]: DEBUG nova.compute.manager [req-c201e605-5680-4153-b859-31c0d437c793 req-88a3bf65-3a55-42d6-8aed-d36162a40cfe service nova] [instance: ead6aae1-36b5-4f57-9129-3bb02cf103ce] Received event network-vif-deleted-606b786e-f948-4d1c-a109-6390d9349ea8 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 886.037256] env[61957]: DEBUG nova.compute.manager [req-c201e605-5680-4153-b859-31c0d437c793 req-88a3bf65-3a55-42d6-8aed-d36162a40cfe service nova] [instance: 28ab7d23-487f-4ae1-8fe4-58db55b59918] Received event network-vif-plugged-1018a669-af2d-42cb-9558-0043b6d775b2 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 886.037450] env[61957]: DEBUG oslo_concurrency.lockutils [req-c201e605-5680-4153-b859-31c0d437c793 req-88a3bf65-3a55-42d6-8aed-d36162a40cfe service nova] Acquiring lock "28ab7d23-487f-4ae1-8fe4-58db55b59918-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 886.037656] env[61957]: DEBUG oslo_concurrency.lockutils [req-c201e605-5680-4153-b859-31c0d437c793 req-88a3bf65-3a55-42d6-8aed-d36162a40cfe service nova] Lock "28ab7d23-487f-4ae1-8fe4-58db55b59918-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 886.037824] env[61957]: DEBUG oslo_concurrency.lockutils [req-c201e605-5680-4153-b859-31c0d437c793 req-88a3bf65-3a55-42d6-8aed-d36162a40cfe service nova] Lock "28ab7d23-487f-4ae1-8fe4-58db55b59918-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 886.037988] env[61957]: DEBUG nova.compute.manager [req-c201e605-5680-4153-b859-31c0d437c793 req-88a3bf65-3a55-42d6-8aed-d36162a40cfe service nova] [instance: 28ab7d23-487f-4ae1-8fe4-58db55b59918] No waiting events found dispatching network-vif-plugged-1018a669-af2d-42cb-9558-0043b6d775b2 {{(pid=61957) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 886.038169] env[61957]: WARNING nova.compute.manager [req-c201e605-5680-4153-b859-31c0d437c793 req-88a3bf65-3a55-42d6-8aed-d36162a40cfe service nova] [instance: 28ab7d23-487f-4ae1-8fe4-58db55b59918] Received unexpected event network-vif-plugged-1018a669-af2d-42cb-9558-0043b6d775b2 for instance with vm_state building and task_state spawning. [ 886.038340] env[61957]: DEBUG nova.compute.manager [req-c201e605-5680-4153-b859-31c0d437c793 req-88a3bf65-3a55-42d6-8aed-d36162a40cfe service nova] [instance: 28ab7d23-487f-4ae1-8fe4-58db55b59918] Received event network-changed-1018a669-af2d-42cb-9558-0043b6d775b2 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 886.038496] env[61957]: DEBUG nova.compute.manager [req-c201e605-5680-4153-b859-31c0d437c793 req-88a3bf65-3a55-42d6-8aed-d36162a40cfe service nova] [instance: 28ab7d23-487f-4ae1-8fe4-58db55b59918] Refreshing instance network info cache due to event network-changed-1018a669-af2d-42cb-9558-0043b6d775b2. {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 886.038679] env[61957]: DEBUG oslo_concurrency.lockutils [req-c201e605-5680-4153-b859-31c0d437c793 req-88a3bf65-3a55-42d6-8aed-d36162a40cfe service nova] Acquiring lock "refresh_cache-28ab7d23-487f-4ae1-8fe4-58db55b59918" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 886.038818] env[61957]: DEBUG oslo_concurrency.lockutils [req-c201e605-5680-4153-b859-31c0d437c793 req-88a3bf65-3a55-42d6-8aed-d36162a40cfe service nova] Acquired lock "refresh_cache-28ab7d23-487f-4ae1-8fe4-58db55b59918" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 886.038972] env[61957]: DEBUG nova.network.neutron [req-c201e605-5680-4153-b859-31c0d437c793 req-88a3bf65-3a55-42d6-8aed-d36162a40cfe service nova] [instance: 28ab7d23-487f-4ae1-8fe4-58db55b59918] Refreshing network info cache for port 1018a669-af2d-42cb-9558-0043b6d775b2 {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 886.072528] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a31df9a-e5f7-401c-a04a-a085f3f00c09 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.085021] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b79c2623-88ae-4c8e-a7d1-9e0cd0c9984c {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.098227] env[61957]: DEBUG nova.compute.provider_tree [None req-95a860a5-ca00-4c71-b827-6e3af6a25042 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 886.117451] env[61957]: DEBUG oslo_vmware.api [None req-06bfc4e2-99e2-4510-98f7-f5b198d6231a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': task-1277622, 'name': PowerOnVM_Task, 'duration_secs': 0.588972} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.117736] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-06bfc4e2-99e2-4510-98f7-f5b198d6231a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 1a0327c2-5671-4970-9db7-c7cc912d8678] Powered on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 886.118339] env[61957]: INFO nova.compute.manager [None req-06bfc4e2-99e2-4510-98f7-f5b198d6231a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 1a0327c2-5671-4970-9db7-c7cc912d8678] Took 9.59 seconds to spawn the instance on the hypervisor. [ 886.118339] env[61957]: DEBUG nova.compute.manager [None req-06bfc4e2-99e2-4510-98f7-f5b198d6231a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 1a0327c2-5671-4970-9db7-c7cc912d8678] Checking state {{(pid=61957) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 886.121332] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17e2f2f4-2afa-46c5-8970-f62567ecbd4c {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.238154] env[61957]: DEBUG oslo_concurrency.lockutils [None req-bafb915b-a249-4473-bf3a-4c41863ba98f tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Acquiring lock "refresh_cache-28ab7d23-487f-4ae1-8fe4-58db55b59918" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 886.238876] env[61957]: INFO nova.compute.manager [-] [instance: ead6aae1-36b5-4f57-9129-3bb02cf103ce] Took 1.55 seconds to deallocate network for instance. [ 886.416213] env[61957]: DEBUG oslo_vmware.api [None req-4f69433c-53e0-4b6d-8d33-8d933f713309 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': task-1277620, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.582577] env[61957]: DEBUG nova.network.neutron [req-c201e605-5680-4153-b859-31c0d437c793 req-88a3bf65-3a55-42d6-8aed-d36162a40cfe service nova] [instance: 28ab7d23-487f-4ae1-8fe4-58db55b59918] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 886.601824] env[61957]: DEBUG nova.scheduler.client.report [None req-95a860a5-ca00-4c71-b827-6e3af6a25042 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 886.640908] env[61957]: INFO nova.compute.manager [None req-06bfc4e2-99e2-4510-98f7-f5b198d6231a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 1a0327c2-5671-4970-9db7-c7cc912d8678] Took 39.92 seconds to build instance. [ 886.701935] env[61957]: DEBUG nova.network.neutron [req-c201e605-5680-4153-b859-31c0d437c793 req-88a3bf65-3a55-42d6-8aed-d36162a40cfe service nova] [instance: 28ab7d23-487f-4ae1-8fe4-58db55b59918] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 886.746088] env[61957]: DEBUG oslo_concurrency.lockutils [None req-a3e467c9-b0de-4881-ac92-2329baed7e90 tempest-ServersTestJSON-1093633677 tempest-ServersTestJSON-1093633677-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 886.915968] env[61957]: DEBUG oslo_vmware.api [None req-4f69433c-53e0-4b6d-8d33-8d933f713309 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': task-1277620, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.107555] env[61957]: DEBUG oslo_concurrency.lockutils [None req-95a860a5-ca00-4c71-b827-6e3af6a25042 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.911s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 887.111339] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ca59f931-a2d6-4c9e-954c-a383ffdeea61 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.772s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 887.114543] env[61957]: INFO nova.compute.claims [None req-ca59f931-a2d6-4c9e-954c-a383ffdeea61 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] [instance: d5d10a31-0e78-4ed7-b944-9208138a4861] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 887.138526] env[61957]: INFO nova.scheduler.client.report [None req-95a860a5-ca00-4c71-b827-6e3af6a25042 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Deleted allocations for instance 37ce46a1-8f9a-4d15-bd81-e40845a0e48a [ 887.143733] env[61957]: DEBUG oslo_concurrency.lockutils [None req-06bfc4e2-99e2-4510-98f7-f5b198d6231a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Lock "1a0327c2-5671-4970-9db7-c7cc912d8678" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 68.506s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 887.208142] env[61957]: DEBUG oslo_concurrency.lockutils [req-c201e605-5680-4153-b859-31c0d437c793 req-88a3bf65-3a55-42d6-8aed-d36162a40cfe service nova] Releasing lock "refresh_cache-28ab7d23-487f-4ae1-8fe4-58db55b59918" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 887.208142] env[61957]: DEBUG oslo_concurrency.lockutils [None req-bafb915b-a249-4473-bf3a-4c41863ba98f tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Acquired lock "refresh_cache-28ab7d23-487f-4ae1-8fe4-58db55b59918" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 887.208142] env[61957]: DEBUG nova.network.neutron [None req-bafb915b-a249-4473-bf3a-4c41863ba98f tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] [instance: 28ab7d23-487f-4ae1-8fe4-58db55b59918] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 887.417611] env[61957]: DEBUG oslo_vmware.api [None req-4f69433c-53e0-4b6d-8d33-8d933f713309 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': task-1277620, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.646862] env[61957]: DEBUG oslo_concurrency.lockutils [None req-95a860a5-ca00-4c71-b827-6e3af6a25042 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Lock "37ce46a1-8f9a-4d15-bd81-e40845a0e48a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 29.655s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 887.750852] env[61957]: DEBUG nova.network.neutron [None req-bafb915b-a249-4473-bf3a-4c41863ba98f tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] [instance: 28ab7d23-487f-4ae1-8fe4-58db55b59918] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 887.916461] env[61957]: DEBUG oslo_vmware.api [None req-4f69433c-53e0-4b6d-8d33-8d933f713309 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': task-1277620, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.960161] env[61957]: DEBUG nova.network.neutron [None req-bafb915b-a249-4473-bf3a-4c41863ba98f tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] [instance: 28ab7d23-487f-4ae1-8fe4-58db55b59918] Updating instance_info_cache with network_info: [{"id": "1018a669-af2d-42cb-9558-0043b6d775b2", "address": "fa:16:3e:1b:1f:70", "network": {"id": "41a1701c-d9b6-4414-9e92-2e999cc916e3", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-727135203-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d849d922cfad41848a33e8d3bc83ea0f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f1b507ed-cd2d-4c09-9d96-c47bde6a7774", "external-id": "nsx-vlan-transportzone-980", "segmentation_id": 980, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1018a669-af", "ovs_interfaceid": "1018a669-af2d-42cb-9558-0043b6d775b2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 888.418113] env[61957]: DEBUG oslo_vmware.api [None req-4f69433c-53e0-4b6d-8d33-8d933f713309 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': task-1277620, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.467020] env[61957]: DEBUG oslo_concurrency.lockutils [None req-bafb915b-a249-4473-bf3a-4c41863ba98f tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Releasing lock "refresh_cache-28ab7d23-487f-4ae1-8fe4-58db55b59918" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 888.467433] env[61957]: DEBUG nova.compute.manager [None req-bafb915b-a249-4473-bf3a-4c41863ba98f tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] [instance: 28ab7d23-487f-4ae1-8fe4-58db55b59918] Instance network_info: |[{"id": "1018a669-af2d-42cb-9558-0043b6d775b2", "address": "fa:16:3e:1b:1f:70", "network": {"id": "41a1701c-d9b6-4414-9e92-2e999cc916e3", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-727135203-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d849d922cfad41848a33e8d3bc83ea0f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f1b507ed-cd2d-4c09-9d96-c47bde6a7774", "external-id": "nsx-vlan-transportzone-980", "segmentation_id": 980, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1018a669-af", "ovs_interfaceid": "1018a669-af2d-42cb-9558-0043b6d775b2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 888.468197] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-bafb915b-a249-4473-bf3a-4c41863ba98f tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] [instance: 28ab7d23-487f-4ae1-8fe4-58db55b59918] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:1b:1f:70', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f1b507ed-cd2d-4c09-9d96-c47bde6a7774', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1018a669-af2d-42cb-9558-0043b6d775b2', 'vif_model': 'vmxnet3'}] {{(pid=61957) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 888.477083] env[61957]: DEBUG oslo.service.loopingcall [None req-bafb915b-a249-4473-bf3a-4c41863ba98f tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 888.477361] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 28ab7d23-487f-4ae1-8fe4-58db55b59918] Creating VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 888.477610] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-22cb2365-90f5-4774-95e6-21e2df84deab {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.494990] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-606b096a-2386-4a06-9cf7-069cdea8e5da {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.503692] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9da54175-3ea5-4843-b141-1382b2facb63 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.508290] env[61957]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 888.508290] env[61957]: value = "task-1277623" [ 888.508290] env[61957]: _type = "Task" [ 888.508290] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.541471] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e31fbf10-4719-447f-ad0a-ad29f9f12bba {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.544158] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277623, 'name': CreateVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.551759] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bb99811-2fa3-4c82-9425-5ba285f5bd61 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.566904] env[61957]: DEBUG nova.compute.provider_tree [None req-ca59f931-a2d6-4c9e-954c-a383ffdeea61 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 888.922256] env[61957]: DEBUG oslo_vmware.api [None req-4f69433c-53e0-4b6d-8d33-8d933f713309 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': task-1277620, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.023854] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277623, 'name': CreateVM_Task} progress is 99%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.070160] env[61957]: DEBUG nova.scheduler.client.report [None req-ca59f931-a2d6-4c9e-954c-a383ffdeea61 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 889.419559] env[61957]: DEBUG oslo_vmware.api [None req-4f69433c-53e0-4b6d-8d33-8d933f713309 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': task-1277620, 'name': ReconfigVM_Task} progress is 18%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.487787] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d61e8152-d9da-413c-b7fd-8eed0b01a69d tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Acquiring lock "b598b8bb-919e-4404-b264-7b76161b0f79" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 889.488059] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d61e8152-d9da-413c-b7fd-8eed0b01a69d tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Lock "b598b8bb-919e-4404-b264-7b76161b0f79" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 889.525319] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277623, 'name': CreateVM_Task, 'duration_secs': 0.639444} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.525569] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 28ab7d23-487f-4ae1-8fe4-58db55b59918] Created VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 889.526554] env[61957]: DEBUG oslo_concurrency.lockutils [None req-bafb915b-a249-4473-bf3a-4c41863ba98f tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 889.526731] env[61957]: DEBUG oslo_concurrency.lockutils [None req-bafb915b-a249-4473-bf3a-4c41863ba98f tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 889.527102] env[61957]: DEBUG oslo_concurrency.lockutils [None req-bafb915b-a249-4473-bf3a-4c41863ba98f tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 889.527380] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-94972de3-b7b7-463c-a6f2-d864c5df2537 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.533230] env[61957]: DEBUG oslo_vmware.api [None req-bafb915b-a249-4473-bf3a-4c41863ba98f tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Waiting for the task: (returnval){ [ 889.533230] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52f0a8a1-306e-6c47-4b0d-ce8307590327" [ 889.533230] env[61957]: _type = "Task" [ 889.533230] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.544790] env[61957]: DEBUG oslo_vmware.api [None req-bafb915b-a249-4473-bf3a-4c41863ba98f tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52f0a8a1-306e-6c47-4b0d-ce8307590327, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.577862] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ca59f931-a2d6-4c9e-954c-a383ffdeea61 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.466s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 889.578651] env[61957]: DEBUG nova.compute.manager [None req-ca59f931-a2d6-4c9e-954c-a383ffdeea61 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] [instance: d5d10a31-0e78-4ed7-b944-9208138a4861] Start building networks asynchronously for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 889.582242] env[61957]: DEBUG oslo_concurrency.lockutils [None req-7916de58-8954-4a2d-b3f9-5ec9b9a16797 tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.729s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 889.583777] env[61957]: INFO nova.compute.claims [None req-7916de58-8954-4a2d-b3f9-5ec9b9a16797 tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] [instance: 0d0f2d34-de35-4e80-8d9f-12693add0786] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 889.922707] env[61957]: DEBUG oslo_vmware.api [None req-4f69433c-53e0-4b6d-8d33-8d933f713309 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': task-1277620, 'name': ReconfigVM_Task, 'duration_secs': 5.850387} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.923157] env[61957]: DEBUG oslo_concurrency.lockutils [None req-4f69433c-53e0-4b6d-8d33-8d933f713309 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Releasing lock "4951132e-7247-4772-8f88-3664c6a7e61e" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 889.923433] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-4f69433c-53e0-4b6d-8d33-8d933f713309 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 4951132e-7247-4772-8f88-3664c6a7e61e] Reconfigured VM to detach interface {{(pid=61957) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1984}} [ 889.941475] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d28d4db3-4ba8-4b6b-b020-dcb0a978b0ab tempest-InstanceActionsNegativeTestJSON-259524954 tempest-InstanceActionsNegativeTestJSON-259524954-project-member] Acquiring lock "7e2db5f6-1419-415e-b957-66cf2e3ec19d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 889.943701] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d28d4db3-4ba8-4b6b-b020-dcb0a978b0ab tempest-InstanceActionsNegativeTestJSON-259524954 tempest-InstanceActionsNegativeTestJSON-259524954-project-member] Lock "7e2db5f6-1419-415e-b957-66cf2e3ec19d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.002s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 889.994234] env[61957]: DEBUG nova.compute.manager [None req-d61e8152-d9da-413c-b7fd-8eed0b01a69d tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: b598b8bb-919e-4404-b264-7b76161b0f79] Starting instance... {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 890.050573] env[61957]: DEBUG oslo_vmware.api [None req-bafb915b-a249-4473-bf3a-4c41863ba98f tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52f0a8a1-306e-6c47-4b0d-ce8307590327, 'name': SearchDatastore_Task, 'duration_secs': 0.028753} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.050847] env[61957]: DEBUG oslo_concurrency.lockutils [None req-bafb915b-a249-4473-bf3a-4c41863ba98f tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 890.050977] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-bafb915b-a249-4473-bf3a-4c41863ba98f tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] [instance: 28ab7d23-487f-4ae1-8fe4-58db55b59918] Processing image 11c76a2c-f705-470a-ba9d-4657858bab38 {{(pid=61957) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 890.051186] env[61957]: DEBUG oslo_concurrency.lockutils [None req-bafb915b-a249-4473-bf3a-4c41863ba98f tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 890.051341] env[61957]: DEBUG oslo_concurrency.lockutils [None req-bafb915b-a249-4473-bf3a-4c41863ba98f tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 890.051538] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-bafb915b-a249-4473-bf3a-4c41863ba98f tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 890.052259] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8ff64b9e-00d0-4a82-a092-4e4f21bf786e {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.065832] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-bafb915b-a249-4473-bf3a-4c41863ba98f tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 890.065832] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-bafb915b-a249-4473-bf3a-4c41863ba98f tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61957) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 890.067090] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-030e19e0-6122-4804-961c-6ee95338a444 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.076257] env[61957]: DEBUG oslo_vmware.api [None req-bafb915b-a249-4473-bf3a-4c41863ba98f tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Waiting for the task: (returnval){ [ 890.076257] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]520a961b-cf2a-eff1-37ca-d45e773c392a" [ 890.076257] env[61957]: _type = "Task" [ 890.076257] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.083885] env[61957]: DEBUG nova.compute.utils [None req-ca59f931-a2d6-4c9e-954c-a383ffdeea61 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Using /dev/sd instead of None {{(pid=61957) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 890.093036] env[61957]: DEBUG nova.compute.manager [None req-ca59f931-a2d6-4c9e-954c-a383ffdeea61 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] [instance: d5d10a31-0e78-4ed7-b944-9208138a4861] Allocating IP information in the background. {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 890.093036] env[61957]: DEBUG nova.network.neutron [None req-ca59f931-a2d6-4c9e-954c-a383ffdeea61 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] [instance: d5d10a31-0e78-4ed7-b944-9208138a4861] allocate_for_instance() {{(pid=61957) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 890.097790] env[61957]: DEBUG oslo_vmware.api [None req-bafb915b-a249-4473-bf3a-4c41863ba98f tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]520a961b-cf2a-eff1-37ca-d45e773c392a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.170034] env[61957]: DEBUG nova.policy [None req-ca59f931-a2d6-4c9e-954c-a383ffdeea61 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9e3ccf0e318349b38c7eff2c89440e54', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd849d922cfad41848a33e8d3bc83ea0f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61957) authorize /opt/stack/nova/nova/policy.py:203}} [ 890.340106] env[61957]: DEBUG nova.compute.manager [req-199b7f90-2972-4864-adf3-5ad49e60179c req-85df7269-e605-440b-9088-0869b666b9a7 service nova] [instance: 4951132e-7247-4772-8f88-3664c6a7e61e] Received event network-vif-deleted-a093d1c9-1f94-4000-8f87-f76e9db98ec8 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 890.340409] env[61957]: INFO nova.compute.manager [req-199b7f90-2972-4864-adf3-5ad49e60179c req-85df7269-e605-440b-9088-0869b666b9a7 service nova] [instance: 4951132e-7247-4772-8f88-3664c6a7e61e] Neutron deleted interface a093d1c9-1f94-4000-8f87-f76e9db98ec8; detaching it from the instance and deleting it from the info cache [ 890.340642] env[61957]: DEBUG nova.network.neutron [req-199b7f90-2972-4864-adf3-5ad49e60179c req-85df7269-e605-440b-9088-0869b666b9a7 service nova] [instance: 4951132e-7247-4772-8f88-3664c6a7e61e] Updating instance_info_cache with network_info: [{"id": "3016c8eb-2ad3-40b1-8d88-eaec447de9ea", "address": "fa:16:3e:59:19:f2", "network": {"id": "364cabcb-046d-4f91-b1ee-5dd3adfae6d7", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1432396967-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.193", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8713aa35bcb24b86ad0b58ca9fc991ba", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b7a73c01-1bb9-4612-a1a7-16d71b732e81", "external-id": "nsx-vlan-transportzone-711", "segmentation_id": 711, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3016c8eb-2a", "ovs_interfaceid": "3016c8eb-2ad3-40b1-8d88-eaec447de9ea", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 890.525255] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d61e8152-d9da-413c-b7fd-8eed0b01a69d tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 890.543018] env[61957]: DEBUG nova.network.neutron [None req-ca59f931-a2d6-4c9e-954c-a383ffdeea61 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] [instance: d5d10a31-0e78-4ed7-b944-9208138a4861] Successfully created port: 572549f2-6ec4-411e-8905-0c33119b31c8 {{(pid=61957) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 890.589938] env[61957]: DEBUG oslo_vmware.api [None req-bafb915b-a249-4473-bf3a-4c41863ba98f tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]520a961b-cf2a-eff1-37ca-d45e773c392a, 'name': SearchDatastore_Task, 'duration_secs': 0.018539} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.590966] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ed993272-1171-46ba-8fe1-025371cbb0e6 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.598259] env[61957]: DEBUG oslo_vmware.api [None req-bafb915b-a249-4473-bf3a-4c41863ba98f tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Waiting for the task: (returnval){ [ 890.598259] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52c1ce98-6049-9c2d-6876-a10d8bc8db15" [ 890.598259] env[61957]: _type = "Task" [ 890.598259] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.599920] env[61957]: DEBUG nova.compute.manager [None req-ca59f931-a2d6-4c9e-954c-a383ffdeea61 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] [instance: d5d10a31-0e78-4ed7-b944-9208138a4861] Start building block device mappings for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 890.626918] env[61957]: DEBUG oslo_vmware.api [None req-bafb915b-a249-4473-bf3a-4c41863ba98f tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52c1ce98-6049-9c2d-6876-a10d8bc8db15, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.845050] env[61957]: DEBUG oslo_concurrency.lockutils [req-199b7f90-2972-4864-adf3-5ad49e60179c req-85df7269-e605-440b-9088-0869b666b9a7 service nova] Acquiring lock "4951132e-7247-4772-8f88-3664c6a7e61e" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 890.845441] env[61957]: DEBUG oslo_concurrency.lockutils [req-199b7f90-2972-4864-adf3-5ad49e60179c req-85df7269-e605-440b-9088-0869b666b9a7 service nova] Acquired lock "4951132e-7247-4772-8f88-3664c6a7e61e" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 890.850121] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-039d55a1-51a5-4118-baee-d2bc7dbb8ff4 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.870873] env[61957]: DEBUG oslo_concurrency.lockutils [req-199b7f90-2972-4864-adf3-5ad49e60179c req-85df7269-e605-440b-9088-0869b666b9a7 service nova] Releasing lock "4951132e-7247-4772-8f88-3664c6a7e61e" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 890.871149] env[61957]: WARNING nova.compute.manager [req-199b7f90-2972-4864-adf3-5ad49e60179c req-85df7269-e605-440b-9088-0869b666b9a7 service nova] [instance: 4951132e-7247-4772-8f88-3664c6a7e61e] Detach interface failed, port_id=a093d1c9-1f94-4000-8f87-f76e9db98ec8, reason: No device with interface-id a093d1c9-1f94-4000-8f87-f76e9db98ec8 exists on VM: nova.exception.NotFound: No device with interface-id a093d1c9-1f94-4000-8f87-f76e9db98ec8 exists on VM [ 891.041753] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fc651f5-f042-4ae3-bdfd-ce56f581efd9 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.052080] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a353e58-4ec7-457a-a156-5d9b1adef428 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.089871] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ebeaadb-65c9-404f-85db-42ddfbdf1a11 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.100270] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1941b2f6-3e5f-4977-b818-30626b715ad0 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.119571] env[61957]: DEBUG oslo_vmware.api [None req-bafb915b-a249-4473-bf3a-4c41863ba98f tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52c1ce98-6049-9c2d-6876-a10d8bc8db15, 'name': SearchDatastore_Task, 'duration_secs': 0.027801} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.127788] env[61957]: DEBUG oslo_concurrency.lockutils [None req-bafb915b-a249-4473-bf3a-4c41863ba98f tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 891.127969] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-bafb915b-a249-4473-bf3a-4c41863ba98f tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore2] 28ab7d23-487f-4ae1-8fe4-58db55b59918/28ab7d23-487f-4ae1-8fe4-58db55b59918.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 891.128446] env[61957]: DEBUG nova.compute.provider_tree [None req-7916de58-8954-4a2d-b3f9-5ec9b9a16797 tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 891.130084] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2e18e76c-19d8-41e0-b9cd-4b4632860388 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.138726] env[61957]: DEBUG oslo_vmware.api [None req-bafb915b-a249-4473-bf3a-4c41863ba98f tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Waiting for the task: (returnval){ [ 891.138726] env[61957]: value = "task-1277624" [ 891.138726] env[61957]: _type = "Task" [ 891.138726] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.149070] env[61957]: DEBUG oslo_vmware.api [None req-bafb915b-a249-4473-bf3a-4c41863ba98f tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Task: {'id': task-1277624, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.309898] env[61957]: DEBUG oslo_concurrency.lockutils [None req-4f69433c-53e0-4b6d-8d33-8d933f713309 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Acquiring lock "refresh_cache-4951132e-7247-4772-8f88-3664c6a7e61e" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 891.310159] env[61957]: DEBUG oslo_concurrency.lockutils [None req-4f69433c-53e0-4b6d-8d33-8d933f713309 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Acquired lock "refresh_cache-4951132e-7247-4772-8f88-3664c6a7e61e" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 891.311724] env[61957]: DEBUG nova.network.neutron [None req-4f69433c-53e0-4b6d-8d33-8d933f713309 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 4951132e-7247-4772-8f88-3664c6a7e61e] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 891.453427] env[61957]: DEBUG oslo_concurrency.lockutils [None req-9314189b-c158-4aed-834e-ffe55cc4ba51 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Acquiring lock "4951132e-7247-4772-8f88-3664c6a7e61e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 891.453427] env[61957]: DEBUG oslo_concurrency.lockutils [None req-9314189b-c158-4aed-834e-ffe55cc4ba51 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Lock "4951132e-7247-4772-8f88-3664c6a7e61e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 891.453427] env[61957]: DEBUG oslo_concurrency.lockutils [None req-9314189b-c158-4aed-834e-ffe55cc4ba51 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Acquiring lock "4951132e-7247-4772-8f88-3664c6a7e61e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 891.453427] env[61957]: DEBUG oslo_concurrency.lockutils [None req-9314189b-c158-4aed-834e-ffe55cc4ba51 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Lock "4951132e-7247-4772-8f88-3664c6a7e61e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 891.453427] env[61957]: DEBUG oslo_concurrency.lockutils [None req-9314189b-c158-4aed-834e-ffe55cc4ba51 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Lock "4951132e-7247-4772-8f88-3664c6a7e61e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 891.455846] env[61957]: INFO nova.compute.manager [None req-9314189b-c158-4aed-834e-ffe55cc4ba51 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 4951132e-7247-4772-8f88-3664c6a7e61e] Terminating instance [ 891.457870] env[61957]: DEBUG nova.compute.manager [None req-9314189b-c158-4aed-834e-ffe55cc4ba51 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 4951132e-7247-4772-8f88-3664c6a7e61e] Start destroying the instance on the hypervisor. {{(pid=61957) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 891.458084] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-9314189b-c158-4aed-834e-ffe55cc4ba51 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 4951132e-7247-4772-8f88-3664c6a7e61e] Destroying instance {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 891.458919] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bba9c98e-4c6e-4d7b-818e-0a68846b15e9 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.471208] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-9314189b-c158-4aed-834e-ffe55cc4ba51 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 4951132e-7247-4772-8f88-3664c6a7e61e] Powering off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 891.471208] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e95f31c2-1d87-4259-948d-0c9f8aafd43d {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.480162] env[61957]: DEBUG oslo_vmware.api [None req-9314189b-c158-4aed-834e-ffe55cc4ba51 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Waiting for the task: (returnval){ [ 891.480162] env[61957]: value = "task-1277625" [ 891.480162] env[61957]: _type = "Task" [ 891.480162] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.494447] env[61957]: DEBUG oslo_vmware.api [None req-9314189b-c158-4aed-834e-ffe55cc4ba51 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': task-1277625, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.621721] env[61957]: DEBUG nova.compute.manager [None req-ca59f931-a2d6-4c9e-954c-a383ffdeea61 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] [instance: d5d10a31-0e78-4ed7-b944-9208138a4861] Start spawning the instance on the hypervisor. {{(pid=61957) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 891.634332] env[61957]: DEBUG nova.scheduler.client.report [None req-7916de58-8954-4a2d-b3f9-5ec9b9a16797 tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 891.663300] env[61957]: DEBUG oslo_vmware.api [None req-bafb915b-a249-4473-bf3a-4c41863ba98f tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Task: {'id': task-1277624, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.489499} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.669835] env[61957]: DEBUG nova.virt.hardware [None req-ca59f931-a2d6-4c9e-954c-a383ffdeea61 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T17:22:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T17:21:50Z,direct_url=,disk_format='vmdk',id=11c76a2c-f705-470a-ba9d-4657858bab38,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='81d3096d0f094373913fc3dc8f5c3c6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T17:21:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 891.670125] env[61957]: DEBUG nova.virt.hardware [None req-ca59f931-a2d6-4c9e-954c-a383ffdeea61 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Flavor limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 891.670309] env[61957]: DEBUG nova.virt.hardware [None req-ca59f931-a2d6-4c9e-954c-a383ffdeea61 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Image limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 891.670568] env[61957]: DEBUG nova.virt.hardware [None req-ca59f931-a2d6-4c9e-954c-a383ffdeea61 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Flavor pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 891.670722] env[61957]: DEBUG nova.virt.hardware [None req-ca59f931-a2d6-4c9e-954c-a383ffdeea61 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Image pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 891.670873] env[61957]: DEBUG nova.virt.hardware [None req-ca59f931-a2d6-4c9e-954c-a383ffdeea61 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 891.671096] env[61957]: DEBUG nova.virt.hardware [None req-ca59f931-a2d6-4c9e-954c-a383ffdeea61 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 891.671262] env[61957]: DEBUG nova.virt.hardware [None req-ca59f931-a2d6-4c9e-954c-a383ffdeea61 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 891.671478] env[61957]: DEBUG nova.virt.hardware [None req-ca59f931-a2d6-4c9e-954c-a383ffdeea61 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Got 1 possible topologies {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 891.671629] env[61957]: DEBUG nova.virt.hardware [None req-ca59f931-a2d6-4c9e-954c-a383ffdeea61 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 891.671818] env[61957]: DEBUG nova.virt.hardware [None req-ca59f931-a2d6-4c9e-954c-a383ffdeea61 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 891.672792] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-bafb915b-a249-4473-bf3a-4c41863ba98f tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore2] 28ab7d23-487f-4ae1-8fe4-58db55b59918/28ab7d23-487f-4ae1-8fe4-58db55b59918.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 891.673099] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-bafb915b-a249-4473-bf3a-4c41863ba98f tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] [instance: 28ab7d23-487f-4ae1-8fe4-58db55b59918] Extending root virtual disk to 1048576 {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 891.673895] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ed4c01b-14b7-473c-8e12-3d2e4aa9021a {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.680887] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b589595b-14a0-4328-8776-ffd30c6ef07e {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.690413] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a65e0b33-0fd5-4d2d-aa81-c37d7d4e4847 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.712398] env[61957]: DEBUG oslo_vmware.api [None req-bafb915b-a249-4473-bf3a-4c41863ba98f tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Waiting for the task: (returnval){ [ 891.712398] env[61957]: value = "task-1277626" [ 891.712398] env[61957]: _type = "Task" [ 891.712398] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.724260] env[61957]: DEBUG oslo_vmware.api [None req-bafb915b-a249-4473-bf3a-4c41863ba98f tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Task: {'id': task-1277626, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.995422] env[61957]: DEBUG oslo_vmware.api [None req-9314189b-c158-4aed-834e-ffe55cc4ba51 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': task-1277625, 'name': PowerOffVM_Task, 'duration_secs': 0.243593} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.995422] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-9314189b-c158-4aed-834e-ffe55cc4ba51 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 4951132e-7247-4772-8f88-3664c6a7e61e] Powered off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 891.995557] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-9314189b-c158-4aed-834e-ffe55cc4ba51 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 4951132e-7247-4772-8f88-3664c6a7e61e] Unregistering the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 891.995863] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e088291e-33cf-439c-b32b-e3535566edfc {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.084643] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-9314189b-c158-4aed-834e-ffe55cc4ba51 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 4951132e-7247-4772-8f88-3664c6a7e61e] Unregistered the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 892.084978] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-9314189b-c158-4aed-834e-ffe55cc4ba51 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 4951132e-7247-4772-8f88-3664c6a7e61e] Deleting contents of the VM from datastore datastore1 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 892.085379] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-9314189b-c158-4aed-834e-ffe55cc4ba51 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Deleting the datastore file [datastore1] 4951132e-7247-4772-8f88-3664c6a7e61e {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 892.085379] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7ba3db2b-ccfd-420a-9b27-3ff3b111a307 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.093971] env[61957]: DEBUG oslo_vmware.api [None req-9314189b-c158-4aed-834e-ffe55cc4ba51 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Waiting for the task: (returnval){ [ 892.093971] env[61957]: value = "task-1277628" [ 892.093971] env[61957]: _type = "Task" [ 892.093971] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.106742] env[61957]: DEBUG oslo_vmware.api [None req-9314189b-c158-4aed-834e-ffe55cc4ba51 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': task-1277628, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.154455] env[61957]: DEBUG oslo_concurrency.lockutils [None req-7916de58-8954-4a2d-b3f9-5ec9b9a16797 tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.572s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 892.155321] env[61957]: DEBUG nova.compute.manager [None req-7916de58-8954-4a2d-b3f9-5ec9b9a16797 tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] [instance: 0d0f2d34-de35-4e80-8d9f-12693add0786] Start building networks asynchronously for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 892.161032] env[61957]: DEBUG oslo_concurrency.lockutils [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 25.387s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 892.161139] env[61957]: DEBUG oslo_concurrency.lockutils [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 892.163876] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61957) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 892.163876] env[61957]: DEBUG oslo_concurrency.lockutils [None req-accdd3a1-4d57-48f0-a206-1602b908905b tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.343s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 892.167893] env[61957]: INFO nova.compute.claims [None req-accdd3a1-4d57-48f0-a206-1602b908905b tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] [instance: a8ca1ad6-d636-4fa2-b0ac-53b020e392ce] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 892.171981] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86004f99-352f-44ed-b8ae-5c4f414bf1f3 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.182029] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab729b6a-7e0b-4b78-a99f-e2d9f4d4326f {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.200791] env[61957]: DEBUG nova.network.neutron [None req-4f69433c-53e0-4b6d-8d33-8d933f713309 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 4951132e-7247-4772-8f88-3664c6a7e61e] Updating instance_info_cache with network_info: [{"id": "3016c8eb-2ad3-40b1-8d88-eaec447de9ea", "address": "fa:16:3e:59:19:f2", "network": {"id": "364cabcb-046d-4f91-b1ee-5dd3adfae6d7", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1432396967-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.193", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8713aa35bcb24b86ad0b58ca9fc991ba", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b7a73c01-1bb9-4612-a1a7-16d71b732e81", "external-id": "nsx-vlan-transportzone-711", "segmentation_id": 711, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3016c8eb-2a", "ovs_interfaceid": "3016c8eb-2ad3-40b1-8d88-eaec447de9ea", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 892.205859] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39b1739b-13d4-4eff-9253-a7efe0289488 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.214928] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bb687d2-a300-4ef7-b2d5-cdf973074830 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.230823] env[61957]: DEBUG oslo_vmware.api [None req-bafb915b-a249-4473-bf3a-4c41863ba98f tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Task: {'id': task-1277626, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.421268} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.261995] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-bafb915b-a249-4473-bf3a-4c41863ba98f tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] [instance: 28ab7d23-487f-4ae1-8fe4-58db55b59918] Extended root virtual disk {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 892.262827] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181083MB free_disk=141GB free_vcpus=48 pci_devices=None {{(pid=61957) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 892.262827] env[61957]: DEBUG oslo_concurrency.lockutils [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 892.264052] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37bb9233-3bbf-4871-b841-da5ee9fa1efc {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.295408] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-bafb915b-a249-4473-bf3a-4c41863ba98f tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] [instance: 28ab7d23-487f-4ae1-8fe4-58db55b59918] Reconfiguring VM instance instance-00000046 to attach disk [datastore2] 28ab7d23-487f-4ae1-8fe4-58db55b59918/28ab7d23-487f-4ae1-8fe4-58db55b59918.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 892.295589] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-efda5250-af20-46ab-8931-2d5178fd1606 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.312083] env[61957]: DEBUG nova.network.neutron [None req-ca59f931-a2d6-4c9e-954c-a383ffdeea61 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] [instance: d5d10a31-0e78-4ed7-b944-9208138a4861] Successfully updated port: 572549f2-6ec4-411e-8905-0c33119b31c8 {{(pid=61957) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 892.320317] env[61957]: DEBUG oslo_vmware.api [None req-bafb915b-a249-4473-bf3a-4c41863ba98f tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Waiting for the task: (returnval){ [ 892.320317] env[61957]: value = "task-1277629" [ 892.320317] env[61957]: _type = "Task" [ 892.320317] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.334946] env[61957]: DEBUG oslo_vmware.api [None req-bafb915b-a249-4473-bf3a-4c41863ba98f tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Task: {'id': task-1277629, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.374964] env[61957]: DEBUG nova.compute.manager [req-bc038d76-ac92-4993-8c4c-5a60d5c58d5c req-9e95dee1-3298-4402-a7e7-c5c84b76a63c service nova] [instance: d5d10a31-0e78-4ed7-b944-9208138a4861] Received event network-vif-plugged-572549f2-6ec4-411e-8905-0c33119b31c8 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 892.374964] env[61957]: DEBUG oslo_concurrency.lockutils [req-bc038d76-ac92-4993-8c4c-5a60d5c58d5c req-9e95dee1-3298-4402-a7e7-c5c84b76a63c service nova] Acquiring lock "d5d10a31-0e78-4ed7-b944-9208138a4861-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 892.374964] env[61957]: DEBUG oslo_concurrency.lockutils [req-bc038d76-ac92-4993-8c4c-5a60d5c58d5c req-9e95dee1-3298-4402-a7e7-c5c84b76a63c service nova] Lock "d5d10a31-0e78-4ed7-b944-9208138a4861-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 892.374964] env[61957]: DEBUG oslo_concurrency.lockutils [req-bc038d76-ac92-4993-8c4c-5a60d5c58d5c req-9e95dee1-3298-4402-a7e7-c5c84b76a63c service nova] Lock "d5d10a31-0e78-4ed7-b944-9208138a4861-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 892.375417] env[61957]: DEBUG nova.compute.manager [req-bc038d76-ac92-4993-8c4c-5a60d5c58d5c req-9e95dee1-3298-4402-a7e7-c5c84b76a63c service nova] [instance: d5d10a31-0e78-4ed7-b944-9208138a4861] No waiting events found dispatching network-vif-plugged-572549f2-6ec4-411e-8905-0c33119b31c8 {{(pid=61957) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 892.375417] env[61957]: WARNING nova.compute.manager [req-bc038d76-ac92-4993-8c4c-5a60d5c58d5c req-9e95dee1-3298-4402-a7e7-c5c84b76a63c service nova] [instance: d5d10a31-0e78-4ed7-b944-9208138a4861] Received unexpected event network-vif-plugged-572549f2-6ec4-411e-8905-0c33119b31c8 for instance with vm_state building and task_state spawning. [ 892.375525] env[61957]: DEBUG nova.compute.manager [req-bc038d76-ac92-4993-8c4c-5a60d5c58d5c req-9e95dee1-3298-4402-a7e7-c5c84b76a63c service nova] [instance: d5d10a31-0e78-4ed7-b944-9208138a4861] Received event network-changed-572549f2-6ec4-411e-8905-0c33119b31c8 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 892.375684] env[61957]: DEBUG nova.compute.manager [req-bc038d76-ac92-4993-8c4c-5a60d5c58d5c req-9e95dee1-3298-4402-a7e7-c5c84b76a63c service nova] [instance: d5d10a31-0e78-4ed7-b944-9208138a4861] Refreshing instance network info cache due to event network-changed-572549f2-6ec4-411e-8905-0c33119b31c8. {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 892.375887] env[61957]: DEBUG oslo_concurrency.lockutils [req-bc038d76-ac92-4993-8c4c-5a60d5c58d5c req-9e95dee1-3298-4402-a7e7-c5c84b76a63c service nova] Acquiring lock "refresh_cache-d5d10a31-0e78-4ed7-b944-9208138a4861" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 892.376096] env[61957]: DEBUG oslo_concurrency.lockutils [req-bc038d76-ac92-4993-8c4c-5a60d5c58d5c req-9e95dee1-3298-4402-a7e7-c5c84b76a63c service nova] Acquired lock "refresh_cache-d5d10a31-0e78-4ed7-b944-9208138a4861" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 892.376632] env[61957]: DEBUG nova.network.neutron [req-bc038d76-ac92-4993-8c4c-5a60d5c58d5c req-9e95dee1-3298-4402-a7e7-c5c84b76a63c service nova] [instance: d5d10a31-0e78-4ed7-b944-9208138a4861] Refreshing network info cache for port 572549f2-6ec4-411e-8905-0c33119b31c8 {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 892.604565] env[61957]: DEBUG oslo_vmware.api [None req-9314189b-c158-4aed-834e-ffe55cc4ba51 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': task-1277628, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.332441} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.605288] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-9314189b-c158-4aed-834e-ffe55cc4ba51 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Deleted the datastore file {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 892.605523] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-9314189b-c158-4aed-834e-ffe55cc4ba51 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 4951132e-7247-4772-8f88-3664c6a7e61e] Deleted contents of the VM from datastore datastore1 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 892.605702] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-9314189b-c158-4aed-834e-ffe55cc4ba51 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 4951132e-7247-4772-8f88-3664c6a7e61e] Instance destroyed {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 892.605878] env[61957]: INFO nova.compute.manager [None req-9314189b-c158-4aed-834e-ffe55cc4ba51 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 4951132e-7247-4772-8f88-3664c6a7e61e] Took 1.15 seconds to destroy the instance on the hypervisor. [ 892.606182] env[61957]: DEBUG oslo.service.loopingcall [None req-9314189b-c158-4aed-834e-ffe55cc4ba51 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 892.606392] env[61957]: DEBUG nova.compute.manager [-] [instance: 4951132e-7247-4772-8f88-3664c6a7e61e] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 892.606482] env[61957]: DEBUG nova.network.neutron [-] [instance: 4951132e-7247-4772-8f88-3664c6a7e61e] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 892.673089] env[61957]: DEBUG nova.compute.utils [None req-7916de58-8954-4a2d-b3f9-5ec9b9a16797 tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Using /dev/sd instead of None {{(pid=61957) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 892.674855] env[61957]: DEBUG nova.compute.manager [None req-7916de58-8954-4a2d-b3f9-5ec9b9a16797 tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] [instance: 0d0f2d34-de35-4e80-8d9f-12693add0786] Not allocating networking since 'none' was specified. {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 892.707010] env[61957]: DEBUG oslo_concurrency.lockutils [None req-1036c4ee-30e7-4291-a697-7231cf748564 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Acquiring lock "9b0053fe-420e-4b92-86a8-ed44f6a6ec49" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 892.707010] env[61957]: DEBUG oslo_concurrency.lockutils [None req-1036c4ee-30e7-4291-a697-7231cf748564 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Lock "9b0053fe-420e-4b92-86a8-ed44f6a6ec49" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 892.708426] env[61957]: DEBUG oslo_concurrency.lockutils [None req-4f69433c-53e0-4b6d-8d33-8d933f713309 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Releasing lock "refresh_cache-4951132e-7247-4772-8f88-3664c6a7e61e" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 892.818046] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ca59f931-a2d6-4c9e-954c-a383ffdeea61 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Acquiring lock "refresh_cache-d5d10a31-0e78-4ed7-b944-9208138a4861" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 892.834278] env[61957]: DEBUG oslo_vmware.api [None req-bafb915b-a249-4473-bf3a-4c41863ba98f tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Task: {'id': task-1277629, 'name': ReconfigVM_Task, 'duration_secs': 0.334752} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.834549] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-bafb915b-a249-4473-bf3a-4c41863ba98f tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] [instance: 28ab7d23-487f-4ae1-8fe4-58db55b59918] Reconfigured VM instance instance-00000046 to attach disk [datastore2] 28ab7d23-487f-4ae1-8fe4-58db55b59918/28ab7d23-487f-4ae1-8fe4-58db55b59918.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 892.835210] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-80dbe218-a2d8-4ab6-a5e4-e664756fade7 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.842887] env[61957]: DEBUG oslo_vmware.api [None req-bafb915b-a249-4473-bf3a-4c41863ba98f tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Waiting for the task: (returnval){ [ 892.842887] env[61957]: value = "task-1277630" [ 892.842887] env[61957]: _type = "Task" [ 892.842887] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.853954] env[61957]: DEBUG oslo_vmware.api [None req-bafb915b-a249-4473-bf3a-4c41863ba98f tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Task: {'id': task-1277630, 'name': Rename_Task} progress is 5%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.922728] env[61957]: DEBUG nova.network.neutron [req-bc038d76-ac92-4993-8c4c-5a60d5c58d5c req-9e95dee1-3298-4402-a7e7-c5c84b76a63c service nova] [instance: d5d10a31-0e78-4ed7-b944-9208138a4861] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 893.065780] env[61957]: DEBUG nova.network.neutron [req-bc038d76-ac92-4993-8c4c-5a60d5c58d5c req-9e95dee1-3298-4402-a7e7-c5c84b76a63c service nova] [instance: d5d10a31-0e78-4ed7-b944-9208138a4861] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 893.177044] env[61957]: DEBUG nova.compute.manager [None req-7916de58-8954-4a2d-b3f9-5ec9b9a16797 tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] [instance: 0d0f2d34-de35-4e80-8d9f-12693add0786] Start building block device mappings for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 893.212392] env[61957]: DEBUG oslo_concurrency.lockutils [None req-4f69433c-53e0-4b6d-8d33-8d933f713309 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Lock "interface-4951132e-7247-4772-8f88-3664c6a7e61e-a093d1c9-1f94-4000-8f87-f76e9db98ec8" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 9.895s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 893.363292] env[61957]: DEBUG oslo_vmware.api [None req-bafb915b-a249-4473-bf3a-4c41863ba98f tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Task: {'id': task-1277630, 'name': Rename_Task, 'duration_secs': 0.149849} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.363292] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-bafb915b-a249-4473-bf3a-4c41863ba98f tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] [instance: 28ab7d23-487f-4ae1-8fe4-58db55b59918] Powering on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 893.363292] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-deb2a29a-32ed-455e-89b6-6dd7ed71e172 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.370947] env[61957]: DEBUG oslo_vmware.api [None req-bafb915b-a249-4473-bf3a-4c41863ba98f tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Waiting for the task: (returnval){ [ 893.370947] env[61957]: value = "task-1277631" [ 893.370947] env[61957]: _type = "Task" [ 893.370947] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.384134] env[61957]: DEBUG oslo_vmware.api [None req-bafb915b-a249-4473-bf3a-4c41863ba98f tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Task: {'id': task-1277631, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.553335] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0af6843-7fe4-4a8f-97b6-50a358c60030 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.561144] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bbb64da-f53a-42d5-91d5-bc1c6d836635 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.567657] env[61957]: DEBUG oslo_concurrency.lockutils [req-bc038d76-ac92-4993-8c4c-5a60d5c58d5c req-9e95dee1-3298-4402-a7e7-c5c84b76a63c service nova] Releasing lock "refresh_cache-d5d10a31-0e78-4ed7-b944-9208138a4861" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 893.593326] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ca59f931-a2d6-4c9e-954c-a383ffdeea61 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Acquired lock "refresh_cache-d5d10a31-0e78-4ed7-b944-9208138a4861" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 893.593461] env[61957]: DEBUG nova.network.neutron [None req-ca59f931-a2d6-4c9e-954c-a383ffdeea61 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] [instance: d5d10a31-0e78-4ed7-b944-9208138a4861] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 893.595393] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef1a0534-55e8-4546-9aec-9195843e2d60 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.604556] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86cdcbf5-e529-494c-b115-0fd13b25e2ec {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.620163] env[61957]: DEBUG nova.compute.provider_tree [None req-accdd3a1-4d57-48f0-a206-1602b908905b tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 893.637153] env[61957]: DEBUG nova.network.neutron [-] [instance: 4951132e-7247-4772-8f88-3664c6a7e61e] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 893.881423] env[61957]: DEBUG oslo_vmware.api [None req-bafb915b-a249-4473-bf3a-4c41863ba98f tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Task: {'id': task-1277631, 'name': PowerOnVM_Task, 'duration_secs': 0.46994} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.881646] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-bafb915b-a249-4473-bf3a-4c41863ba98f tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] [instance: 28ab7d23-487f-4ae1-8fe4-58db55b59918] Powered on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 893.881865] env[61957]: INFO nova.compute.manager [None req-bafb915b-a249-4473-bf3a-4c41863ba98f tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] [instance: 28ab7d23-487f-4ae1-8fe4-58db55b59918] Took 8.70 seconds to spawn the instance on the hypervisor. [ 893.882058] env[61957]: DEBUG nova.compute.manager [None req-bafb915b-a249-4473-bf3a-4c41863ba98f tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] [instance: 28ab7d23-487f-4ae1-8fe4-58db55b59918] Checking state {{(pid=61957) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 893.882818] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f931d04e-9a84-46b6-99c1-bb72f07a6027 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.127572] env[61957]: DEBUG nova.scheduler.client.report [None req-accdd3a1-4d57-48f0-a206-1602b908905b tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 894.131584] env[61957]: DEBUG nova.network.neutron [None req-ca59f931-a2d6-4c9e-954c-a383ffdeea61 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] [instance: d5d10a31-0e78-4ed7-b944-9208138a4861] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 894.140482] env[61957]: INFO nova.compute.manager [-] [instance: 4951132e-7247-4772-8f88-3664c6a7e61e] Took 1.53 seconds to deallocate network for instance. [ 894.193487] env[61957]: DEBUG nova.compute.manager [None req-7916de58-8954-4a2d-b3f9-5ec9b9a16797 tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] [instance: 0d0f2d34-de35-4e80-8d9f-12693add0786] Start spawning the instance on the hypervisor. {{(pid=61957) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 894.220261] env[61957]: DEBUG nova.virt.hardware [None req-7916de58-8954-4a2d-b3f9-5ec9b9a16797 tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T17:22:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T17:21:50Z,direct_url=,disk_format='vmdk',id=11c76a2c-f705-470a-ba9d-4657858bab38,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='81d3096d0f094373913fc3dc8f5c3c6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T17:21:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 894.220549] env[61957]: DEBUG nova.virt.hardware [None req-7916de58-8954-4a2d-b3f9-5ec9b9a16797 tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Flavor limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 894.220721] env[61957]: DEBUG nova.virt.hardware [None req-7916de58-8954-4a2d-b3f9-5ec9b9a16797 tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Image limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 894.220914] env[61957]: DEBUG nova.virt.hardware [None req-7916de58-8954-4a2d-b3f9-5ec9b9a16797 tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Flavor pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 894.221074] env[61957]: DEBUG nova.virt.hardware [None req-7916de58-8954-4a2d-b3f9-5ec9b9a16797 tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Image pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 894.221229] env[61957]: DEBUG nova.virt.hardware [None req-7916de58-8954-4a2d-b3f9-5ec9b9a16797 tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 894.221471] env[61957]: DEBUG nova.virt.hardware [None req-7916de58-8954-4a2d-b3f9-5ec9b9a16797 tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 894.221646] env[61957]: DEBUG nova.virt.hardware [None req-7916de58-8954-4a2d-b3f9-5ec9b9a16797 tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 894.221819] env[61957]: DEBUG nova.virt.hardware [None req-7916de58-8954-4a2d-b3f9-5ec9b9a16797 tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Got 1 possible topologies {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 894.221984] env[61957]: DEBUG nova.virt.hardware [None req-7916de58-8954-4a2d-b3f9-5ec9b9a16797 tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 894.222255] env[61957]: DEBUG nova.virt.hardware [None req-7916de58-8954-4a2d-b3f9-5ec9b9a16797 tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 894.223044] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-462dc5ba-e3e7-4c3d-8cfa-b9e1b3075367 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.233891] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4141c9c3-b95c-4022-a4e8-5292d915ca87 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.249183] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-7916de58-8954-4a2d-b3f9-5ec9b9a16797 tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] [instance: 0d0f2d34-de35-4e80-8d9f-12693add0786] Instance VIF info [] {{(pid=61957) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 894.254936] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-7916de58-8954-4a2d-b3f9-5ec9b9a16797 tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Creating folder: Project (c65b7d0c9c334c27b3c9aa12b45ca377). Parent ref: group-v274445. {{(pid=61957) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 894.257694] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4be09603-4b96-4e4e-8f50-c008d83930c0 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.272057] env[61957]: INFO nova.virt.vmwareapi.vm_util [None req-7916de58-8954-4a2d-b3f9-5ec9b9a16797 tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Created folder: Project (c65b7d0c9c334c27b3c9aa12b45ca377) in parent group-v274445. [ 894.272588] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-7916de58-8954-4a2d-b3f9-5ec9b9a16797 tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Creating folder: Instances. Parent ref: group-v274536. {{(pid=61957) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 894.272653] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-89b7621a-4379-477b-93ab-aed4eb8ac513 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.286039] env[61957]: INFO nova.virt.vmwareapi.vm_util [None req-7916de58-8954-4a2d-b3f9-5ec9b9a16797 tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Created folder: Instances in parent group-v274536. [ 894.286382] env[61957]: DEBUG oslo.service.loopingcall [None req-7916de58-8954-4a2d-b3f9-5ec9b9a16797 tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 894.286801] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0d0f2d34-de35-4e80-8d9f-12693add0786] Creating VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 894.287108] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d214b8c9-f6c2-41a9-823c-6a5e91468a8e {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.302725] env[61957]: DEBUG nova.network.neutron [None req-ca59f931-a2d6-4c9e-954c-a383ffdeea61 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] [instance: d5d10a31-0e78-4ed7-b944-9208138a4861] Updating instance_info_cache with network_info: [{"id": "572549f2-6ec4-411e-8905-0c33119b31c8", "address": "fa:16:3e:c2:97:2f", "network": {"id": "41a1701c-d9b6-4414-9e92-2e999cc916e3", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-727135203-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d849d922cfad41848a33e8d3bc83ea0f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f1b507ed-cd2d-4c09-9d96-c47bde6a7774", "external-id": "nsx-vlan-transportzone-980", "segmentation_id": 980, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap572549f2-6e", "ovs_interfaceid": "572549f2-6ec4-411e-8905-0c33119b31c8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 894.310393] env[61957]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 894.310393] env[61957]: value = "task-1277634" [ 894.310393] env[61957]: _type = "Task" [ 894.310393] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.318325] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277634, 'name': CreateVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.398600] env[61957]: INFO nova.compute.manager [None req-bafb915b-a249-4473-bf3a-4c41863ba98f tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] [instance: 28ab7d23-487f-4ae1-8fe4-58db55b59918] Took 34.13 seconds to build instance. [ 894.401137] env[61957]: DEBUG nova.compute.manager [req-65ac2b44-7937-45de-a7b9-505d3bd85785 req-e1bda8e8-ce30-4967-b1fe-3ff39fda8f88 service nova] [instance: 4951132e-7247-4772-8f88-3664c6a7e61e] Received event network-vif-deleted-3016c8eb-2ad3-40b1-8d88-eaec447de9ea {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 894.634880] env[61957]: DEBUG oslo_concurrency.lockutils [None req-accdd3a1-4d57-48f0-a206-1602b908905b tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.473s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 894.635401] env[61957]: DEBUG nova.compute.manager [None req-accdd3a1-4d57-48f0-a206-1602b908905b tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] [instance: a8ca1ad6-d636-4fa2-b0ac-53b020e392ce] Start building networks asynchronously for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 894.638073] env[61957]: DEBUG oslo_concurrency.lockutils [None req-61ad082c-18ce-4de4-a90d-5e84d15388ff tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 25.983s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 894.638351] env[61957]: DEBUG nova.objects.instance [None req-61ad082c-18ce-4de4-a90d-5e84d15388ff tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Lazy-loading 'resources' on Instance uuid 6438fe40-046c-45d5-9986-8f182ecde49f {{(pid=61957) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 894.646458] env[61957]: DEBUG oslo_concurrency.lockutils [None req-9314189b-c158-4aed-834e-ffe55cc4ba51 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 894.805454] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ca59f931-a2d6-4c9e-954c-a383ffdeea61 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Releasing lock "refresh_cache-d5d10a31-0e78-4ed7-b944-9208138a4861" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 894.805958] env[61957]: DEBUG nova.compute.manager [None req-ca59f931-a2d6-4c9e-954c-a383ffdeea61 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] [instance: d5d10a31-0e78-4ed7-b944-9208138a4861] Instance network_info: |[{"id": "572549f2-6ec4-411e-8905-0c33119b31c8", "address": "fa:16:3e:c2:97:2f", "network": {"id": "41a1701c-d9b6-4414-9e92-2e999cc916e3", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-727135203-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d849d922cfad41848a33e8d3bc83ea0f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f1b507ed-cd2d-4c09-9d96-c47bde6a7774", "external-id": "nsx-vlan-transportzone-980", "segmentation_id": 980, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap572549f2-6e", "ovs_interfaceid": "572549f2-6ec4-411e-8905-0c33119b31c8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 894.806442] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-ca59f931-a2d6-4c9e-954c-a383ffdeea61 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] [instance: d5d10a31-0e78-4ed7-b944-9208138a4861] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c2:97:2f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f1b507ed-cd2d-4c09-9d96-c47bde6a7774', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '572549f2-6ec4-411e-8905-0c33119b31c8', 'vif_model': 'vmxnet3'}] {{(pid=61957) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 894.814854] env[61957]: DEBUG oslo.service.loopingcall [None req-ca59f931-a2d6-4c9e-954c-a383ffdeea61 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 894.815189] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d5d10a31-0e78-4ed7-b944-9208138a4861] Creating VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 894.819535] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1af0169a-e30e-4a11-9c06-ee5414cfadf0 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.841538] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277634, 'name': CreateVM_Task, 'duration_secs': 0.329044} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.842780] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0d0f2d34-de35-4e80-8d9f-12693add0786] Created VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 894.842999] env[61957]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 894.842999] env[61957]: value = "task-1277635" [ 894.842999] env[61957]: _type = "Task" [ 894.842999] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.843430] env[61957]: DEBUG oslo_concurrency.lockutils [None req-7916de58-8954-4a2d-b3f9-5ec9b9a16797 tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 894.843596] env[61957]: DEBUG oslo_concurrency.lockutils [None req-7916de58-8954-4a2d-b3f9-5ec9b9a16797 tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 894.843926] env[61957]: DEBUG oslo_concurrency.lockutils [None req-7916de58-8954-4a2d-b3f9-5ec9b9a16797 tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 894.844240] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c1c61a25-3dda-45d7-aa1e-04e0260082b0 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.856848] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277635, 'name': CreateVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.857203] env[61957]: DEBUG oslo_vmware.api [None req-7916de58-8954-4a2d-b3f9-5ec9b9a16797 tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Waiting for the task: (returnval){ [ 894.857203] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]5267de20-84ce-49d6-1ddb-7b639d8b2d66" [ 894.857203] env[61957]: _type = "Task" [ 894.857203] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.867132] env[61957]: DEBUG oslo_vmware.api [None req-7916de58-8954-4a2d-b3f9-5ec9b9a16797 tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]5267de20-84ce-49d6-1ddb-7b639d8b2d66, 'name': SearchDatastore_Task, 'duration_secs': 0.009648} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.867536] env[61957]: DEBUG oslo_concurrency.lockutils [None req-7916de58-8954-4a2d-b3f9-5ec9b9a16797 tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 894.867809] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-7916de58-8954-4a2d-b3f9-5ec9b9a16797 tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] [instance: 0d0f2d34-de35-4e80-8d9f-12693add0786] Processing image 11c76a2c-f705-470a-ba9d-4657858bab38 {{(pid=61957) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 894.868080] env[61957]: DEBUG oslo_concurrency.lockutils [None req-7916de58-8954-4a2d-b3f9-5ec9b9a16797 tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 894.868238] env[61957]: DEBUG oslo_concurrency.lockutils [None req-7916de58-8954-4a2d-b3f9-5ec9b9a16797 tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 894.868486] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-7916de58-8954-4a2d-b3f9-5ec9b9a16797 tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 894.868781] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6348e2bd-7ccb-4881-8499-57ea896fde80 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.876350] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-7916de58-8954-4a2d-b3f9-5ec9b9a16797 tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 894.876528] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-7916de58-8954-4a2d-b3f9-5ec9b9a16797 tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61957) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 894.877254] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4079a103-6a1f-4bfa-a599-fba2048c05f3 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.882657] env[61957]: DEBUG oslo_vmware.api [None req-7916de58-8954-4a2d-b3f9-5ec9b9a16797 tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Waiting for the task: (returnval){ [ 894.882657] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52afebe7-ff07-c280-fc04-fc08fcc42f9c" [ 894.882657] env[61957]: _type = "Task" [ 894.882657] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.891032] env[61957]: DEBUG oslo_vmware.api [None req-7916de58-8954-4a2d-b3f9-5ec9b9a16797 tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52afebe7-ff07-c280-fc04-fc08fcc42f9c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.902776] env[61957]: DEBUG oslo_concurrency.lockutils [None req-bafb915b-a249-4473-bf3a-4c41863ba98f tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Lock "28ab7d23-487f-4ae1-8fe4-58db55b59918" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 76.007s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 895.145950] env[61957]: DEBUG nova.compute.utils [None req-accdd3a1-4d57-48f0-a206-1602b908905b tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Using /dev/sd instead of None {{(pid=61957) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 895.147605] env[61957]: DEBUG nova.compute.manager [None req-accdd3a1-4d57-48f0-a206-1602b908905b tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] [instance: a8ca1ad6-d636-4fa2-b0ac-53b020e392ce] Not allocating networking since 'none' was specified. {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 895.356262] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277635, 'name': CreateVM_Task, 'duration_secs': 0.358892} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.358446] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d5d10a31-0e78-4ed7-b944-9208138a4861] Created VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 895.359526] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ca59f931-a2d6-4c9e-954c-a383ffdeea61 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 895.359526] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ca59f931-a2d6-4c9e-954c-a383ffdeea61 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 895.359792] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ca59f931-a2d6-4c9e-954c-a383ffdeea61 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 895.360057] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-370353b4-02e8-4696-bd60-f09b2095963f {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.365345] env[61957]: DEBUG oslo_vmware.api [None req-ca59f931-a2d6-4c9e-954c-a383ffdeea61 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Waiting for the task: (returnval){ [ 895.365345] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]5287f6e4-3883-777d-1c93-a3a9d3cfa0b1" [ 895.365345] env[61957]: _type = "Task" [ 895.365345] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.375011] env[61957]: DEBUG oslo_vmware.api [None req-ca59f931-a2d6-4c9e-954c-a383ffdeea61 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]5287f6e4-3883-777d-1c93-a3a9d3cfa0b1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.391256] env[61957]: DEBUG oslo_vmware.api [None req-7916de58-8954-4a2d-b3f9-5ec9b9a16797 tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52afebe7-ff07-c280-fc04-fc08fcc42f9c, 'name': SearchDatastore_Task, 'duration_secs': 0.009562} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.394073] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-43930417-1a48-45b2-889d-a3eda6581da7 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.399210] env[61957]: DEBUG oslo_vmware.api [None req-7916de58-8954-4a2d-b3f9-5ec9b9a16797 tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Waiting for the task: (returnval){ [ 895.399210] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52baa4d1-f60b-a181-fa6d-2010b8f40e0f" [ 895.399210] env[61957]: _type = "Task" [ 895.399210] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.405905] env[61957]: DEBUG nova.compute.manager [None req-d28d4db3-4ba8-4b6b-b020-dcb0a978b0ab tempest-InstanceActionsNegativeTestJSON-259524954 tempest-InstanceActionsNegativeTestJSON-259524954-project-member] [instance: 7e2db5f6-1419-415e-b957-66cf2e3ec19d] Starting instance... {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 895.411548] env[61957]: DEBUG oslo_vmware.api [None req-7916de58-8954-4a2d-b3f9-5ec9b9a16797 tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52baa4d1-f60b-a181-fa6d-2010b8f40e0f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.435172] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42ce26b7-1177-4912-a54c-1dea073ab776 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.443400] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82c35878-69c0-4fbd-ac90-d4dec0d3dcd2 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.473960] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77c32da4-1d76-43a3-9925-283643f41e10 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.482428] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e915e7e-7efc-4576-a0ae-3c47cc817655 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.497792] env[61957]: DEBUG nova.compute.provider_tree [None req-61ad082c-18ce-4de4-a90d-5e84d15388ff tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 895.649447] env[61957]: DEBUG nova.compute.manager [None req-accdd3a1-4d57-48f0-a206-1602b908905b tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] [instance: a8ca1ad6-d636-4fa2-b0ac-53b020e392ce] Start building block device mappings for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 895.875225] env[61957]: DEBUG oslo_vmware.api [None req-ca59f931-a2d6-4c9e-954c-a383ffdeea61 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]5287f6e4-3883-777d-1c93-a3a9d3cfa0b1, 'name': SearchDatastore_Task, 'duration_secs': 0.009657} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.875537] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ca59f931-a2d6-4c9e-954c-a383ffdeea61 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 895.875774] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-ca59f931-a2d6-4c9e-954c-a383ffdeea61 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] [instance: d5d10a31-0e78-4ed7-b944-9208138a4861] Processing image 11c76a2c-f705-470a-ba9d-4657858bab38 {{(pid=61957) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 895.875989] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ca59f931-a2d6-4c9e-954c-a383ffdeea61 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 895.908965] env[61957]: DEBUG oslo_vmware.api [None req-7916de58-8954-4a2d-b3f9-5ec9b9a16797 tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52baa4d1-f60b-a181-fa6d-2010b8f40e0f, 'name': SearchDatastore_Task, 'duration_secs': 0.012138} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.909187] env[61957]: DEBUG oslo_concurrency.lockutils [None req-7916de58-8954-4a2d-b3f9-5ec9b9a16797 tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 895.909438] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-7916de58-8954-4a2d-b3f9-5ec9b9a16797 tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore2] 0d0f2d34-de35-4e80-8d9f-12693add0786/0d0f2d34-de35-4e80-8d9f-12693add0786.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 895.909709] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ca59f931-a2d6-4c9e-954c-a383ffdeea61 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 895.909900] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-ca59f931-a2d6-4c9e-954c-a383ffdeea61 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 895.910142] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ee0dcc17-b116-4703-a91d-888c5b2522e4 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.912177] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-008049a3-1fab-46ff-9c15-ac752fe0432c {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.921303] env[61957]: DEBUG oslo_vmware.api [None req-7916de58-8954-4a2d-b3f9-5ec9b9a16797 tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Waiting for the task: (returnval){ [ 895.921303] env[61957]: value = "task-1277636" [ 895.921303] env[61957]: _type = "Task" [ 895.921303] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.925916] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-ca59f931-a2d6-4c9e-954c-a383ffdeea61 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 895.926133] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-ca59f931-a2d6-4c9e-954c-a383ffdeea61 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61957) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 895.927113] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-84ce7cfe-69bb-40b8-8d7f-323153ed9251 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.932686] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d28d4db3-4ba8-4b6b-b020-dcb0a978b0ab tempest-InstanceActionsNegativeTestJSON-259524954 tempest-InstanceActionsNegativeTestJSON-259524954-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 895.932959] env[61957]: DEBUG oslo_vmware.api [None req-7916de58-8954-4a2d-b3f9-5ec9b9a16797 tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Task: {'id': task-1277636, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.935917] env[61957]: DEBUG oslo_vmware.api [None req-ca59f931-a2d6-4c9e-954c-a383ffdeea61 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Waiting for the task: (returnval){ [ 895.935917] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52fd8c68-0ffe-5656-2e40-0f429b8c305d" [ 895.935917] env[61957]: _type = "Task" [ 895.935917] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.944150] env[61957]: DEBUG oslo_vmware.api [None req-ca59f931-a2d6-4c9e-954c-a383ffdeea61 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52fd8c68-0ffe-5656-2e40-0f429b8c305d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.000511] env[61957]: DEBUG nova.scheduler.client.report [None req-61ad082c-18ce-4de4-a90d-5e84d15388ff tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 896.431816] env[61957]: DEBUG oslo_vmware.api [None req-7916de58-8954-4a2d-b3f9-5ec9b9a16797 tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Task: {'id': task-1277636, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.441775} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.432150] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-7916de58-8954-4a2d-b3f9-5ec9b9a16797 tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore2] 0d0f2d34-de35-4e80-8d9f-12693add0786/0d0f2d34-de35-4e80-8d9f-12693add0786.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 896.432380] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-7916de58-8954-4a2d-b3f9-5ec9b9a16797 tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] [instance: 0d0f2d34-de35-4e80-8d9f-12693add0786] Extending root virtual disk to 1048576 {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 896.432688] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a458d140-95a6-4c7a-8a1d-7a84d350c6c3 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.441076] env[61957]: DEBUG oslo_vmware.api [None req-7916de58-8954-4a2d-b3f9-5ec9b9a16797 tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Waiting for the task: (returnval){ [ 896.441076] env[61957]: value = "task-1277637" [ 896.441076] env[61957]: _type = "Task" [ 896.441076] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.447813] env[61957]: DEBUG oslo_vmware.api [None req-ca59f931-a2d6-4c9e-954c-a383ffdeea61 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52fd8c68-0ffe-5656-2e40-0f429b8c305d, 'name': SearchDatastore_Task, 'duration_secs': 0.008439} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.448927] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c35f1634-9850-43ad-87af-7e9fe294893a {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.454546] env[61957]: DEBUG oslo_vmware.api [None req-7916de58-8954-4a2d-b3f9-5ec9b9a16797 tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Task: {'id': task-1277637, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.458631] env[61957]: DEBUG oslo_vmware.api [None req-ca59f931-a2d6-4c9e-954c-a383ffdeea61 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Waiting for the task: (returnval){ [ 896.458631] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52f01067-0743-f57e-f9a8-e31dba3626ea" [ 896.458631] env[61957]: _type = "Task" [ 896.458631] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.466912] env[61957]: DEBUG oslo_vmware.api [None req-ca59f931-a2d6-4c9e-954c-a383ffdeea61 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52f01067-0743-f57e-f9a8-e31dba3626ea, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.506131] env[61957]: DEBUG oslo_concurrency.lockutils [None req-61ad082c-18ce-4de4-a90d-5e84d15388ff tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.868s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 896.508679] env[61957]: DEBUG oslo_concurrency.lockutils [None req-99665f7b-964c-4087-97b8-943747701eca tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 27.838s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 896.508925] env[61957]: DEBUG nova.objects.instance [None req-99665f7b-964c-4087-97b8-943747701eca tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Lazy-loading 'resources' on Instance uuid f66db265-887e-4d61-b848-c609e5c884cb {{(pid=61957) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 896.523737] env[61957]: INFO nova.scheduler.client.report [None req-61ad082c-18ce-4de4-a90d-5e84d15388ff tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Deleted allocations for instance 6438fe40-046c-45d5-9986-8f182ecde49f [ 896.659521] env[61957]: DEBUG nova.compute.manager [None req-accdd3a1-4d57-48f0-a206-1602b908905b tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] [instance: a8ca1ad6-d636-4fa2-b0ac-53b020e392ce] Start spawning the instance on the hypervisor. {{(pid=61957) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 896.687116] env[61957]: DEBUG nova.virt.hardware [None req-accdd3a1-4d57-48f0-a206-1602b908905b tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T17:22:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T17:21:50Z,direct_url=,disk_format='vmdk',id=11c76a2c-f705-470a-ba9d-4657858bab38,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='81d3096d0f094373913fc3dc8f5c3c6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T17:21:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 896.687116] env[61957]: DEBUG nova.virt.hardware [None req-accdd3a1-4d57-48f0-a206-1602b908905b tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Flavor limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 896.687256] env[61957]: DEBUG nova.virt.hardware [None req-accdd3a1-4d57-48f0-a206-1602b908905b tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Image limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 896.687358] env[61957]: DEBUG nova.virt.hardware [None req-accdd3a1-4d57-48f0-a206-1602b908905b tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Flavor pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 896.687526] env[61957]: DEBUG nova.virt.hardware [None req-accdd3a1-4d57-48f0-a206-1602b908905b tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Image pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 896.687678] env[61957]: DEBUG nova.virt.hardware [None req-accdd3a1-4d57-48f0-a206-1602b908905b tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 896.687911] env[61957]: DEBUG nova.virt.hardware [None req-accdd3a1-4d57-48f0-a206-1602b908905b tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 896.688113] env[61957]: DEBUG nova.virt.hardware [None req-accdd3a1-4d57-48f0-a206-1602b908905b tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 896.688310] env[61957]: DEBUG nova.virt.hardware [None req-accdd3a1-4d57-48f0-a206-1602b908905b tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Got 1 possible topologies {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 896.688487] env[61957]: DEBUG nova.virt.hardware [None req-accdd3a1-4d57-48f0-a206-1602b908905b tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 896.688684] env[61957]: DEBUG nova.virt.hardware [None req-accdd3a1-4d57-48f0-a206-1602b908905b tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 896.689671] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2b37c97-3a25-4944-8abd-eb648b0e706b {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.706483] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a284391d-8a1d-446d-83b4-e9c256e00ea1 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.722639] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-accdd3a1-4d57-48f0-a206-1602b908905b tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] [instance: a8ca1ad6-d636-4fa2-b0ac-53b020e392ce] Instance VIF info [] {{(pid=61957) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 896.728321] env[61957]: DEBUG oslo.service.loopingcall [None req-accdd3a1-4d57-48f0-a206-1602b908905b tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 896.728626] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a8ca1ad6-d636-4fa2-b0ac-53b020e392ce] Creating VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 896.728858] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-adf07cb4-45f9-4f74-87f6-580505771f9b {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.747669] env[61957]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 896.747669] env[61957]: value = "task-1277638" [ 896.747669] env[61957]: _type = "Task" [ 896.747669] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.756616] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277638, 'name': CreateVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.951769] env[61957]: DEBUG oslo_vmware.api [None req-7916de58-8954-4a2d-b3f9-5ec9b9a16797 tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Task: {'id': task-1277637, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.07181} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.952100] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-7916de58-8954-4a2d-b3f9-5ec9b9a16797 tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] [instance: 0d0f2d34-de35-4e80-8d9f-12693add0786] Extended root virtual disk {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 896.952921] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a578200-b99c-4145-b50a-fbacd32caa82 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.974331] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-7916de58-8954-4a2d-b3f9-5ec9b9a16797 tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] [instance: 0d0f2d34-de35-4e80-8d9f-12693add0786] Reconfiguring VM instance instance-00000048 to attach disk [datastore2] 0d0f2d34-de35-4e80-8d9f-12693add0786/0d0f2d34-de35-4e80-8d9f-12693add0786.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 896.977576] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e31545c5-7da5-4a40-8d92-cc42bc2bad17 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.998262] env[61957]: DEBUG oslo_vmware.api [None req-ca59f931-a2d6-4c9e-954c-a383ffdeea61 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52f01067-0743-f57e-f9a8-e31dba3626ea, 'name': SearchDatastore_Task, 'duration_secs': 0.009219} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.999661] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ca59f931-a2d6-4c9e-954c-a383ffdeea61 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 896.999874] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca59f931-a2d6-4c9e-954c-a383ffdeea61 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore2] d5d10a31-0e78-4ed7-b944-9208138a4861/d5d10a31-0e78-4ed7-b944-9208138a4861.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 897.000222] env[61957]: DEBUG oslo_vmware.api [None req-7916de58-8954-4a2d-b3f9-5ec9b9a16797 tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Waiting for the task: (returnval){ [ 897.000222] env[61957]: value = "task-1277639" [ 897.000222] env[61957]: _type = "Task" [ 897.000222] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.000421] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f8adf7f2-bdba-4d11-b9ff-fbe2879bba49 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.014811] env[61957]: DEBUG oslo_vmware.api [None req-7916de58-8954-4a2d-b3f9-5ec9b9a16797 tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Task: {'id': task-1277639, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.016802] env[61957]: DEBUG oslo_vmware.api [None req-ca59f931-a2d6-4c9e-954c-a383ffdeea61 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Waiting for the task: (returnval){ [ 897.016802] env[61957]: value = "task-1277640" [ 897.016802] env[61957]: _type = "Task" [ 897.016802] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.026049] env[61957]: DEBUG oslo_vmware.api [None req-ca59f931-a2d6-4c9e-954c-a383ffdeea61 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Task: {'id': task-1277640, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.031262] env[61957]: DEBUG oslo_concurrency.lockutils [None req-61ad082c-18ce-4de4-a90d-5e84d15388ff tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Lock "6438fe40-046c-45d5-9986-8f182ecde49f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 32.085s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 897.260784] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277638, 'name': CreateVM_Task} progress is 99%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.337811] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cfe2a76-8e3d-4f41-b9ce-0801bb84e8ec {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.347043] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35812377-4e70-4135-9a22-43bc9e23cc89 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.379532] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1e9b834-2234-4516-a3ff-717ff58cd812 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.388166] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a5ba599-e775-4a35-adf9-3164561e02bc {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.403762] env[61957]: DEBUG nova.compute.provider_tree [None req-99665f7b-964c-4087-97b8-943747701eca tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 897.517628] env[61957]: DEBUG oslo_vmware.api [None req-7916de58-8954-4a2d-b3f9-5ec9b9a16797 tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Task: {'id': task-1277639, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.529154] env[61957]: DEBUG oslo_vmware.api [None req-ca59f931-a2d6-4c9e-954c-a383ffdeea61 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Task: {'id': task-1277640, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.758856] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277638, 'name': CreateVM_Task, 'duration_secs': 0.624419} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.759122] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a8ca1ad6-d636-4fa2-b0ac-53b020e392ce] Created VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 897.759448] env[61957]: DEBUG oslo_concurrency.lockutils [None req-accdd3a1-4d57-48f0-a206-1602b908905b tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 897.759613] env[61957]: DEBUG oslo_concurrency.lockutils [None req-accdd3a1-4d57-48f0-a206-1602b908905b tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Acquired lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 897.759927] env[61957]: DEBUG oslo_concurrency.lockutils [None req-accdd3a1-4d57-48f0-a206-1602b908905b tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 897.760190] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-22ddd261-1708-422a-af41-a5a30bdf50ad {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.764900] env[61957]: DEBUG oslo_vmware.api [None req-accdd3a1-4d57-48f0-a206-1602b908905b tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Waiting for the task: (returnval){ [ 897.764900] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]523e9731-897b-b9cc-e547-426aca3406c7" [ 897.764900] env[61957]: _type = "Task" [ 897.764900] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.772338] env[61957]: DEBUG oslo_vmware.api [None req-accdd3a1-4d57-48f0-a206-1602b908905b tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]523e9731-897b-b9cc-e547-426aca3406c7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.906943] env[61957]: DEBUG nova.scheduler.client.report [None req-99665f7b-964c-4087-97b8-943747701eca tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 898.013980] env[61957]: DEBUG oslo_vmware.api [None req-7916de58-8954-4a2d-b3f9-5ec9b9a16797 tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Task: {'id': task-1277639, 'name': ReconfigVM_Task, 'duration_secs': 0.605222} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.014154] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-7916de58-8954-4a2d-b3f9-5ec9b9a16797 tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] [instance: 0d0f2d34-de35-4e80-8d9f-12693add0786] Reconfigured VM instance instance-00000048 to attach disk [datastore2] 0d0f2d34-de35-4e80-8d9f-12693add0786/0d0f2d34-de35-4e80-8d9f-12693add0786.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 898.014782] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-af93dfef-5879-4ee6-ba69-d1cd37298adb {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.022984] env[61957]: DEBUG oslo_vmware.api [None req-7916de58-8954-4a2d-b3f9-5ec9b9a16797 tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Waiting for the task: (returnval){ [ 898.022984] env[61957]: value = "task-1277641" [ 898.022984] env[61957]: _type = "Task" [ 898.022984] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.026382] env[61957]: DEBUG oslo_vmware.api [None req-ca59f931-a2d6-4c9e-954c-a383ffdeea61 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Task: {'id': task-1277640, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.81} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.029272] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca59f931-a2d6-4c9e-954c-a383ffdeea61 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore2] d5d10a31-0e78-4ed7-b944-9208138a4861/d5d10a31-0e78-4ed7-b944-9208138a4861.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 898.029500] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-ca59f931-a2d6-4c9e-954c-a383ffdeea61 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] [instance: d5d10a31-0e78-4ed7-b944-9208138a4861] Extending root virtual disk to 1048576 {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 898.029897] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0ca9eb27-87eb-42bb-bd40-255914222b01 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.039042] env[61957]: DEBUG oslo_vmware.api [None req-7916de58-8954-4a2d-b3f9-5ec9b9a16797 tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Task: {'id': task-1277641, 'name': Rename_Task} progress is 6%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.040113] env[61957]: DEBUG oslo_vmware.api [None req-ca59f931-a2d6-4c9e-954c-a383ffdeea61 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Waiting for the task: (returnval){ [ 898.040113] env[61957]: value = "task-1277642" [ 898.040113] env[61957]: _type = "Task" [ 898.040113] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.049279] env[61957]: DEBUG oslo_vmware.api [None req-ca59f931-a2d6-4c9e-954c-a383ffdeea61 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Task: {'id': task-1277642, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.277493] env[61957]: DEBUG oslo_vmware.api [None req-accdd3a1-4d57-48f0-a206-1602b908905b tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]523e9731-897b-b9cc-e547-426aca3406c7, 'name': SearchDatastore_Task, 'duration_secs': 0.050994} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.277493] env[61957]: DEBUG oslo_concurrency.lockutils [None req-accdd3a1-4d57-48f0-a206-1602b908905b tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Releasing lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 898.278339] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-accdd3a1-4d57-48f0-a206-1602b908905b tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] [instance: a8ca1ad6-d636-4fa2-b0ac-53b020e392ce] Processing image 11c76a2c-f705-470a-ba9d-4657858bab38 {{(pid=61957) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 898.278339] env[61957]: DEBUG oslo_concurrency.lockutils [None req-accdd3a1-4d57-48f0-a206-1602b908905b tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 898.278339] env[61957]: DEBUG oslo_concurrency.lockutils [None req-accdd3a1-4d57-48f0-a206-1602b908905b tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Acquired lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 898.278339] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-accdd3a1-4d57-48f0-a206-1602b908905b tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 898.278625] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8d3c04f9-e8ec-4904-bc22-932915d1c3f4 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.288277] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-accdd3a1-4d57-48f0-a206-1602b908905b tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 898.288522] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-accdd3a1-4d57-48f0-a206-1602b908905b tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61957) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 898.289267] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-aa203ed5-f8dd-4221-ab5a-7b09211a5368 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.294770] env[61957]: DEBUG oslo_vmware.api [None req-accdd3a1-4d57-48f0-a206-1602b908905b tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Waiting for the task: (returnval){ [ 898.294770] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]527995b8-01b6-8043-aed6-a0ff24eec728" [ 898.294770] env[61957]: _type = "Task" [ 898.294770] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.303111] env[61957]: DEBUG oslo_vmware.api [None req-accdd3a1-4d57-48f0-a206-1602b908905b tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]527995b8-01b6-8043-aed6-a0ff24eec728, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.413066] env[61957]: DEBUG oslo_concurrency.lockutils [None req-99665f7b-964c-4087-97b8-943747701eca tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.904s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 898.414740] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ca2fb2e2-e99f-422a-be65-983bd81b6e15 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.412s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 898.416295] env[61957]: INFO nova.compute.claims [None req-ca2fb2e2-e99f-422a-be65-983bd81b6e15 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] [instance: 736eabcf-70f8-4e1e-9830-270d9971fc9a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 898.431223] env[61957]: INFO nova.scheduler.client.report [None req-99665f7b-964c-4087-97b8-943747701eca tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Deleted allocations for instance f66db265-887e-4d61-b848-c609e5c884cb [ 898.538190] env[61957]: DEBUG oslo_vmware.api [None req-7916de58-8954-4a2d-b3f9-5ec9b9a16797 tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Task: {'id': task-1277641, 'name': Rename_Task, 'duration_secs': 0.154916} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.538526] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-7916de58-8954-4a2d-b3f9-5ec9b9a16797 tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] [instance: 0d0f2d34-de35-4e80-8d9f-12693add0786] Powering on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 898.538764] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8a75f2a1-3a99-48dc-81fc-76cbca624ec2 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.546230] env[61957]: DEBUG oslo_vmware.api [None req-7916de58-8954-4a2d-b3f9-5ec9b9a16797 tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Waiting for the task: (returnval){ [ 898.546230] env[61957]: value = "task-1277643" [ 898.546230] env[61957]: _type = "Task" [ 898.546230] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.549094] env[61957]: DEBUG oslo_vmware.api [None req-ca59f931-a2d6-4c9e-954c-a383ffdeea61 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Task: {'id': task-1277642, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.256419} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.551870] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-ca59f931-a2d6-4c9e-954c-a383ffdeea61 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] [instance: d5d10a31-0e78-4ed7-b944-9208138a4861] Extended root virtual disk {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 898.552600] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-092eafd2-585a-4375-b3a1-443da486b9b1 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.575634] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-ca59f931-a2d6-4c9e-954c-a383ffdeea61 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] [instance: d5d10a31-0e78-4ed7-b944-9208138a4861] Reconfiguring VM instance instance-00000047 to attach disk [datastore2] d5d10a31-0e78-4ed7-b944-9208138a4861/d5d10a31-0e78-4ed7-b944-9208138a4861.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 898.578627] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-178fb88d-e042-4f25-8382-b1d19913f36d {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.594355] env[61957]: DEBUG oslo_vmware.api [None req-7916de58-8954-4a2d-b3f9-5ec9b9a16797 tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Task: {'id': task-1277643, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.601063] env[61957]: DEBUG oslo_vmware.api [None req-ca59f931-a2d6-4c9e-954c-a383ffdeea61 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Waiting for the task: (returnval){ [ 898.601063] env[61957]: value = "task-1277644" [ 898.601063] env[61957]: _type = "Task" [ 898.601063] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.610112] env[61957]: DEBUG oslo_vmware.api [None req-ca59f931-a2d6-4c9e-954c-a383ffdeea61 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Task: {'id': task-1277644, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.805760] env[61957]: DEBUG oslo_vmware.api [None req-accdd3a1-4d57-48f0-a206-1602b908905b tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]527995b8-01b6-8043-aed6-a0ff24eec728, 'name': SearchDatastore_Task, 'duration_secs': 0.076316} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.806631] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-40829e80-7d2a-46d6-a67d-f6ccd2d4f807 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.813749] env[61957]: DEBUG oslo_vmware.api [None req-accdd3a1-4d57-48f0-a206-1602b908905b tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Waiting for the task: (returnval){ [ 898.813749] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]5214a2ab-c42b-348d-ad63-ac212b4a5537" [ 898.813749] env[61957]: _type = "Task" [ 898.813749] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.821903] env[61957]: DEBUG oslo_vmware.api [None req-accdd3a1-4d57-48f0-a206-1602b908905b tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]5214a2ab-c42b-348d-ad63-ac212b4a5537, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.938641] env[61957]: DEBUG oslo_concurrency.lockutils [None req-99665f7b-964c-4087-97b8-943747701eca tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Lock "f66db265-887e-4d61-b848-c609e5c884cb" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 33.914s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 899.062213] env[61957]: DEBUG oslo_vmware.api [None req-7916de58-8954-4a2d-b3f9-5ec9b9a16797 tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Task: {'id': task-1277643, 'name': PowerOnVM_Task, 'duration_secs': 0.513657} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.062933] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-7916de58-8954-4a2d-b3f9-5ec9b9a16797 tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] [instance: 0d0f2d34-de35-4e80-8d9f-12693add0786] Powered on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 899.063208] env[61957]: INFO nova.compute.manager [None req-7916de58-8954-4a2d-b3f9-5ec9b9a16797 tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] [instance: 0d0f2d34-de35-4e80-8d9f-12693add0786] Took 4.87 seconds to spawn the instance on the hypervisor. [ 899.063429] env[61957]: DEBUG nova.compute.manager [None req-7916de58-8954-4a2d-b3f9-5ec9b9a16797 tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] [instance: 0d0f2d34-de35-4e80-8d9f-12693add0786] Checking state {{(pid=61957) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 899.064230] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21ac5d0e-74ff-4a71-be44-d20725cd2e74 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.112141] env[61957]: DEBUG oslo_vmware.api [None req-ca59f931-a2d6-4c9e-954c-a383ffdeea61 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Task: {'id': task-1277644, 'name': ReconfigVM_Task, 'duration_secs': 0.463117} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.112500] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-ca59f931-a2d6-4c9e-954c-a383ffdeea61 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] [instance: d5d10a31-0e78-4ed7-b944-9208138a4861] Reconfigured VM instance instance-00000047 to attach disk [datastore2] d5d10a31-0e78-4ed7-b944-9208138a4861/d5d10a31-0e78-4ed7-b944-9208138a4861.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 899.113188] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f2f6532f-286f-4b0b-8747-be313350c351 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.120281] env[61957]: DEBUG oslo_vmware.api [None req-ca59f931-a2d6-4c9e-954c-a383ffdeea61 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Waiting for the task: (returnval){ [ 899.120281] env[61957]: value = "task-1277645" [ 899.120281] env[61957]: _type = "Task" [ 899.120281] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.128154] env[61957]: DEBUG oslo_vmware.api [None req-ca59f931-a2d6-4c9e-954c-a383ffdeea61 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Task: {'id': task-1277645, 'name': Rename_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.325024] env[61957]: DEBUG oslo_vmware.api [None req-accdd3a1-4d57-48f0-a206-1602b908905b tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]5214a2ab-c42b-348d-ad63-ac212b4a5537, 'name': SearchDatastore_Task, 'duration_secs': 0.010241} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.325024] env[61957]: DEBUG oslo_concurrency.lockutils [None req-accdd3a1-4d57-48f0-a206-1602b908905b tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Releasing lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 899.325268] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-accdd3a1-4d57-48f0-a206-1602b908905b tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore1] a8ca1ad6-d636-4fa2-b0ac-53b020e392ce/a8ca1ad6-d636-4fa2-b0ac-53b020e392ce.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 899.325468] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0402001c-29af-48ae-abc6-29b802685f66 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.334720] env[61957]: DEBUG oslo_vmware.api [None req-accdd3a1-4d57-48f0-a206-1602b908905b tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Waiting for the task: (returnval){ [ 899.334720] env[61957]: value = "task-1277646" [ 899.334720] env[61957]: _type = "Task" [ 899.334720] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.345905] env[61957]: DEBUG oslo_vmware.api [None req-accdd3a1-4d57-48f0-a206-1602b908905b tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Task: {'id': task-1277646, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.584451] env[61957]: INFO nova.compute.manager [None req-7916de58-8954-4a2d-b3f9-5ec9b9a16797 tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] [instance: 0d0f2d34-de35-4e80-8d9f-12693add0786] Took 34.75 seconds to build instance. [ 899.637514] env[61957]: DEBUG oslo_vmware.api [None req-ca59f931-a2d6-4c9e-954c-a383ffdeea61 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Task: {'id': task-1277645, 'name': Rename_Task, 'duration_secs': 0.27759} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.637823] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca59f931-a2d6-4c9e-954c-a383ffdeea61 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] [instance: d5d10a31-0e78-4ed7-b944-9208138a4861] Powering on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 899.638108] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0e821a10-864c-46b5-beeb-abed9981794b {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.651387] env[61957]: DEBUG oslo_vmware.api [None req-ca59f931-a2d6-4c9e-954c-a383ffdeea61 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Waiting for the task: (returnval){ [ 899.651387] env[61957]: value = "task-1277647" [ 899.651387] env[61957]: _type = "Task" [ 899.651387] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.666469] env[61957]: DEBUG oslo_vmware.api [None req-ca59f931-a2d6-4c9e-954c-a383ffdeea61 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Task: {'id': task-1277647, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.761028] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34610ef9-126b-46d5-a11d-bcf0de18541a {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.775384] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c0bc896-758b-40a8-be50-7018df623dc7 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.820209] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fddbcc8d-cf3d-4795-a8d9-aceb47b81ce9 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.835027] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93915c3c-ab8e-4fc1-985d-0a34772a9861 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.849099] env[61957]: DEBUG oslo_vmware.api [None req-accdd3a1-4d57-48f0-a206-1602b908905b tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Task: {'id': task-1277646, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.859561] env[61957]: DEBUG nova.compute.provider_tree [None req-ca2fb2e2-e99f-422a-be65-983bd81b6e15 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 900.086963] env[61957]: DEBUG oslo_concurrency.lockutils [None req-7916de58-8954-4a2d-b3f9-5ec9b9a16797 tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Lock "0d0f2d34-de35-4e80-8d9f-12693add0786" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 80.306s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 900.171126] env[61957]: DEBUG oslo_vmware.api [None req-ca59f931-a2d6-4c9e-954c-a383ffdeea61 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Task: {'id': task-1277647, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.348752] env[61957]: DEBUG oslo_vmware.api [None req-accdd3a1-4d57-48f0-a206-1602b908905b tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Task: {'id': task-1277646, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.698687} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.349091] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-accdd3a1-4d57-48f0-a206-1602b908905b tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore1] a8ca1ad6-d636-4fa2-b0ac-53b020e392ce/a8ca1ad6-d636-4fa2-b0ac-53b020e392ce.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 900.349313] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-accdd3a1-4d57-48f0-a206-1602b908905b tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] [instance: a8ca1ad6-d636-4fa2-b0ac-53b020e392ce] Extending root virtual disk to 1048576 {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 900.349586] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-01b72bd9-dda0-4bfd-b2d3-5644856d7b13 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.359959] env[61957]: DEBUG oslo_vmware.api [None req-accdd3a1-4d57-48f0-a206-1602b908905b tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Waiting for the task: (returnval){ [ 900.359959] env[61957]: value = "task-1277648" [ 900.359959] env[61957]: _type = "Task" [ 900.359959] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.364263] env[61957]: DEBUG nova.scheduler.client.report [None req-ca2fb2e2-e99f-422a-be65-983bd81b6e15 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 900.376410] env[61957]: DEBUG oslo_vmware.api [None req-accdd3a1-4d57-48f0-a206-1602b908905b tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Task: {'id': task-1277648, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.589819] env[61957]: DEBUG nova.compute.manager [None req-1036c4ee-30e7-4291-a697-7231cf748564 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] [instance: 9b0053fe-420e-4b92-86a8-ed44f6a6ec49] Starting instance... {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 900.666363] env[61957]: DEBUG oslo_vmware.api [None req-ca59f931-a2d6-4c9e-954c-a383ffdeea61 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Task: {'id': task-1277647, 'name': PowerOnVM_Task, 'duration_secs': 0.75057} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.666649] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca59f931-a2d6-4c9e-954c-a383ffdeea61 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] [instance: d5d10a31-0e78-4ed7-b944-9208138a4861] Powered on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 900.666876] env[61957]: INFO nova.compute.manager [None req-ca59f931-a2d6-4c9e-954c-a383ffdeea61 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] [instance: d5d10a31-0e78-4ed7-b944-9208138a4861] Took 9.04 seconds to spawn the instance on the hypervisor. [ 900.667066] env[61957]: DEBUG nova.compute.manager [None req-ca59f931-a2d6-4c9e-954c-a383ffdeea61 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] [instance: d5d10a31-0e78-4ed7-b944-9208138a4861] Checking state {{(pid=61957) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 900.668344] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad1a4279-14ab-4299-a8e1-0c3f5e9ed3f0 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.723202] env[61957]: DEBUG oslo_concurrency.lockutils [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Acquiring lock "50e09db7-9f8f-452b-8232-bd7473e9e63a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 900.724644] env[61957]: DEBUG oslo_concurrency.lockutils [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Lock "50e09db7-9f8f-452b-8232-bd7473e9e63a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 900.754708] env[61957]: DEBUG oslo_concurrency.lockutils [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Acquiring lock "2e32aae5-edd7-402f-98ad-75a93d26f7a1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 900.755276] env[61957]: DEBUG oslo_concurrency.lockutils [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Lock "2e32aae5-edd7-402f-98ad-75a93d26f7a1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 900.870992] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ca2fb2e2-e99f-422a-be65-983bd81b6e15 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.456s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 900.871487] env[61957]: DEBUG nova.compute.manager [None req-ca2fb2e2-e99f-422a-be65-983bd81b6e15 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] [instance: 736eabcf-70f8-4e1e-9830-270d9971fc9a] Start building networks asynchronously for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 900.874248] env[61957]: DEBUG oslo_vmware.api [None req-accdd3a1-4d57-48f0-a206-1602b908905b tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Task: {'id': task-1277648, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.277792} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.874579] env[61957]: DEBUG oslo_concurrency.lockutils [None req-3c689a75-14e7-47a1-b918-cef3025704e3 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.590s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 900.875929] env[61957]: INFO nova.compute.claims [None req-3c689a75-14e7-47a1-b918-cef3025704e3 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: dbeeb200-70b9-4cb4-b5a4-182389d21918] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 900.879586] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-accdd3a1-4d57-48f0-a206-1602b908905b tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] [instance: a8ca1ad6-d636-4fa2-b0ac-53b020e392ce] Extended root virtual disk {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 900.880791] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85318a10-2472-4635-b8a8-d7758caff861 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.903907] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-accdd3a1-4d57-48f0-a206-1602b908905b tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] [instance: a8ca1ad6-d636-4fa2-b0ac-53b020e392ce] Reconfiguring VM instance instance-00000049 to attach disk [datastore1] a8ca1ad6-d636-4fa2-b0ac-53b020e392ce/a8ca1ad6-d636-4fa2-b0ac-53b020e392ce.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 900.904042] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9cdee142-8429-4431-b5dd-ecfbba1bd2bb {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.928223] env[61957]: DEBUG oslo_vmware.api [None req-accdd3a1-4d57-48f0-a206-1602b908905b tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Waiting for the task: (returnval){ [ 900.928223] env[61957]: value = "task-1277649" [ 900.928223] env[61957]: _type = "Task" [ 900.928223] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.937998] env[61957]: DEBUG oslo_vmware.api [None req-accdd3a1-4d57-48f0-a206-1602b908905b tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Task: {'id': task-1277649, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.116178] env[61957]: DEBUG oslo_concurrency.lockutils [None req-1036c4ee-30e7-4291-a697-7231cf748564 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 901.188166] env[61957]: INFO nova.compute.manager [None req-ca59f931-a2d6-4c9e-954c-a383ffdeea61 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] [instance: d5d10a31-0e78-4ed7-b944-9208138a4861] Took 37.87 seconds to build instance. [ 901.383893] env[61957]: DEBUG nova.compute.utils [None req-ca2fb2e2-e99f-422a-be65-983bd81b6e15 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Using /dev/sd instead of None {{(pid=61957) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 901.385371] env[61957]: DEBUG nova.compute.manager [None req-ca2fb2e2-e99f-422a-be65-983bd81b6e15 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] [instance: 736eabcf-70f8-4e1e-9830-270d9971fc9a] Allocating IP information in the background. {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 901.385538] env[61957]: DEBUG nova.network.neutron [None req-ca2fb2e2-e99f-422a-be65-983bd81b6e15 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] [instance: 736eabcf-70f8-4e1e-9830-270d9971fc9a] allocate_for_instance() {{(pid=61957) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 901.438704] env[61957]: DEBUG oslo_vmware.api [None req-accdd3a1-4d57-48f0-a206-1602b908905b tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Task: {'id': task-1277649, 'name': ReconfigVM_Task, 'duration_secs': 0.285612} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.439015] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-accdd3a1-4d57-48f0-a206-1602b908905b tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] [instance: a8ca1ad6-d636-4fa2-b0ac-53b020e392ce] Reconfigured VM instance instance-00000049 to attach disk [datastore1] a8ca1ad6-d636-4fa2-b0ac-53b020e392ce/a8ca1ad6-d636-4fa2-b0ac-53b020e392ce.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 901.439654] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a879a67b-91ed-40c1-b1dc-e28e001ea20f {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.447301] env[61957]: DEBUG nova.policy [None req-ca2fb2e2-e99f-422a-be65-983bd81b6e15 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8a534d0b5fc24c3babe7a24cc6b6d941', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '55742b1d4fbb401ea61f990ecaa2def6', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61957) authorize /opt/stack/nova/nova/policy.py:203}} [ 901.451260] env[61957]: DEBUG oslo_vmware.api [None req-accdd3a1-4d57-48f0-a206-1602b908905b tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Waiting for the task: (returnval){ [ 901.451260] env[61957]: value = "task-1277650" [ 901.451260] env[61957]: _type = "Task" [ 901.451260] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.460657] env[61957]: DEBUG oslo_vmware.api [None req-accdd3a1-4d57-48f0-a206-1602b908905b tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Task: {'id': task-1277650, 'name': Rename_Task} progress is 5%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.690855] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ca59f931-a2d6-4c9e-954c-a383ffdeea61 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Lock "d5d10a31-0e78-4ed7-b944-9208138a4861" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 82.135s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 901.768654] env[61957]: DEBUG nova.network.neutron [None req-ca2fb2e2-e99f-422a-be65-983bd81b6e15 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] [instance: 736eabcf-70f8-4e1e-9830-270d9971fc9a] Successfully created port: 95763a65-6383-409f-a9d6-1f37aaf44d34 {{(pid=61957) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 901.890023] env[61957]: DEBUG nova.compute.manager [None req-ca2fb2e2-e99f-422a-be65-983bd81b6e15 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] [instance: 736eabcf-70f8-4e1e-9830-270d9971fc9a] Start building block device mappings for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 901.964070] env[61957]: DEBUG oslo_vmware.api [None req-accdd3a1-4d57-48f0-a206-1602b908905b tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Task: {'id': task-1277650, 'name': Rename_Task, 'duration_secs': 0.166679} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.964383] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-accdd3a1-4d57-48f0-a206-1602b908905b tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] [instance: a8ca1ad6-d636-4fa2-b0ac-53b020e392ce] Powering on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 901.965907] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d7589d65-3200-415e-8c7d-0f7a8c8eb18f {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.972392] env[61957]: DEBUG oslo_vmware.api [None req-accdd3a1-4d57-48f0-a206-1602b908905b tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Waiting for the task: (returnval){ [ 901.972392] env[61957]: value = "task-1277651" [ 901.972392] env[61957]: _type = "Task" [ 901.972392] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.981996] env[61957]: DEBUG oslo_vmware.api [None req-accdd3a1-4d57-48f0-a206-1602b908905b tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Task: {'id': task-1277651, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.194070] env[61957]: DEBUG nova.compute.manager [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: 50e09db7-9f8f-452b-8232-bd7473e9e63a] Starting instance... {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 902.221958] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ec9ceac-6f4a-49a2-a1fd-72af697a9ce3 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.231646] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bf43199-d15e-4b32-ab03-34b4cf11e8bb {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.266411] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2606bd9-5190-430e-9bc0-43d89d1936e0 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.275586] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d3996f0-be40-480e-8dcd-1b8abf64b30e {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.291425] env[61957]: DEBUG nova.compute.provider_tree [None req-3c689a75-14e7-47a1-b918-cef3025704e3 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 902.482527] env[61957]: DEBUG oslo_vmware.api [None req-accdd3a1-4d57-48f0-a206-1602b908905b tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Task: {'id': task-1277651, 'name': PowerOnVM_Task, 'duration_secs': 0.433016} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.482811] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-accdd3a1-4d57-48f0-a206-1602b908905b tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] [instance: a8ca1ad6-d636-4fa2-b0ac-53b020e392ce] Powered on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 902.483032] env[61957]: INFO nova.compute.manager [None req-accdd3a1-4d57-48f0-a206-1602b908905b tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] [instance: a8ca1ad6-d636-4fa2-b0ac-53b020e392ce] Took 5.82 seconds to spawn the instance on the hypervisor. [ 902.483218] env[61957]: DEBUG nova.compute.manager [None req-accdd3a1-4d57-48f0-a206-1602b908905b tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] [instance: a8ca1ad6-d636-4fa2-b0ac-53b020e392ce] Checking state {{(pid=61957) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 902.484009] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11f206a8-4793-441b-a687-0613725ed1a0 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.719943] env[61957]: DEBUG oslo_concurrency.lockutils [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 902.796687] env[61957]: DEBUG nova.scheduler.client.report [None req-3c689a75-14e7-47a1-b918-cef3025704e3 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 902.904454] env[61957]: DEBUG nova.compute.manager [None req-ca2fb2e2-e99f-422a-be65-983bd81b6e15 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] [instance: 736eabcf-70f8-4e1e-9830-270d9971fc9a] Start spawning the instance on the hypervisor. {{(pid=61957) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 902.935839] env[61957]: DEBUG nova.virt.hardware [None req-ca2fb2e2-e99f-422a-be65-983bd81b6e15 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T17:22:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T17:21:50Z,direct_url=,disk_format='vmdk',id=11c76a2c-f705-470a-ba9d-4657858bab38,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='81d3096d0f094373913fc3dc8f5c3c6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T17:21:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 902.936913] env[61957]: DEBUG nova.virt.hardware [None req-ca2fb2e2-e99f-422a-be65-983bd81b6e15 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Flavor limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 902.936913] env[61957]: DEBUG nova.virt.hardware [None req-ca2fb2e2-e99f-422a-be65-983bd81b6e15 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Image limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 902.937223] env[61957]: DEBUG nova.virt.hardware [None req-ca2fb2e2-e99f-422a-be65-983bd81b6e15 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Flavor pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 902.937223] env[61957]: DEBUG nova.virt.hardware [None req-ca2fb2e2-e99f-422a-be65-983bd81b6e15 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Image pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 902.937568] env[61957]: DEBUG nova.virt.hardware [None req-ca2fb2e2-e99f-422a-be65-983bd81b6e15 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 902.937568] env[61957]: DEBUG nova.virt.hardware [None req-ca2fb2e2-e99f-422a-be65-983bd81b6e15 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 902.937734] env[61957]: DEBUG nova.virt.hardware [None req-ca2fb2e2-e99f-422a-be65-983bd81b6e15 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 902.937988] env[61957]: DEBUG nova.virt.hardware [None req-ca2fb2e2-e99f-422a-be65-983bd81b6e15 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Got 1 possible topologies {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 902.938256] env[61957]: DEBUG nova.virt.hardware [None req-ca2fb2e2-e99f-422a-be65-983bd81b6e15 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 902.938467] env[61957]: DEBUG nova.virt.hardware [None req-ca2fb2e2-e99f-422a-be65-983bd81b6e15 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 902.939789] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4d57db0-845d-4119-aa39-5fbdf703805d {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.950532] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed0a8034-9792-427b-8638-aa57aa4cada4 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.003795] env[61957]: INFO nova.compute.manager [None req-accdd3a1-4d57-48f0-a206-1602b908905b tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] [instance: a8ca1ad6-d636-4fa2-b0ac-53b020e392ce] Took 36.22 seconds to build instance. [ 903.206935] env[61957]: DEBUG nova.compute.manager [req-572231b3-4d26-4dc7-894d-40dbaa165e16 req-e3a8a32f-e33c-4545-a7e7-ce3434144b18 service nova] [instance: 736eabcf-70f8-4e1e-9830-270d9971fc9a] Received event network-vif-plugged-95763a65-6383-409f-a9d6-1f37aaf44d34 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 903.206935] env[61957]: DEBUG oslo_concurrency.lockutils [req-572231b3-4d26-4dc7-894d-40dbaa165e16 req-e3a8a32f-e33c-4545-a7e7-ce3434144b18 service nova] Acquiring lock "736eabcf-70f8-4e1e-9830-270d9971fc9a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 903.206935] env[61957]: DEBUG oslo_concurrency.lockutils [req-572231b3-4d26-4dc7-894d-40dbaa165e16 req-e3a8a32f-e33c-4545-a7e7-ce3434144b18 service nova] Lock "736eabcf-70f8-4e1e-9830-270d9971fc9a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 903.206935] env[61957]: DEBUG oslo_concurrency.lockutils [req-572231b3-4d26-4dc7-894d-40dbaa165e16 req-e3a8a32f-e33c-4545-a7e7-ce3434144b18 service nova] Lock "736eabcf-70f8-4e1e-9830-270d9971fc9a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 903.206935] env[61957]: DEBUG nova.compute.manager [req-572231b3-4d26-4dc7-894d-40dbaa165e16 req-e3a8a32f-e33c-4545-a7e7-ce3434144b18 service nova] [instance: 736eabcf-70f8-4e1e-9830-270d9971fc9a] No waiting events found dispatching network-vif-plugged-95763a65-6383-409f-a9d6-1f37aaf44d34 {{(pid=61957) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 903.206935] env[61957]: WARNING nova.compute.manager [req-572231b3-4d26-4dc7-894d-40dbaa165e16 req-e3a8a32f-e33c-4545-a7e7-ce3434144b18 service nova] [instance: 736eabcf-70f8-4e1e-9830-270d9971fc9a] Received unexpected event network-vif-plugged-95763a65-6383-409f-a9d6-1f37aaf44d34 for instance with vm_state building and task_state spawning. [ 903.304086] env[61957]: DEBUG nova.network.neutron [None req-ca2fb2e2-e99f-422a-be65-983bd81b6e15 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] [instance: 736eabcf-70f8-4e1e-9830-270d9971fc9a] Successfully updated port: 95763a65-6383-409f-a9d6-1f37aaf44d34 {{(pid=61957) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 903.304086] env[61957]: DEBUG oslo_concurrency.lockutils [None req-3c689a75-14e7-47a1-b918-cef3025704e3 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.427s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 903.304086] env[61957]: DEBUG nova.compute.manager [None req-3c689a75-14e7-47a1-b918-cef3025704e3 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: dbeeb200-70b9-4cb4-b5a4-182389d21918] Start building networks asynchronously for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 903.306189] env[61957]: DEBUG oslo_concurrency.lockutils [None req-fd686081-caee-4453-9fa7-68efd0fac70e tempest-AttachInterfacesV270Test-2032541732 tempest-AttachInterfacesV270Test-2032541732-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 27.937s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 903.306479] env[61957]: DEBUG nova.objects.instance [None req-fd686081-caee-4453-9fa7-68efd0fac70e tempest-AttachInterfacesV270Test-2032541732 tempest-AttachInterfacesV270Test-2032541732-project-member] Lazy-loading 'resources' on Instance uuid 281b9644-0e7d-48b9-a7b6-45fd6102d558 {{(pid=61957) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 903.505835] env[61957]: DEBUG oslo_concurrency.lockutils [None req-accdd3a1-4d57-48f0-a206-1602b908905b tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Lock "a8ca1ad6-d636-4fa2-b0ac-53b020e392ce" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 83.489s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 903.613909] env[61957]: DEBUG oslo_concurrency.lockutils [None req-8e2c2702-9da9-4cbe-8ea6-ee174b7beb38 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Acquiring lock "e40fa112-4648-428f-a403-b3e3b8319ea8" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 903.615041] env[61957]: DEBUG oslo_concurrency.lockutils [None req-8e2c2702-9da9-4cbe-8ea6-ee174b7beb38 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Lock "e40fa112-4648-428f-a403-b3e3b8319ea8" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 903.615441] env[61957]: DEBUG nova.compute.manager [None req-8e2c2702-9da9-4cbe-8ea6-ee174b7beb38 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] [instance: e40fa112-4648-428f-a403-b3e3b8319ea8] Checking state {{(pid=61957) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 903.616430] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26d7e9c7-b85f-40c3-81c8-747abaa59920 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.624102] env[61957]: DEBUG nova.compute.manager [None req-8e2c2702-9da9-4cbe-8ea6-ee174b7beb38 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] [instance: e40fa112-4648-428f-a403-b3e3b8319ea8] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61957) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 903.624677] env[61957]: DEBUG nova.objects.instance [None req-8e2c2702-9da9-4cbe-8ea6-ee174b7beb38 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Lazy-loading 'flavor' on Instance uuid e40fa112-4648-428f-a403-b3e3b8319ea8 {{(pid=61957) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 903.807580] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ca2fb2e2-e99f-422a-be65-983bd81b6e15 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Acquiring lock "refresh_cache-736eabcf-70f8-4e1e-9830-270d9971fc9a" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 903.807921] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ca2fb2e2-e99f-422a-be65-983bd81b6e15 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Acquired lock "refresh_cache-736eabcf-70f8-4e1e-9830-270d9971fc9a" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 903.808078] env[61957]: DEBUG nova.network.neutron [None req-ca2fb2e2-e99f-422a-be65-983bd81b6e15 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] [instance: 736eabcf-70f8-4e1e-9830-270d9971fc9a] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 903.811835] env[61957]: DEBUG nova.compute.utils [None req-3c689a75-14e7-47a1-b918-cef3025704e3 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Using /dev/sd instead of None {{(pid=61957) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 903.815333] env[61957]: DEBUG nova.compute.manager [None req-3c689a75-14e7-47a1-b918-cef3025704e3 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: dbeeb200-70b9-4cb4-b5a4-182389d21918] Allocating IP information in the background. {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 903.815333] env[61957]: DEBUG nova.network.neutron [None req-3c689a75-14e7-47a1-b918-cef3025704e3 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: dbeeb200-70b9-4cb4-b5a4-182389d21918] allocate_for_instance() {{(pid=61957) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 903.864582] env[61957]: DEBUG nova.policy [None req-3c689a75-14e7-47a1-b918-cef3025704e3 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3136bf0d1db3411fb1ff1ebe9c6045f6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1dbdc699741a48af93c52068d88a5357', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61957) authorize /opt/stack/nova/nova/policy.py:203}} [ 904.008522] env[61957]: DEBUG nova.compute.manager [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: 2e32aae5-edd7-402f-98ad-75a93d26f7a1] Starting instance... {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 904.073356] env[61957]: DEBUG oslo_concurrency.lockutils [None req-c02d7885-6c3b-4a1f-ab8a-8e683ed8e586 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Acquiring lock "c23141ee-0cbb-4d1b-8390-c3073fe354f1" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 904.073591] env[61957]: DEBUG oslo_concurrency.lockutils [None req-c02d7885-6c3b-4a1f-ab8a-8e683ed8e586 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Lock "c23141ee-0cbb-4d1b-8390-c3073fe354f1" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 904.116610] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fff2b0e3-1c9a-449c-8889-7fde2c331a00 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.129962] env[61957]: DEBUG nova.network.neutron [None req-3c689a75-14e7-47a1-b918-cef3025704e3 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: dbeeb200-70b9-4cb4-b5a4-182389d21918] Successfully created port: 86553395-da58-4c44-b9f4-c67db304d3fe {{(pid=61957) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 904.133045] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e5d01d8-dbbd-461f-a839-ce7fbdf2519a {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.136477] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-8e2c2702-9da9-4cbe-8ea6-ee174b7beb38 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] [instance: e40fa112-4648-428f-a403-b3e3b8319ea8] Powering off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 904.137021] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fe73b39f-afca-4fb0-a3f3-8dc5c0255d9c {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.173901] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30fcae55-19ab-4397-82cb-5e07db87cf36 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.177090] env[61957]: DEBUG oslo_vmware.api [None req-8e2c2702-9da9-4cbe-8ea6-ee174b7beb38 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Waiting for the task: (returnval){ [ 904.177090] env[61957]: value = "task-1277652" [ 904.177090] env[61957]: _type = "Task" [ 904.177090] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.184327] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50f950d1-8d63-42e8-b89c-ee103172590f {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.190954] env[61957]: DEBUG oslo_vmware.api [None req-8e2c2702-9da9-4cbe-8ea6-ee174b7beb38 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Task: {'id': task-1277652, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.201027] env[61957]: INFO nova.compute.manager [None req-35c8e096-fa6b-43c8-874d-91a51735ac8e tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] [instance: a8ca1ad6-d636-4fa2-b0ac-53b020e392ce] Rebuilding instance [ 904.203192] env[61957]: DEBUG nova.compute.provider_tree [None req-fd686081-caee-4453-9fa7-68efd0fac70e tempest-AttachInterfacesV270Test-2032541732 tempest-AttachInterfacesV270Test-2032541732-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 904.243925] env[61957]: DEBUG nova.compute.manager [None req-35c8e096-fa6b-43c8-874d-91a51735ac8e tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] [instance: a8ca1ad6-d636-4fa2-b0ac-53b020e392ce] Checking state {{(pid=61957) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 904.244807] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0202508-ba6f-47c4-a9c8-4d4ba121e407 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.319439] env[61957]: DEBUG nova.compute.manager [None req-3c689a75-14e7-47a1-b918-cef3025704e3 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: dbeeb200-70b9-4cb4-b5a4-182389d21918] Start building block device mappings for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 904.354187] env[61957]: DEBUG nova.network.neutron [None req-ca2fb2e2-e99f-422a-be65-983bd81b6e15 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] [instance: 736eabcf-70f8-4e1e-9830-270d9971fc9a] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 904.525128] env[61957]: DEBUG nova.network.neutron [None req-ca2fb2e2-e99f-422a-be65-983bd81b6e15 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] [instance: 736eabcf-70f8-4e1e-9830-270d9971fc9a] Updating instance_info_cache with network_info: [{"id": "95763a65-6383-409f-a9d6-1f37aaf44d34", "address": "fa:16:3e:64:07:83", "network": {"id": "b303ad01-ef9b-4119-9722-3be27ffb30bd", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-96064402-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "55742b1d4fbb401ea61f990ecaa2def6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a8b99a46-3e7f-4ef1-9e45-58e6cd17f210", "external-id": "nsx-vlan-transportzone-704", "segmentation_id": 704, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap95763a65-63", "ovs_interfaceid": "95763a65-6383-409f-a9d6-1f37aaf44d34", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 904.532592] env[61957]: DEBUG oslo_concurrency.lockutils [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 904.576579] env[61957]: DEBUG nova.compute.utils [None req-c02d7885-6c3b-4a1f-ab8a-8e683ed8e586 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Using /dev/sd instead of None {{(pid=61957) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 904.700785] env[61957]: DEBUG oslo_vmware.api [None req-8e2c2702-9da9-4cbe-8ea6-ee174b7beb38 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Task: {'id': task-1277652, 'name': PowerOffVM_Task} progress is 100%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.705824] env[61957]: DEBUG nova.scheduler.client.report [None req-fd686081-caee-4453-9fa7-68efd0fac70e tempest-AttachInterfacesV270Test-2032541732 tempest-AttachInterfacesV270Test-2032541732-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 904.756574] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-35c8e096-fa6b-43c8-874d-91a51735ac8e tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] [instance: a8ca1ad6-d636-4fa2-b0ac-53b020e392ce] Powering off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 904.756877] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-662d6322-f10b-48e4-b5af-63347fda95c3 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.764729] env[61957]: DEBUG oslo_vmware.api [None req-35c8e096-fa6b-43c8-874d-91a51735ac8e tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Waiting for the task: (returnval){ [ 904.764729] env[61957]: value = "task-1277653" [ 904.764729] env[61957]: _type = "Task" [ 904.764729] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.774057] env[61957]: DEBUG oslo_vmware.api [None req-35c8e096-fa6b-43c8-874d-91a51735ac8e tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Task: {'id': task-1277653, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.028090] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ca2fb2e2-e99f-422a-be65-983bd81b6e15 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Releasing lock "refresh_cache-736eabcf-70f8-4e1e-9830-270d9971fc9a" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 905.028577] env[61957]: DEBUG nova.compute.manager [None req-ca2fb2e2-e99f-422a-be65-983bd81b6e15 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] [instance: 736eabcf-70f8-4e1e-9830-270d9971fc9a] Instance network_info: |[{"id": "95763a65-6383-409f-a9d6-1f37aaf44d34", "address": "fa:16:3e:64:07:83", "network": {"id": "b303ad01-ef9b-4119-9722-3be27ffb30bd", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-96064402-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "55742b1d4fbb401ea61f990ecaa2def6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a8b99a46-3e7f-4ef1-9e45-58e6cd17f210", "external-id": "nsx-vlan-transportzone-704", "segmentation_id": 704, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap95763a65-63", "ovs_interfaceid": "95763a65-6383-409f-a9d6-1f37aaf44d34", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 905.029283] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-ca2fb2e2-e99f-422a-be65-983bd81b6e15 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] [instance: 736eabcf-70f8-4e1e-9830-270d9971fc9a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:64:07:83', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a8b99a46-3e7f-4ef1-9e45-58e6cd17f210', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '95763a65-6383-409f-a9d6-1f37aaf44d34', 'vif_model': 'vmxnet3'}] {{(pid=61957) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 905.037498] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca2fb2e2-e99f-422a-be65-983bd81b6e15 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Creating folder: Project (55742b1d4fbb401ea61f990ecaa2def6). Parent ref: group-v274445. {{(pid=61957) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 905.037810] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b11f8c93-667d-4f20-87da-95cfbd6bd473 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.050954] env[61957]: INFO nova.virt.vmwareapi.vm_util [None req-ca2fb2e2-e99f-422a-be65-983bd81b6e15 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Created folder: Project (55742b1d4fbb401ea61f990ecaa2def6) in parent group-v274445. [ 905.051178] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca2fb2e2-e99f-422a-be65-983bd81b6e15 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Creating folder: Instances. Parent ref: group-v274541. {{(pid=61957) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 905.051461] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-70b6025f-b3fd-4961-ba42-02c012c4d62f {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.063180] env[61957]: INFO nova.virt.vmwareapi.vm_util [None req-ca2fb2e2-e99f-422a-be65-983bd81b6e15 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Created folder: Instances in parent group-v274541. [ 905.063455] env[61957]: DEBUG oslo.service.loopingcall [None req-ca2fb2e2-e99f-422a-be65-983bd81b6e15 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 905.063665] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 736eabcf-70f8-4e1e-9830-270d9971fc9a] Creating VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 905.063882] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ad028f88-5e5d-4f8f-a908-63a87986f3ab {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.079166] env[61957]: DEBUG oslo_concurrency.lockutils [None req-c02d7885-6c3b-4a1f-ab8a-8e683ed8e586 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Lock "c23141ee-0cbb-4d1b-8390-c3073fe354f1" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.005s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 905.087184] env[61957]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 905.087184] env[61957]: value = "task-1277656" [ 905.087184] env[61957]: _type = "Task" [ 905.087184] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.095778] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277656, 'name': CreateVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.189430] env[61957]: DEBUG oslo_vmware.api [None req-8e2c2702-9da9-4cbe-8ea6-ee174b7beb38 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Task: {'id': task-1277652, 'name': PowerOffVM_Task, 'duration_secs': 0.972632} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.189792] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-8e2c2702-9da9-4cbe-8ea6-ee174b7beb38 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] [instance: e40fa112-4648-428f-a403-b3e3b8319ea8] Powered off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 905.190010] env[61957]: DEBUG nova.compute.manager [None req-8e2c2702-9da9-4cbe-8ea6-ee174b7beb38 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] [instance: e40fa112-4648-428f-a403-b3e3b8319ea8] Checking state {{(pid=61957) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 905.190967] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ed3eace-79f6-4e24-8bac-2f15195a6d52 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.211107] env[61957]: DEBUG oslo_concurrency.lockutils [None req-fd686081-caee-4453-9fa7-68efd0fac70e tempest-AttachInterfacesV270Test-2032541732 tempest-AttachInterfacesV270Test-2032541732-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.904s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 905.213387] env[61957]: DEBUG oslo_concurrency.lockutils [None req-a7399dbc-0983-4a91-82d1-9c1b173d4f53 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.343s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 905.216239] env[61957]: INFO nova.compute.claims [None req-a7399dbc-0983-4a91-82d1-9c1b173d4f53 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 39b8b1df-efdd-4c22-9f31-85c85be4f0eb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 905.232753] env[61957]: INFO nova.scheduler.client.report [None req-fd686081-caee-4453-9fa7-68efd0fac70e tempest-AttachInterfacesV270Test-2032541732 tempest-AttachInterfacesV270Test-2032541732-project-member] Deleted allocations for instance 281b9644-0e7d-48b9-a7b6-45fd6102d558 [ 905.248351] env[61957]: DEBUG nova.compute.manager [req-e0c5d734-ff6f-4ee6-82e6-5a3447e4eb2d req-b3b04ef6-3a98-414e-97e4-297c877676da service nova] [instance: 736eabcf-70f8-4e1e-9830-270d9971fc9a] Received event network-changed-95763a65-6383-409f-a9d6-1f37aaf44d34 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 905.248981] env[61957]: DEBUG nova.compute.manager [req-e0c5d734-ff6f-4ee6-82e6-5a3447e4eb2d req-b3b04ef6-3a98-414e-97e4-297c877676da service nova] [instance: 736eabcf-70f8-4e1e-9830-270d9971fc9a] Refreshing instance network info cache due to event network-changed-95763a65-6383-409f-a9d6-1f37aaf44d34. {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 905.248981] env[61957]: DEBUG oslo_concurrency.lockutils [req-e0c5d734-ff6f-4ee6-82e6-5a3447e4eb2d req-b3b04ef6-3a98-414e-97e4-297c877676da service nova] Acquiring lock "refresh_cache-736eabcf-70f8-4e1e-9830-270d9971fc9a" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 905.249206] env[61957]: DEBUG oslo_concurrency.lockutils [req-e0c5d734-ff6f-4ee6-82e6-5a3447e4eb2d req-b3b04ef6-3a98-414e-97e4-297c877676da service nova] Acquired lock "refresh_cache-736eabcf-70f8-4e1e-9830-270d9971fc9a" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 905.249239] env[61957]: DEBUG nova.network.neutron [req-e0c5d734-ff6f-4ee6-82e6-5a3447e4eb2d req-b3b04ef6-3a98-414e-97e4-297c877676da service nova] [instance: 736eabcf-70f8-4e1e-9830-270d9971fc9a] Refreshing network info cache for port 95763a65-6383-409f-a9d6-1f37aaf44d34 {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 905.277200] env[61957]: DEBUG oslo_vmware.api [None req-35c8e096-fa6b-43c8-874d-91a51735ac8e tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Task: {'id': task-1277653, 'name': PowerOffVM_Task, 'duration_secs': 0.21081} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.277496] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-35c8e096-fa6b-43c8-874d-91a51735ac8e tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] [instance: a8ca1ad6-d636-4fa2-b0ac-53b020e392ce] Powered off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 905.277720] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-35c8e096-fa6b-43c8-874d-91a51735ac8e tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] [instance: a8ca1ad6-d636-4fa2-b0ac-53b020e392ce] Destroying instance {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 905.278634] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ede2af1b-18c2-4e78-83ba-5a8446047e10 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.290609] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-35c8e096-fa6b-43c8-874d-91a51735ac8e tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] [instance: a8ca1ad6-d636-4fa2-b0ac-53b020e392ce] Unregistering the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 905.291258] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2f13d850-b421-42e7-921f-cbccdcd8f25c {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.322209] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-35c8e096-fa6b-43c8-874d-91a51735ac8e tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] [instance: a8ca1ad6-d636-4fa2-b0ac-53b020e392ce] Unregistered the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 905.322468] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-35c8e096-fa6b-43c8-874d-91a51735ac8e tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] [instance: a8ca1ad6-d636-4fa2-b0ac-53b020e392ce] Deleting contents of the VM from datastore datastore1 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 905.322678] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-35c8e096-fa6b-43c8-874d-91a51735ac8e tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Deleting the datastore file [datastore1] a8ca1ad6-d636-4fa2-b0ac-53b020e392ce {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 905.323030] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1dcea21b-67e9-48fa-bdcb-7aa3c9b74a18 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.329684] env[61957]: DEBUG nova.compute.manager [None req-3c689a75-14e7-47a1-b918-cef3025704e3 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: dbeeb200-70b9-4cb4-b5a4-182389d21918] Start spawning the instance on the hypervisor. {{(pid=61957) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 905.333320] env[61957]: DEBUG oslo_vmware.api [None req-35c8e096-fa6b-43c8-874d-91a51735ac8e tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Waiting for the task: (returnval){ [ 905.333320] env[61957]: value = "task-1277658" [ 905.333320] env[61957]: _type = "Task" [ 905.333320] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.343124] env[61957]: DEBUG oslo_vmware.api [None req-35c8e096-fa6b-43c8-874d-91a51735ac8e tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Task: {'id': task-1277658, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.358354] env[61957]: DEBUG nova.virt.hardware [None req-3c689a75-14e7-47a1-b918-cef3025704e3 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T17:22:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T17:21:50Z,direct_url=,disk_format='vmdk',id=11c76a2c-f705-470a-ba9d-4657858bab38,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='81d3096d0f094373913fc3dc8f5c3c6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T17:21:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 905.358525] env[61957]: DEBUG nova.virt.hardware [None req-3c689a75-14e7-47a1-b918-cef3025704e3 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Flavor limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 905.358747] env[61957]: DEBUG nova.virt.hardware [None req-3c689a75-14e7-47a1-b918-cef3025704e3 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Image limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 905.358964] env[61957]: DEBUG nova.virt.hardware [None req-3c689a75-14e7-47a1-b918-cef3025704e3 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Flavor pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 905.359138] env[61957]: DEBUG nova.virt.hardware [None req-3c689a75-14e7-47a1-b918-cef3025704e3 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Image pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 905.359291] env[61957]: DEBUG nova.virt.hardware [None req-3c689a75-14e7-47a1-b918-cef3025704e3 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 905.359502] env[61957]: DEBUG nova.virt.hardware [None req-3c689a75-14e7-47a1-b918-cef3025704e3 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 905.359687] env[61957]: DEBUG nova.virt.hardware [None req-3c689a75-14e7-47a1-b918-cef3025704e3 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 905.359862] env[61957]: DEBUG nova.virt.hardware [None req-3c689a75-14e7-47a1-b918-cef3025704e3 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Got 1 possible topologies {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 905.360074] env[61957]: DEBUG nova.virt.hardware [None req-3c689a75-14e7-47a1-b918-cef3025704e3 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 905.360266] env[61957]: DEBUG nova.virt.hardware [None req-3c689a75-14e7-47a1-b918-cef3025704e3 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 905.361202] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d919fad3-f030-4ef0-8538-c6d488a73234 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.370246] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a535b262-8b91-456e-862a-86dd31f25fd8 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.597601] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277656, 'name': CreateVM_Task, 'duration_secs': 0.397998} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.597803] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 736eabcf-70f8-4e1e-9830-270d9971fc9a] Created VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 905.598512] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ca2fb2e2-e99f-422a-be65-983bd81b6e15 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 905.598689] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ca2fb2e2-e99f-422a-be65-983bd81b6e15 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Acquired lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 905.599031] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ca2fb2e2-e99f-422a-be65-983bd81b6e15 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 905.599293] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3b0959fb-d329-468f-ba43-1f80d2d44d23 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.605109] env[61957]: DEBUG oslo_vmware.api [None req-ca2fb2e2-e99f-422a-be65-983bd81b6e15 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Waiting for the task: (returnval){ [ 905.605109] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]524f2f2c-98f1-2374-97da-97f84d4d2000" [ 905.605109] env[61957]: _type = "Task" [ 905.605109] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.613709] env[61957]: DEBUG oslo_vmware.api [None req-ca2fb2e2-e99f-422a-be65-983bd81b6e15 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]524f2f2c-98f1-2374-97da-97f84d4d2000, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.705022] env[61957]: DEBUG oslo_concurrency.lockutils [None req-8e2c2702-9da9-4cbe-8ea6-ee174b7beb38 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Lock "e40fa112-4648-428f-a403-b3e3b8319ea8" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.090s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 905.739024] env[61957]: DEBUG oslo_concurrency.lockutils [None req-fd686081-caee-4453-9fa7-68efd0fac70e tempest-AttachInterfacesV270Test-2032541732 tempest-AttachInterfacesV270Test-2032541732-project-member] Lock "281b9644-0e7d-48b9-a7b6-45fd6102d558" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 33.320s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 905.765715] env[61957]: DEBUG nova.network.neutron [None req-3c689a75-14e7-47a1-b918-cef3025704e3 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: dbeeb200-70b9-4cb4-b5a4-182389d21918] Successfully updated port: 86553395-da58-4c44-b9f4-c67db304d3fe {{(pid=61957) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 905.843129] env[61957]: DEBUG oslo_vmware.api [None req-35c8e096-fa6b-43c8-874d-91a51735ac8e tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Task: {'id': task-1277658, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.188578} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.843427] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-35c8e096-fa6b-43c8-874d-91a51735ac8e tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Deleted the datastore file {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 905.843645] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-35c8e096-fa6b-43c8-874d-91a51735ac8e tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] [instance: a8ca1ad6-d636-4fa2-b0ac-53b020e392ce] Deleted contents of the VM from datastore datastore1 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 905.843866] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-35c8e096-fa6b-43c8-874d-91a51735ac8e tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] [instance: a8ca1ad6-d636-4fa2-b0ac-53b020e392ce] Instance destroyed {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 906.105422] env[61957]: DEBUG nova.network.neutron [req-e0c5d734-ff6f-4ee6-82e6-5a3447e4eb2d req-b3b04ef6-3a98-414e-97e4-297c877676da service nova] [instance: 736eabcf-70f8-4e1e-9830-270d9971fc9a] Updated VIF entry in instance network info cache for port 95763a65-6383-409f-a9d6-1f37aaf44d34. {{(pid=61957) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 906.105793] env[61957]: DEBUG nova.network.neutron [req-e0c5d734-ff6f-4ee6-82e6-5a3447e4eb2d req-b3b04ef6-3a98-414e-97e4-297c877676da service nova] [instance: 736eabcf-70f8-4e1e-9830-270d9971fc9a] Updating instance_info_cache with network_info: [{"id": "95763a65-6383-409f-a9d6-1f37aaf44d34", "address": "fa:16:3e:64:07:83", "network": {"id": "b303ad01-ef9b-4119-9722-3be27ffb30bd", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-96064402-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "55742b1d4fbb401ea61f990ecaa2def6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a8b99a46-3e7f-4ef1-9e45-58e6cd17f210", "external-id": "nsx-vlan-transportzone-704", "segmentation_id": 704, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap95763a65-63", "ovs_interfaceid": "95763a65-6383-409f-a9d6-1f37aaf44d34", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 906.117261] env[61957]: DEBUG oslo_vmware.api [None req-ca2fb2e2-e99f-422a-be65-983bd81b6e15 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]524f2f2c-98f1-2374-97da-97f84d4d2000, 'name': SearchDatastore_Task, 'duration_secs': 0.017334} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.118094] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ca2fb2e2-e99f-422a-be65-983bd81b6e15 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Releasing lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 906.118337] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-ca2fb2e2-e99f-422a-be65-983bd81b6e15 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] [instance: 736eabcf-70f8-4e1e-9830-270d9971fc9a] Processing image 11c76a2c-f705-470a-ba9d-4657858bab38 {{(pid=61957) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 906.118577] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ca2fb2e2-e99f-422a-be65-983bd81b6e15 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 906.118725] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ca2fb2e2-e99f-422a-be65-983bd81b6e15 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Acquired lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 906.118905] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-ca2fb2e2-e99f-422a-be65-983bd81b6e15 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 906.119415] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8dc9e2de-dc1e-4692-972e-51f9b3b6f69c {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.128296] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-ca2fb2e2-e99f-422a-be65-983bd81b6e15 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 906.128482] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-ca2fb2e2-e99f-422a-be65-983bd81b6e15 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61957) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 906.129207] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-30a120ae-e2d2-459e-995e-fd842d673004 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.134880] env[61957]: DEBUG oslo_vmware.api [None req-ca2fb2e2-e99f-422a-be65-983bd81b6e15 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Waiting for the task: (returnval){ [ 906.134880] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52935abf-735e-2cab-3c23-db274c579a64" [ 906.134880] env[61957]: _type = "Task" [ 906.134880] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.142228] env[61957]: DEBUG oslo_vmware.api [None req-ca2fb2e2-e99f-422a-be65-983bd81b6e15 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52935abf-735e-2cab-3c23-db274c579a64, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.150918] env[61957]: DEBUG oslo_concurrency.lockutils [None req-c02d7885-6c3b-4a1f-ab8a-8e683ed8e586 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Acquiring lock "c23141ee-0cbb-4d1b-8390-c3073fe354f1" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 906.151168] env[61957]: DEBUG oslo_concurrency.lockutils [None req-c02d7885-6c3b-4a1f-ab8a-8e683ed8e586 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Lock "c23141ee-0cbb-4d1b-8390-c3073fe354f1" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 906.151392] env[61957]: INFO nova.compute.manager [None req-c02d7885-6c3b-4a1f-ab8a-8e683ed8e586 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] [instance: c23141ee-0cbb-4d1b-8390-c3073fe354f1] Attaching volume 734d094d-c382-4070-87c9-699fa1355b79 to /dev/sdb [ 906.185371] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3d549b5-c312-4166-aaf9-5dcec7a2c3af {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.193763] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf31f91e-68e3-4de2-93a3-ebf81bbbc8bf {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.210942] env[61957]: DEBUG nova.virt.block_device [None req-c02d7885-6c3b-4a1f-ab8a-8e683ed8e586 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] [instance: c23141ee-0cbb-4d1b-8390-c3073fe354f1] Updating existing volume attachment record: e7935eaa-0519-4020-a8d4-d31350f7ca0d {{(pid=61957) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 906.217821] env[61957]: DEBUG nova.objects.instance [None req-6c9b1d92-3397-4471-b73b-76c509e782fc tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Lazy-loading 'flavor' on Instance uuid e40fa112-4648-428f-a403-b3e3b8319ea8 {{(pid=61957) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 906.273380] env[61957]: DEBUG oslo_concurrency.lockutils [None req-3c689a75-14e7-47a1-b918-cef3025704e3 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Acquiring lock "refresh_cache-dbeeb200-70b9-4cb4-b5a4-182389d21918" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 906.273380] env[61957]: DEBUG oslo_concurrency.lockutils [None req-3c689a75-14e7-47a1-b918-cef3025704e3 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Acquired lock "refresh_cache-dbeeb200-70b9-4cb4-b5a4-182389d21918" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 906.273380] env[61957]: DEBUG nova.network.neutron [None req-3c689a75-14e7-47a1-b918-cef3025704e3 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: dbeeb200-70b9-4cb4-b5a4-182389d21918] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 906.563608] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47219c1e-2f37-412b-a968-127dbd46a04c {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.572428] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a85560d-9f80-466d-9680-9591ef8d12e3 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.615189] env[61957]: DEBUG oslo_concurrency.lockutils [req-e0c5d734-ff6f-4ee6-82e6-5a3447e4eb2d req-b3b04ef6-3a98-414e-97e4-297c877676da service nova] Releasing lock "refresh_cache-736eabcf-70f8-4e1e-9830-270d9971fc9a" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 906.615189] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4aec1300-984c-4011-bd22-a55c7aa7efd6 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.622748] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-befb4be2-8d18-405d-97d1-5011d534d75c {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.637343] env[61957]: DEBUG nova.compute.provider_tree [None req-a7399dbc-0983-4a91-82d1-9c1b173d4f53 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 906.648897] env[61957]: DEBUG oslo_vmware.api [None req-ca2fb2e2-e99f-422a-be65-983bd81b6e15 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52935abf-735e-2cab-3c23-db274c579a64, 'name': SearchDatastore_Task, 'duration_secs': 0.010237} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.649725] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-38ae88e0-7c40-4ac7-ad90-a681cb8ba784 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.656615] env[61957]: DEBUG oslo_vmware.api [None req-ca2fb2e2-e99f-422a-be65-983bd81b6e15 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Waiting for the task: (returnval){ [ 906.656615] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52a9bfce-89aa-a8f6-4e18-f33f8460d400" [ 906.656615] env[61957]: _type = "Task" [ 906.656615] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.665278] env[61957]: DEBUG oslo_vmware.api [None req-ca2fb2e2-e99f-422a-be65-983bd81b6e15 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52a9bfce-89aa-a8f6-4e18-f33f8460d400, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.726281] env[61957]: DEBUG oslo_concurrency.lockutils [None req-6c9b1d92-3397-4471-b73b-76c509e782fc tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Acquiring lock "refresh_cache-e40fa112-4648-428f-a403-b3e3b8319ea8" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 906.726281] env[61957]: DEBUG oslo_concurrency.lockutils [None req-6c9b1d92-3397-4471-b73b-76c509e782fc tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Acquired lock "refresh_cache-e40fa112-4648-428f-a403-b3e3b8319ea8" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 906.726523] env[61957]: DEBUG nova.network.neutron [None req-6c9b1d92-3397-4471-b73b-76c509e782fc tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] [instance: e40fa112-4648-428f-a403-b3e3b8319ea8] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 906.726523] env[61957]: DEBUG nova.objects.instance [None req-6c9b1d92-3397-4471-b73b-76c509e782fc tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Lazy-loading 'info_cache' on Instance uuid e40fa112-4648-428f-a403-b3e3b8319ea8 {{(pid=61957) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 906.820716] env[61957]: DEBUG nova.network.neutron [None req-3c689a75-14e7-47a1-b918-cef3025704e3 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: dbeeb200-70b9-4cb4-b5a4-182389d21918] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 906.891126] env[61957]: DEBUG nova.virt.hardware [None req-35c8e096-fa6b-43c8-874d-91a51735ac8e tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T17:22:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T17:21:50Z,direct_url=,disk_format='vmdk',id=11c76a2c-f705-470a-ba9d-4657858bab38,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='81d3096d0f094373913fc3dc8f5c3c6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T17:21:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 906.891378] env[61957]: DEBUG nova.virt.hardware [None req-35c8e096-fa6b-43c8-874d-91a51735ac8e tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Flavor limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 906.891557] env[61957]: DEBUG nova.virt.hardware [None req-35c8e096-fa6b-43c8-874d-91a51735ac8e tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Image limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 906.891810] env[61957]: DEBUG nova.virt.hardware [None req-35c8e096-fa6b-43c8-874d-91a51735ac8e tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Flavor pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 906.891993] env[61957]: DEBUG nova.virt.hardware [None req-35c8e096-fa6b-43c8-874d-91a51735ac8e tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Image pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 906.892177] env[61957]: DEBUG nova.virt.hardware [None req-35c8e096-fa6b-43c8-874d-91a51735ac8e tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 906.893283] env[61957]: DEBUG nova.virt.hardware [None req-35c8e096-fa6b-43c8-874d-91a51735ac8e tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 906.893283] env[61957]: DEBUG nova.virt.hardware [None req-35c8e096-fa6b-43c8-874d-91a51735ac8e tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 906.893283] env[61957]: DEBUG nova.virt.hardware [None req-35c8e096-fa6b-43c8-874d-91a51735ac8e tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Got 1 possible topologies {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 906.893283] env[61957]: DEBUG nova.virt.hardware [None req-35c8e096-fa6b-43c8-874d-91a51735ac8e tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 906.893283] env[61957]: DEBUG nova.virt.hardware [None req-35c8e096-fa6b-43c8-874d-91a51735ac8e tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 906.894604] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f3d77aa-b880-4a27-b843-d320304e2ee6 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.907023] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85319814-ffcf-469b-b64f-b1c056a7627a {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.920363] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-35c8e096-fa6b-43c8-874d-91a51735ac8e tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] [instance: a8ca1ad6-d636-4fa2-b0ac-53b020e392ce] Instance VIF info [] {{(pid=61957) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 906.926693] env[61957]: DEBUG oslo.service.loopingcall [None req-35c8e096-fa6b-43c8-874d-91a51735ac8e tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 906.927944] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a8ca1ad6-d636-4fa2-b0ac-53b020e392ce] Creating VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 906.928350] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2f0b805e-e47d-474c-bc2f-28df6a92627b {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.956416] env[61957]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 906.956416] env[61957]: value = "task-1277662" [ 906.956416] env[61957]: _type = "Task" [ 906.956416] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.965279] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277662, 'name': CreateVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.074604] env[61957]: DEBUG nova.network.neutron [None req-3c689a75-14e7-47a1-b918-cef3025704e3 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: dbeeb200-70b9-4cb4-b5a4-182389d21918] Updating instance_info_cache with network_info: [{"id": "86553395-da58-4c44-b9f4-c67db304d3fe", "address": "fa:16:3e:a4:1b:84", "network": {"id": "6fbccc17-6674-4cd5-8f1b-aa73570baac7", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-394544811-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1dbdc699741a48af93c52068d88a5357", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "20e3f794-c7a3-4696-9488-ecf34c570ef9", "external-id": "nsx-vlan-transportzone-509", "segmentation_id": 509, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap86553395-da", "ovs_interfaceid": "86553395-da58-4c44-b9f4-c67db304d3fe", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 907.143468] env[61957]: DEBUG nova.scheduler.client.report [None req-a7399dbc-0983-4a91-82d1-9c1b173d4f53 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 907.171026] env[61957]: DEBUG oslo_vmware.api [None req-ca2fb2e2-e99f-422a-be65-983bd81b6e15 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52a9bfce-89aa-a8f6-4e18-f33f8460d400, 'name': SearchDatastore_Task, 'duration_secs': 0.010755} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.171026] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ca2fb2e2-e99f-422a-be65-983bd81b6e15 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Releasing lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 907.171026] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca2fb2e2-e99f-422a-be65-983bd81b6e15 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore1] 736eabcf-70f8-4e1e-9830-270d9971fc9a/736eabcf-70f8-4e1e-9830-270d9971fc9a.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 907.171026] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5b328b8a-7bbb-4ad9-88fc-a6f92abfe775 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.185198] env[61957]: DEBUG oslo_vmware.api [None req-ca2fb2e2-e99f-422a-be65-983bd81b6e15 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Waiting for the task: (returnval){ [ 907.185198] env[61957]: value = "task-1277663" [ 907.185198] env[61957]: _type = "Task" [ 907.185198] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.196196] env[61957]: DEBUG oslo_vmware.api [None req-ca2fb2e2-e99f-422a-be65-983bd81b6e15 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Task: {'id': task-1277663, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.232559] env[61957]: DEBUG nova.objects.base [None req-6c9b1d92-3397-4471-b73b-76c509e782fc tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Object Instance lazy-loaded attributes: flavor,info_cache {{(pid=61957) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 907.289637] env[61957]: DEBUG nova.compute.manager [req-bc702898-b386-41b1-a4ae-d81beb1c2dc5 req-5078b003-9681-417e-97b0-9fa1c212a9b8 service nova] [instance: dbeeb200-70b9-4cb4-b5a4-182389d21918] Received event network-vif-plugged-86553395-da58-4c44-b9f4-c67db304d3fe {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 907.289953] env[61957]: DEBUG oslo_concurrency.lockutils [req-bc702898-b386-41b1-a4ae-d81beb1c2dc5 req-5078b003-9681-417e-97b0-9fa1c212a9b8 service nova] Acquiring lock "dbeeb200-70b9-4cb4-b5a4-182389d21918-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 907.291321] env[61957]: DEBUG oslo_concurrency.lockutils [req-bc702898-b386-41b1-a4ae-d81beb1c2dc5 req-5078b003-9681-417e-97b0-9fa1c212a9b8 service nova] Lock "dbeeb200-70b9-4cb4-b5a4-182389d21918-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 907.293339] env[61957]: DEBUG oslo_concurrency.lockutils [req-bc702898-b386-41b1-a4ae-d81beb1c2dc5 req-5078b003-9681-417e-97b0-9fa1c212a9b8 service nova] Lock "dbeeb200-70b9-4cb4-b5a4-182389d21918-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 907.293659] env[61957]: DEBUG nova.compute.manager [req-bc702898-b386-41b1-a4ae-d81beb1c2dc5 req-5078b003-9681-417e-97b0-9fa1c212a9b8 service nova] [instance: dbeeb200-70b9-4cb4-b5a4-182389d21918] No waiting events found dispatching network-vif-plugged-86553395-da58-4c44-b9f4-c67db304d3fe {{(pid=61957) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 907.293901] env[61957]: WARNING nova.compute.manager [req-bc702898-b386-41b1-a4ae-d81beb1c2dc5 req-5078b003-9681-417e-97b0-9fa1c212a9b8 service nova] [instance: dbeeb200-70b9-4cb4-b5a4-182389d21918] Received unexpected event network-vif-plugged-86553395-da58-4c44-b9f4-c67db304d3fe for instance with vm_state building and task_state spawning. [ 907.294165] env[61957]: DEBUG nova.compute.manager [req-bc702898-b386-41b1-a4ae-d81beb1c2dc5 req-5078b003-9681-417e-97b0-9fa1c212a9b8 service nova] [instance: dbeeb200-70b9-4cb4-b5a4-182389d21918] Received event network-changed-86553395-da58-4c44-b9f4-c67db304d3fe {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 907.294387] env[61957]: DEBUG nova.compute.manager [req-bc702898-b386-41b1-a4ae-d81beb1c2dc5 req-5078b003-9681-417e-97b0-9fa1c212a9b8 service nova] [instance: dbeeb200-70b9-4cb4-b5a4-182389d21918] Refreshing instance network info cache due to event network-changed-86553395-da58-4c44-b9f4-c67db304d3fe. {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 907.294712] env[61957]: DEBUG oslo_concurrency.lockutils [req-bc702898-b386-41b1-a4ae-d81beb1c2dc5 req-5078b003-9681-417e-97b0-9fa1c212a9b8 service nova] Acquiring lock "refresh_cache-dbeeb200-70b9-4cb4-b5a4-182389d21918" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 907.474322] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277662, 'name': CreateVM_Task, 'duration_secs': 0.288549} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.474532] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a8ca1ad6-d636-4fa2-b0ac-53b020e392ce] Created VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 907.474922] env[61957]: DEBUG oslo_concurrency.lockutils [None req-35c8e096-fa6b-43c8-874d-91a51735ac8e tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 907.475108] env[61957]: DEBUG oslo_concurrency.lockutils [None req-35c8e096-fa6b-43c8-874d-91a51735ac8e tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Acquired lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 907.475478] env[61957]: DEBUG oslo_concurrency.lockutils [None req-35c8e096-fa6b-43c8-874d-91a51735ac8e tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 907.475782] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7a57e2cf-8a8d-4cac-9505-bb770c5052f6 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.482061] env[61957]: DEBUG oslo_vmware.api [None req-35c8e096-fa6b-43c8-874d-91a51735ac8e tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Waiting for the task: (returnval){ [ 907.482061] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]5271f692-04c4-03de-6ca3-61c8119f4f5b" [ 907.482061] env[61957]: _type = "Task" [ 907.482061] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.493601] env[61957]: DEBUG oslo_vmware.api [None req-35c8e096-fa6b-43c8-874d-91a51735ac8e tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]5271f692-04c4-03de-6ca3-61c8119f4f5b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.578290] env[61957]: DEBUG oslo_concurrency.lockutils [None req-3c689a75-14e7-47a1-b918-cef3025704e3 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Releasing lock "refresh_cache-dbeeb200-70b9-4cb4-b5a4-182389d21918" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 907.578290] env[61957]: DEBUG nova.compute.manager [None req-3c689a75-14e7-47a1-b918-cef3025704e3 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: dbeeb200-70b9-4cb4-b5a4-182389d21918] Instance network_info: |[{"id": "86553395-da58-4c44-b9f4-c67db304d3fe", "address": "fa:16:3e:a4:1b:84", "network": {"id": "6fbccc17-6674-4cd5-8f1b-aa73570baac7", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-394544811-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1dbdc699741a48af93c52068d88a5357", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "20e3f794-c7a3-4696-9488-ecf34c570ef9", "external-id": "nsx-vlan-transportzone-509", "segmentation_id": 509, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap86553395-da", "ovs_interfaceid": "86553395-da58-4c44-b9f4-c67db304d3fe", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 907.578290] env[61957]: DEBUG oslo_concurrency.lockutils [req-bc702898-b386-41b1-a4ae-d81beb1c2dc5 req-5078b003-9681-417e-97b0-9fa1c212a9b8 service nova] Acquired lock "refresh_cache-dbeeb200-70b9-4cb4-b5a4-182389d21918" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 907.578762] env[61957]: DEBUG nova.network.neutron [req-bc702898-b386-41b1-a4ae-d81beb1c2dc5 req-5078b003-9681-417e-97b0-9fa1c212a9b8 service nova] [instance: dbeeb200-70b9-4cb4-b5a4-182389d21918] Refreshing network info cache for port 86553395-da58-4c44-b9f4-c67db304d3fe {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 907.579735] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-3c689a75-14e7-47a1-b918-cef3025704e3 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: dbeeb200-70b9-4cb4-b5a4-182389d21918] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a4:1b:84', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '20e3f794-c7a3-4696-9488-ecf34c570ef9', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '86553395-da58-4c44-b9f4-c67db304d3fe', 'vif_model': 'vmxnet3'}] {{(pid=61957) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 907.587730] env[61957]: DEBUG oslo.service.loopingcall [None req-3c689a75-14e7-47a1-b918-cef3025704e3 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 907.588923] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: dbeeb200-70b9-4cb4-b5a4-182389d21918] Creating VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 907.589196] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6e5b8070-8fd0-4cda-939c-0250bf885b27 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.616249] env[61957]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 907.616249] env[61957]: value = "task-1277664" [ 907.616249] env[61957]: _type = "Task" [ 907.616249] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.627800] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277664, 'name': CreateVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.648320] env[61957]: DEBUG oslo_concurrency.lockutils [None req-a7399dbc-0983-4a91-82d1-9c1b173d4f53 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.435s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 907.648982] env[61957]: DEBUG nova.compute.manager [None req-a7399dbc-0983-4a91-82d1-9c1b173d4f53 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 39b8b1df-efdd-4c22-9f31-85c85be4f0eb] Start building networks asynchronously for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 907.651862] env[61957]: DEBUG oslo_concurrency.lockutils [None req-bb69c8c2-41e0-4fa7-aea3-e810870c1532 tempest-ServersNegativeTestMultiTenantJSON-2135399494 tempest-ServersNegativeTestMultiTenantJSON-2135399494-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.450s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 907.654092] env[61957]: INFO nova.compute.claims [None req-bb69c8c2-41e0-4fa7-aea3-e810870c1532 tempest-ServersNegativeTestMultiTenantJSON-2135399494 tempest-ServersNegativeTestMultiTenantJSON-2135399494-project-member] [instance: 19966b0a-53b7-48c5-849c-a9d00dc024f8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 907.703812] env[61957]: DEBUG oslo_vmware.api [None req-ca2fb2e2-e99f-422a-be65-983bd81b6e15 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Task: {'id': task-1277663, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.45896} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.704053] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca2fb2e2-e99f-422a-be65-983bd81b6e15 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore1] 736eabcf-70f8-4e1e-9830-270d9971fc9a/736eabcf-70f8-4e1e-9830-270d9971fc9a.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 907.704291] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-ca2fb2e2-e99f-422a-be65-983bd81b6e15 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] [instance: 736eabcf-70f8-4e1e-9830-270d9971fc9a] Extending root virtual disk to 1048576 {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 907.704546] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0a2221e7-df55-4f84-be5f-bff888ad0f65 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.712765] env[61957]: DEBUG oslo_vmware.api [None req-ca2fb2e2-e99f-422a-be65-983bd81b6e15 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Waiting for the task: (returnval){ [ 907.712765] env[61957]: value = "task-1277665" [ 907.712765] env[61957]: _type = "Task" [ 907.712765] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.721782] env[61957]: DEBUG oslo_vmware.api [None req-ca2fb2e2-e99f-422a-be65-983bd81b6e15 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Task: {'id': task-1277665, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.977648] env[61957]: DEBUG nova.network.neutron [None req-6c9b1d92-3397-4471-b73b-76c509e782fc tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] [instance: e40fa112-4648-428f-a403-b3e3b8319ea8] Updating instance_info_cache with network_info: [{"id": "c407d755-4caa-44d0-b9b2-8e59106cc23e", "address": "fa:16:3e:db:0f:8c", "network": {"id": "41a1701c-d9b6-4414-9e92-2e999cc916e3", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-727135203-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d849d922cfad41848a33e8d3bc83ea0f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f1b507ed-cd2d-4c09-9d96-c47bde6a7774", "external-id": "nsx-vlan-transportzone-980", "segmentation_id": 980, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc407d755-4c", "ovs_interfaceid": "c407d755-4caa-44d0-b9b2-8e59106cc23e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 907.993082] env[61957]: DEBUG oslo_vmware.api [None req-35c8e096-fa6b-43c8-874d-91a51735ac8e tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]5271f692-04c4-03de-6ca3-61c8119f4f5b, 'name': SearchDatastore_Task, 'duration_secs': 0.05671} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.993395] env[61957]: DEBUG oslo_concurrency.lockutils [None req-35c8e096-fa6b-43c8-874d-91a51735ac8e tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Releasing lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 907.993632] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-35c8e096-fa6b-43c8-874d-91a51735ac8e tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] [instance: a8ca1ad6-d636-4fa2-b0ac-53b020e392ce] Processing image 11c76a2c-f705-470a-ba9d-4657858bab38 {{(pid=61957) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 907.993867] env[61957]: DEBUG oslo_concurrency.lockutils [None req-35c8e096-fa6b-43c8-874d-91a51735ac8e tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 907.994059] env[61957]: DEBUG oslo_concurrency.lockutils [None req-35c8e096-fa6b-43c8-874d-91a51735ac8e tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Acquired lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 907.994326] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-35c8e096-fa6b-43c8-874d-91a51735ac8e tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 907.994463] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a09bdd6d-64fb-4600-9196-055cb4b7f3a9 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.003282] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-35c8e096-fa6b-43c8-874d-91a51735ac8e tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 908.003467] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-35c8e096-fa6b-43c8-874d-91a51735ac8e tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61957) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 908.004225] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f888947f-af5f-4774-a6c4-98332a7065de {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.009675] env[61957]: DEBUG oslo_vmware.api [None req-35c8e096-fa6b-43c8-874d-91a51735ac8e tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Waiting for the task: (returnval){ [ 908.009675] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52cf30cc-6c1f-0617-8997-c64f9e751a74" [ 908.009675] env[61957]: _type = "Task" [ 908.009675] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.017706] env[61957]: DEBUG oslo_vmware.api [None req-35c8e096-fa6b-43c8-874d-91a51735ac8e tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52cf30cc-6c1f-0617-8997-c64f9e751a74, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.126659] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277664, 'name': CreateVM_Task, 'duration_secs': 0.388795} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.126829] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: dbeeb200-70b9-4cb4-b5a4-182389d21918] Created VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 908.127655] env[61957]: DEBUG oslo_concurrency.lockutils [None req-3c689a75-14e7-47a1-b918-cef3025704e3 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 908.127655] env[61957]: DEBUG oslo_concurrency.lockutils [None req-3c689a75-14e7-47a1-b918-cef3025704e3 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Acquired lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 908.128819] env[61957]: DEBUG oslo_concurrency.lockutils [None req-3c689a75-14e7-47a1-b918-cef3025704e3 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 908.128819] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5cf045f5-1347-486d-9331-a463b62f996e {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.134011] env[61957]: DEBUG oslo_vmware.api [None req-3c689a75-14e7-47a1-b918-cef3025704e3 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Waiting for the task: (returnval){ [ 908.134011] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]528fa63f-4437-4f59-cc1f-9f1ed4ac75c9" [ 908.134011] env[61957]: _type = "Task" [ 908.134011] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.142616] env[61957]: DEBUG oslo_vmware.api [None req-3c689a75-14e7-47a1-b918-cef3025704e3 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]528fa63f-4437-4f59-cc1f-9f1ed4ac75c9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.158166] env[61957]: DEBUG nova.compute.utils [None req-a7399dbc-0983-4a91-82d1-9c1b173d4f53 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Using /dev/sd instead of None {{(pid=61957) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 908.164025] env[61957]: DEBUG nova.compute.manager [None req-a7399dbc-0983-4a91-82d1-9c1b173d4f53 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 39b8b1df-efdd-4c22-9f31-85c85be4f0eb] Allocating IP information in the background. {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 908.164025] env[61957]: DEBUG nova.network.neutron [None req-a7399dbc-0983-4a91-82d1-9c1b173d4f53 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 39b8b1df-efdd-4c22-9f31-85c85be4f0eb] allocate_for_instance() {{(pid=61957) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 908.222969] env[61957]: DEBUG oslo_vmware.api [None req-ca2fb2e2-e99f-422a-be65-983bd81b6e15 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Task: {'id': task-1277665, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.081693} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.225491] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-ca2fb2e2-e99f-422a-be65-983bd81b6e15 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] [instance: 736eabcf-70f8-4e1e-9830-270d9971fc9a] Extended root virtual disk {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 908.226363] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f91f4d5e-e49b-4546-9629-43701331e835 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.249874] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-ca2fb2e2-e99f-422a-be65-983bd81b6e15 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] [instance: 736eabcf-70f8-4e1e-9830-270d9971fc9a] Reconfiguring VM instance instance-0000004a to attach disk [datastore1] 736eabcf-70f8-4e1e-9830-270d9971fc9a/736eabcf-70f8-4e1e-9830-270d9971fc9a.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 908.252046] env[61957]: DEBUG nova.policy [None req-a7399dbc-0983-4a91-82d1-9c1b173d4f53 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '85d4f8bb3aa64e7285aac42d14087ff1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e16f6dce3f0e44fb96de516f17d4c6f5', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61957) authorize /opt/stack/nova/nova/policy.py:203}} [ 908.253141] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b6f8fb40-4e07-45e1-bde6-e22e6940d6b8 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.280144] env[61957]: DEBUG oslo_vmware.api [None req-ca2fb2e2-e99f-422a-be65-983bd81b6e15 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Waiting for the task: (returnval){ [ 908.280144] env[61957]: value = "task-1277666" [ 908.280144] env[61957]: _type = "Task" [ 908.280144] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.289971] env[61957]: DEBUG oslo_vmware.api [None req-ca2fb2e2-e99f-422a-be65-983bd81b6e15 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Task: {'id': task-1277666, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.412702] env[61957]: DEBUG nova.network.neutron [req-bc702898-b386-41b1-a4ae-d81beb1c2dc5 req-5078b003-9681-417e-97b0-9fa1c212a9b8 service nova] [instance: dbeeb200-70b9-4cb4-b5a4-182389d21918] Updated VIF entry in instance network info cache for port 86553395-da58-4c44-b9f4-c67db304d3fe. {{(pid=61957) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 908.413159] env[61957]: DEBUG nova.network.neutron [req-bc702898-b386-41b1-a4ae-d81beb1c2dc5 req-5078b003-9681-417e-97b0-9fa1c212a9b8 service nova] [instance: dbeeb200-70b9-4cb4-b5a4-182389d21918] Updating instance_info_cache with network_info: [{"id": "86553395-da58-4c44-b9f4-c67db304d3fe", "address": "fa:16:3e:a4:1b:84", "network": {"id": "6fbccc17-6674-4cd5-8f1b-aa73570baac7", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-394544811-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1dbdc699741a48af93c52068d88a5357", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "20e3f794-c7a3-4696-9488-ecf34c570ef9", "external-id": "nsx-vlan-transportzone-509", "segmentation_id": 509, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap86553395-da", "ovs_interfaceid": "86553395-da58-4c44-b9f4-c67db304d3fe", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 908.481287] env[61957]: DEBUG oslo_concurrency.lockutils [None req-6c9b1d92-3397-4471-b73b-76c509e782fc tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Releasing lock "refresh_cache-e40fa112-4648-428f-a403-b3e3b8319ea8" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 908.523031] env[61957]: DEBUG oslo_vmware.api [None req-35c8e096-fa6b-43c8-874d-91a51735ac8e tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52cf30cc-6c1f-0617-8997-c64f9e751a74, 'name': SearchDatastore_Task, 'duration_secs': 0.031534} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.524034] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0f53a15b-f075-47dd-bf59-49d851f6e0d1 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.530708] env[61957]: DEBUG oslo_vmware.api [None req-35c8e096-fa6b-43c8-874d-91a51735ac8e tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Waiting for the task: (returnval){ [ 908.530708] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]521fcb1c-26f5-003f-3904-9899ba4e244f" [ 908.530708] env[61957]: _type = "Task" [ 908.530708] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.542040] env[61957]: DEBUG oslo_vmware.api [None req-35c8e096-fa6b-43c8-874d-91a51735ac8e tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]521fcb1c-26f5-003f-3904-9899ba4e244f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.598309] env[61957]: DEBUG nova.network.neutron [None req-a7399dbc-0983-4a91-82d1-9c1b173d4f53 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 39b8b1df-efdd-4c22-9f31-85c85be4f0eb] Successfully created port: 7e47ddcd-8c57-46be-8e68-62cec87e391b {{(pid=61957) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 908.647895] env[61957]: DEBUG oslo_vmware.api [None req-3c689a75-14e7-47a1-b918-cef3025704e3 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]528fa63f-4437-4f59-cc1f-9f1ed4ac75c9, 'name': SearchDatastore_Task, 'duration_secs': 0.025944} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.648272] env[61957]: DEBUG oslo_concurrency.lockutils [None req-3c689a75-14e7-47a1-b918-cef3025704e3 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Releasing lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 908.648558] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-3c689a75-14e7-47a1-b918-cef3025704e3 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: dbeeb200-70b9-4cb4-b5a4-182389d21918] Processing image 11c76a2c-f705-470a-ba9d-4657858bab38 {{(pid=61957) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 908.648788] env[61957]: DEBUG oslo_concurrency.lockutils [None req-3c689a75-14e7-47a1-b918-cef3025704e3 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 908.666703] env[61957]: DEBUG nova.compute.manager [None req-a7399dbc-0983-4a91-82d1-9c1b173d4f53 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 39b8b1df-efdd-4c22-9f31-85c85be4f0eb] Start building block device mappings for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 908.791411] env[61957]: DEBUG oslo_vmware.api [None req-ca2fb2e2-e99f-422a-be65-983bd81b6e15 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Task: {'id': task-1277666, 'name': ReconfigVM_Task, 'duration_secs': 0.314784} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.791753] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-ca2fb2e2-e99f-422a-be65-983bd81b6e15 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] [instance: 736eabcf-70f8-4e1e-9830-270d9971fc9a] Reconfigured VM instance instance-0000004a to attach disk [datastore1] 736eabcf-70f8-4e1e-9830-270d9971fc9a/736eabcf-70f8-4e1e-9830-270d9971fc9a.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 908.792463] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-45543e97-25a8-420b-8013-83d50baff0be {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.803518] env[61957]: DEBUG oslo_vmware.api [None req-ca2fb2e2-e99f-422a-be65-983bd81b6e15 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Waiting for the task: (returnval){ [ 908.803518] env[61957]: value = "task-1277668" [ 908.803518] env[61957]: _type = "Task" [ 908.803518] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.816156] env[61957]: DEBUG oslo_vmware.api [None req-ca2fb2e2-e99f-422a-be65-983bd81b6e15 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Task: {'id': task-1277668, 'name': Rename_Task} progress is 5%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.916367] env[61957]: DEBUG oslo_concurrency.lockutils [req-bc702898-b386-41b1-a4ae-d81beb1c2dc5 req-5078b003-9681-417e-97b0-9fa1c212a9b8 service nova] Releasing lock "refresh_cache-dbeeb200-70b9-4cb4-b5a4-182389d21918" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 908.987779] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e76870f2-3be0-4f8b-ba6f-faeeda07e841 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.991202] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c9b1d92-3397-4471-b73b-76c509e782fc tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] [instance: e40fa112-4648-428f-a403-b3e3b8319ea8] Powering on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 908.991458] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c831b4ad-8a53-4334-abdc-aab52626271e {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.999895] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdc5e1d8-26fb-4689-a4d8-60bf8294f344 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.002957] env[61957]: DEBUG oslo_vmware.api [None req-6c9b1d92-3397-4471-b73b-76c509e782fc tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Waiting for the task: (returnval){ [ 909.002957] env[61957]: value = "task-1277669" [ 909.002957] env[61957]: _type = "Task" [ 909.002957] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.038326] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-388ce5cc-374a-4d02-bb8c-b2cfa6d819f5 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.040862] env[61957]: DEBUG oslo_vmware.api [None req-6c9b1d92-3397-4471-b73b-76c509e782fc tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Task: {'id': task-1277669, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.047057] env[61957]: DEBUG oslo_vmware.api [None req-35c8e096-fa6b-43c8-874d-91a51735ac8e tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]521fcb1c-26f5-003f-3904-9899ba4e244f, 'name': SearchDatastore_Task, 'duration_secs': 0.011041} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.049156] env[61957]: DEBUG oslo_concurrency.lockutils [None req-35c8e096-fa6b-43c8-874d-91a51735ac8e tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Releasing lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 909.049423] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-35c8e096-fa6b-43c8-874d-91a51735ac8e tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore1] a8ca1ad6-d636-4fa2-b0ac-53b020e392ce/a8ca1ad6-d636-4fa2-b0ac-53b020e392ce.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 909.049749] env[61957]: DEBUG oslo_concurrency.lockutils [None req-3c689a75-14e7-47a1-b918-cef3025704e3 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Acquired lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 909.049966] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-3c689a75-14e7-47a1-b918-cef3025704e3 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 909.050202] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e3a93ca0-67c5-436c-ae01-014ee1af25ab {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.052999] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80323799-887a-4059-868c-fab13545f13b {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.056730] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f80456f5-e780-41fb-a6c4-f6780d2bb921 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.071645] env[61957]: DEBUG oslo_vmware.api [None req-35c8e096-fa6b-43c8-874d-91a51735ac8e tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Waiting for the task: (returnval){ [ 909.071645] env[61957]: value = "task-1277670" [ 909.071645] env[61957]: _type = "Task" [ 909.071645] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.072149] env[61957]: DEBUG nova.compute.provider_tree [None req-bb69c8c2-41e0-4fa7-aea3-e810870c1532 tempest-ServersNegativeTestMultiTenantJSON-2135399494 tempest-ServersNegativeTestMultiTenantJSON-2135399494-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 909.075265] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-3c689a75-14e7-47a1-b918-cef3025704e3 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 909.075454] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-3c689a75-14e7-47a1-b918-cef3025704e3 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61957) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 909.078930] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4c9c9b2f-d2d5-4289-92bb-000b4eaf8d65 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.086104] env[61957]: DEBUG oslo_vmware.api [None req-3c689a75-14e7-47a1-b918-cef3025704e3 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Waiting for the task: (returnval){ [ 909.086104] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]527faed3-a3be-50e4-7485-eb866c81712d" [ 909.086104] env[61957]: _type = "Task" [ 909.086104] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.089368] env[61957]: DEBUG oslo_vmware.api [None req-35c8e096-fa6b-43c8-874d-91a51735ac8e tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Task: {'id': task-1277670, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.098085] env[61957]: DEBUG oslo_vmware.api [None req-3c689a75-14e7-47a1-b918-cef3025704e3 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]527faed3-a3be-50e4-7485-eb866c81712d, 'name': SearchDatastore_Task, 'duration_secs': 0.009322} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.098951] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8f55abc5-6e14-4d30-9e8a-2937b11e282f {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.104144] env[61957]: DEBUG oslo_vmware.api [None req-3c689a75-14e7-47a1-b918-cef3025704e3 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Waiting for the task: (returnval){ [ 909.104144] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]5264d923-0bcf-dc97-7756-17d37b39d9cb" [ 909.104144] env[61957]: _type = "Task" [ 909.104144] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.112594] env[61957]: DEBUG oslo_vmware.api [None req-3c689a75-14e7-47a1-b918-cef3025704e3 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]5264d923-0bcf-dc97-7756-17d37b39d9cb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.314764] env[61957]: DEBUG oslo_vmware.api [None req-ca2fb2e2-e99f-422a-be65-983bd81b6e15 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Task: {'id': task-1277668, 'name': Rename_Task, 'duration_secs': 0.156016} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.315724] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca2fb2e2-e99f-422a-be65-983bd81b6e15 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] [instance: 736eabcf-70f8-4e1e-9830-270d9971fc9a] Powering on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 909.316206] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8e94ce0d-6a67-4132-a696-f2f89f2455ec {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.327331] env[61957]: DEBUG oslo_vmware.api [None req-ca2fb2e2-e99f-422a-be65-983bd81b6e15 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Waiting for the task: (returnval){ [ 909.327331] env[61957]: value = "task-1277671" [ 909.327331] env[61957]: _type = "Task" [ 909.327331] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.335448] env[61957]: DEBUG oslo_vmware.api [None req-ca2fb2e2-e99f-422a-be65-983bd81b6e15 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Task: {'id': task-1277671, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.516355] env[61957]: DEBUG oslo_vmware.api [None req-6c9b1d92-3397-4471-b73b-76c509e782fc tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Task: {'id': task-1277669, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.576407] env[61957]: DEBUG nova.scheduler.client.report [None req-bb69c8c2-41e0-4fa7-aea3-e810870c1532 tempest-ServersNegativeTestMultiTenantJSON-2135399494 tempest-ServersNegativeTestMultiTenantJSON-2135399494-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 909.590925] env[61957]: DEBUG oslo_vmware.api [None req-35c8e096-fa6b-43c8-874d-91a51735ac8e tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Task: {'id': task-1277670, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.615744] env[61957]: DEBUG oslo_vmware.api [None req-3c689a75-14e7-47a1-b918-cef3025704e3 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]5264d923-0bcf-dc97-7756-17d37b39d9cb, 'name': SearchDatastore_Task, 'duration_secs': 0.014323} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.616145] env[61957]: DEBUG oslo_concurrency.lockutils [None req-3c689a75-14e7-47a1-b918-cef3025704e3 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Releasing lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 909.616438] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c689a75-14e7-47a1-b918-cef3025704e3 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore1] dbeeb200-70b9-4cb4-b5a4-182389d21918/dbeeb200-70b9-4cb4-b5a4-182389d21918.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 909.616724] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-63480582-9cb9-403d-9856-95095cb7c596 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.626946] env[61957]: DEBUG oslo_vmware.api [None req-3c689a75-14e7-47a1-b918-cef3025704e3 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Waiting for the task: (returnval){ [ 909.626946] env[61957]: value = "task-1277672" [ 909.626946] env[61957]: _type = "Task" [ 909.626946] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.638278] env[61957]: DEBUG oslo_vmware.api [None req-3c689a75-14e7-47a1-b918-cef3025704e3 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': task-1277672, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.679555] env[61957]: DEBUG nova.compute.manager [None req-a7399dbc-0983-4a91-82d1-9c1b173d4f53 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 39b8b1df-efdd-4c22-9f31-85c85be4f0eb] Start spawning the instance on the hypervisor. {{(pid=61957) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 909.703197] env[61957]: DEBUG nova.virt.hardware [None req-a7399dbc-0983-4a91-82d1-9c1b173d4f53 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T17:22:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T17:21:50Z,direct_url=,disk_format='vmdk',id=11c76a2c-f705-470a-ba9d-4657858bab38,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='81d3096d0f094373913fc3dc8f5c3c6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T17:21:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 909.703525] env[61957]: DEBUG nova.virt.hardware [None req-a7399dbc-0983-4a91-82d1-9c1b173d4f53 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Flavor limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 909.703722] env[61957]: DEBUG nova.virt.hardware [None req-a7399dbc-0983-4a91-82d1-9c1b173d4f53 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Image limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 909.703916] env[61957]: DEBUG nova.virt.hardware [None req-a7399dbc-0983-4a91-82d1-9c1b173d4f53 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Flavor pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 909.704107] env[61957]: DEBUG nova.virt.hardware [None req-a7399dbc-0983-4a91-82d1-9c1b173d4f53 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Image pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 909.704266] env[61957]: DEBUG nova.virt.hardware [None req-a7399dbc-0983-4a91-82d1-9c1b173d4f53 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 909.704495] env[61957]: DEBUG nova.virt.hardware [None req-a7399dbc-0983-4a91-82d1-9c1b173d4f53 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 909.704664] env[61957]: DEBUG nova.virt.hardware [None req-a7399dbc-0983-4a91-82d1-9c1b173d4f53 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 909.704865] env[61957]: DEBUG nova.virt.hardware [None req-a7399dbc-0983-4a91-82d1-9c1b173d4f53 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Got 1 possible topologies {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 909.705103] env[61957]: DEBUG nova.virt.hardware [None req-a7399dbc-0983-4a91-82d1-9c1b173d4f53 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 909.705290] env[61957]: DEBUG nova.virt.hardware [None req-a7399dbc-0983-4a91-82d1-9c1b173d4f53 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 909.706182] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0cbf50a-011f-4ade-80f4-6482b91b07b6 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.715366] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd9ef7ac-5f74-4365-9982-e00a732125d8 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.836066] env[61957]: DEBUG oslo_vmware.api [None req-ca2fb2e2-e99f-422a-be65-983bd81b6e15 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Task: {'id': task-1277671, 'name': PowerOnVM_Task} progress is 99%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.006395] env[61957]: DEBUG nova.compute.manager [req-0f4a3dad-c813-44b8-8671-09dd613ce202 req-51910c67-8c39-4333-bf8a-0450479d9d0d service nova] [instance: 39b8b1df-efdd-4c22-9f31-85c85be4f0eb] Received event network-vif-plugged-7e47ddcd-8c57-46be-8e68-62cec87e391b {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 910.006395] env[61957]: DEBUG oslo_concurrency.lockutils [req-0f4a3dad-c813-44b8-8671-09dd613ce202 req-51910c67-8c39-4333-bf8a-0450479d9d0d service nova] Acquiring lock "39b8b1df-efdd-4c22-9f31-85c85be4f0eb-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 910.006395] env[61957]: DEBUG oslo_concurrency.lockutils [req-0f4a3dad-c813-44b8-8671-09dd613ce202 req-51910c67-8c39-4333-bf8a-0450479d9d0d service nova] Lock "39b8b1df-efdd-4c22-9f31-85c85be4f0eb-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 910.006395] env[61957]: DEBUG oslo_concurrency.lockutils [req-0f4a3dad-c813-44b8-8671-09dd613ce202 req-51910c67-8c39-4333-bf8a-0450479d9d0d service nova] Lock "39b8b1df-efdd-4c22-9f31-85c85be4f0eb-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 910.006395] env[61957]: DEBUG nova.compute.manager [req-0f4a3dad-c813-44b8-8671-09dd613ce202 req-51910c67-8c39-4333-bf8a-0450479d9d0d service nova] [instance: 39b8b1df-efdd-4c22-9f31-85c85be4f0eb] No waiting events found dispatching network-vif-plugged-7e47ddcd-8c57-46be-8e68-62cec87e391b {{(pid=61957) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 910.006395] env[61957]: WARNING nova.compute.manager [req-0f4a3dad-c813-44b8-8671-09dd613ce202 req-51910c67-8c39-4333-bf8a-0450479d9d0d service nova] [instance: 39b8b1df-efdd-4c22-9f31-85c85be4f0eb] Received unexpected event network-vif-plugged-7e47ddcd-8c57-46be-8e68-62cec87e391b for instance with vm_state building and task_state spawning. [ 910.016388] env[61957]: DEBUG oslo_vmware.api [None req-6c9b1d92-3397-4471-b73b-76c509e782fc tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Task: {'id': task-1277669, 'name': PowerOnVM_Task, 'duration_secs': 0.524695} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.016643] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c9b1d92-3397-4471-b73b-76c509e782fc tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] [instance: e40fa112-4648-428f-a403-b3e3b8319ea8] Powered on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 910.016861] env[61957]: DEBUG nova.compute.manager [None req-6c9b1d92-3397-4471-b73b-76c509e782fc tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] [instance: e40fa112-4648-428f-a403-b3e3b8319ea8] Checking state {{(pid=61957) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 910.017620] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e760f849-c67d-4997-9285-98d74329a0f9 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.081688] env[61957]: DEBUG oslo_concurrency.lockutils [None req-bb69c8c2-41e0-4fa7-aea3-e810870c1532 tempest-ServersNegativeTestMultiTenantJSON-2135399494 tempest-ServersNegativeTestMultiTenantJSON-2135399494-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.430s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 910.082421] env[61957]: DEBUG nova.compute.manager [None req-bb69c8c2-41e0-4fa7-aea3-e810870c1532 tempest-ServersNegativeTestMultiTenantJSON-2135399494 tempest-ServersNegativeTestMultiTenantJSON-2135399494-project-member] [instance: 19966b0a-53b7-48c5-849c-a9d00dc024f8] Start building networks asynchronously for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 910.085732] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ba6997f6-9064-4c0a-859c-de0e4f8ea2ee tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 27.650s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 910.085957] env[61957]: DEBUG nova.objects.instance [None req-ba6997f6-9064-4c0a-859c-de0e4f8ea2ee tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Lazy-loading 'resources' on Instance uuid 0adae8e1-8c2f-4110-805b-1f286debc833 {{(pid=61957) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 910.097815] env[61957]: DEBUG oslo_vmware.api [None req-35c8e096-fa6b-43c8-874d-91a51735ac8e tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Task: {'id': task-1277670, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.577838} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.098197] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-35c8e096-fa6b-43c8-874d-91a51735ac8e tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore1] a8ca1ad6-d636-4fa2-b0ac-53b020e392ce/a8ca1ad6-d636-4fa2-b0ac-53b020e392ce.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 910.098364] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-35c8e096-fa6b-43c8-874d-91a51735ac8e tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] [instance: a8ca1ad6-d636-4fa2-b0ac-53b020e392ce] Extending root virtual disk to 1048576 {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 910.098639] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6cb9bc1e-4a80-49cc-b3cc-43d5dc7913d9 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.108823] env[61957]: DEBUG oslo_vmware.api [None req-35c8e096-fa6b-43c8-874d-91a51735ac8e tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Waiting for the task: (returnval){ [ 910.108823] env[61957]: value = "task-1277673" [ 910.108823] env[61957]: _type = "Task" [ 910.108823] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.109629] env[61957]: DEBUG nova.network.neutron [None req-a7399dbc-0983-4a91-82d1-9c1b173d4f53 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 39b8b1df-efdd-4c22-9f31-85c85be4f0eb] Successfully updated port: 7e47ddcd-8c57-46be-8e68-62cec87e391b {{(pid=61957) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 910.123499] env[61957]: DEBUG oslo_vmware.api [None req-35c8e096-fa6b-43c8-874d-91a51735ac8e tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Task: {'id': task-1277673, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.138279] env[61957]: DEBUG oslo_vmware.api [None req-3c689a75-14e7-47a1-b918-cef3025704e3 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': task-1277672, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.335628] env[61957]: DEBUG oslo_vmware.api [None req-ca2fb2e2-e99f-422a-be65-983bd81b6e15 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Task: {'id': task-1277671, 'name': PowerOnVM_Task, 'duration_secs': 0.59158} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.335969] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca2fb2e2-e99f-422a-be65-983bd81b6e15 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] [instance: 736eabcf-70f8-4e1e-9830-270d9971fc9a] Powered on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 910.336065] env[61957]: INFO nova.compute.manager [None req-ca2fb2e2-e99f-422a-be65-983bd81b6e15 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] [instance: 736eabcf-70f8-4e1e-9830-270d9971fc9a] Took 7.43 seconds to spawn the instance on the hypervisor. [ 910.336256] env[61957]: DEBUG nova.compute.manager [None req-ca2fb2e2-e99f-422a-be65-983bd81b6e15 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] [instance: 736eabcf-70f8-4e1e-9830-270d9971fc9a] Checking state {{(pid=61957) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 910.337063] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ba4f934-6abd-4d4b-b0b8-f49859d50703 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.598069] env[61957]: DEBUG nova.compute.utils [None req-bb69c8c2-41e0-4fa7-aea3-e810870c1532 tempest-ServersNegativeTestMultiTenantJSON-2135399494 tempest-ServersNegativeTestMultiTenantJSON-2135399494-project-member] Using /dev/sd instead of None {{(pid=61957) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 910.599931] env[61957]: DEBUG nova.compute.manager [None req-bb69c8c2-41e0-4fa7-aea3-e810870c1532 tempest-ServersNegativeTestMultiTenantJSON-2135399494 tempest-ServersNegativeTestMultiTenantJSON-2135399494-project-member] [instance: 19966b0a-53b7-48c5-849c-a9d00dc024f8] Allocating IP information in the background. {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 910.600419] env[61957]: DEBUG nova.network.neutron [None req-bb69c8c2-41e0-4fa7-aea3-e810870c1532 tempest-ServersNegativeTestMultiTenantJSON-2135399494 tempest-ServersNegativeTestMultiTenantJSON-2135399494-project-member] [instance: 19966b0a-53b7-48c5-849c-a9d00dc024f8] allocate_for_instance() {{(pid=61957) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 910.617761] env[61957]: DEBUG oslo_concurrency.lockutils [None req-a7399dbc-0983-4a91-82d1-9c1b173d4f53 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Acquiring lock "refresh_cache-39b8b1df-efdd-4c22-9f31-85c85be4f0eb" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 910.617761] env[61957]: DEBUG oslo_concurrency.lockutils [None req-a7399dbc-0983-4a91-82d1-9c1b173d4f53 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Acquired lock "refresh_cache-39b8b1df-efdd-4c22-9f31-85c85be4f0eb" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 910.617761] env[61957]: DEBUG nova.network.neutron [None req-a7399dbc-0983-4a91-82d1-9c1b173d4f53 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 39b8b1df-efdd-4c22-9f31-85c85be4f0eb] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 910.625930] env[61957]: DEBUG oslo_vmware.api [None req-35c8e096-fa6b-43c8-874d-91a51735ac8e tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Task: {'id': task-1277673, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072752} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.626290] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-35c8e096-fa6b-43c8-874d-91a51735ac8e tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] [instance: a8ca1ad6-d636-4fa2-b0ac-53b020e392ce] Extended root virtual disk {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 910.627085] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-983fd713-ebd5-40bf-899f-90c0e5d17d13 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.644073] env[61957]: DEBUG oslo_vmware.api [None req-3c689a75-14e7-47a1-b918-cef3025704e3 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': task-1277672, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.662181] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-35c8e096-fa6b-43c8-874d-91a51735ac8e tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] [instance: a8ca1ad6-d636-4fa2-b0ac-53b020e392ce] Reconfiguring VM instance instance-00000049 to attach disk [datastore1] a8ca1ad6-d636-4fa2-b0ac-53b020e392ce/a8ca1ad6-d636-4fa2-b0ac-53b020e392ce.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 910.664071] env[61957]: DEBUG nova.policy [None req-bb69c8c2-41e0-4fa7-aea3-e810870c1532 tempest-ServersNegativeTestMultiTenantJSON-2135399494 tempest-ServersNegativeTestMultiTenantJSON-2135399494-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a4332361414e4fbda17136c3126b3ed5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '219263b31ffc43a0904d8e620e48e085', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61957) authorize /opt/stack/nova/nova/policy.py:203}} [ 910.668233] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2f612a50-775c-476d-bc19-d89ac715efea {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.700774] env[61957]: DEBUG oslo_vmware.api [None req-35c8e096-fa6b-43c8-874d-91a51735ac8e tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Waiting for the task: (returnval){ [ 910.700774] env[61957]: value = "task-1277674" [ 910.700774] env[61957]: _type = "Task" [ 910.700774] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.709538] env[61957]: DEBUG oslo_vmware.api [None req-35c8e096-fa6b-43c8-874d-91a51735ac8e tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Task: {'id': task-1277674, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.772832] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-c02d7885-6c3b-4a1f-ab8a-8e683ed8e586 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] [instance: c23141ee-0cbb-4d1b-8390-c3073fe354f1] Volume attach. Driver type: vmdk {{(pid=61957) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 910.773304] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-c02d7885-6c3b-4a1f-ab8a-8e683ed8e586 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] [instance: c23141ee-0cbb-4d1b-8390-c3073fe354f1] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-274545', 'volume_id': '734d094d-c382-4070-87c9-699fa1355b79', 'name': 'volume-734d094d-c382-4070-87c9-699fa1355b79', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'c23141ee-0cbb-4d1b-8390-c3073fe354f1', 'attached_at': '', 'detached_at': '', 'volume_id': '734d094d-c382-4070-87c9-699fa1355b79', 'serial': '734d094d-c382-4070-87c9-699fa1355b79'} {{(pid=61957) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 910.774230] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14eb5438-fa11-4004-ae14-3a46766bc242 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.798329] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f795c9a-56a7-47c7-a60d-d3f091ba14e1 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.826045] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-c02d7885-6c3b-4a1f-ab8a-8e683ed8e586 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] [instance: c23141ee-0cbb-4d1b-8390-c3073fe354f1] Reconfiguring VM instance instance-0000003f to attach disk [datastore1] volume-734d094d-c382-4070-87c9-699fa1355b79/volume-734d094d-c382-4070-87c9-699fa1355b79.vmdk or device None with type thin {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 910.828948] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7e691742-6542-4b3e-9c51-1f09fbf607ee {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.859410] env[61957]: DEBUG oslo_vmware.api [None req-c02d7885-6c3b-4a1f-ab8a-8e683ed8e586 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Waiting for the task: (returnval){ [ 910.859410] env[61957]: value = "task-1277675" [ 910.859410] env[61957]: _type = "Task" [ 910.859410] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.860144] env[61957]: INFO nova.compute.manager [None req-ca2fb2e2-e99f-422a-be65-983bd81b6e15 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] [instance: 736eabcf-70f8-4e1e-9830-270d9971fc9a] Took 40.88 seconds to build instance. [ 910.871525] env[61957]: DEBUG oslo_vmware.api [None req-c02d7885-6c3b-4a1f-ab8a-8e683ed8e586 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Task: {'id': task-1277675, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.021286] env[61957]: DEBUG nova.network.neutron [None req-bb69c8c2-41e0-4fa7-aea3-e810870c1532 tempest-ServersNegativeTestMultiTenantJSON-2135399494 tempest-ServersNegativeTestMultiTenantJSON-2135399494-project-member] [instance: 19966b0a-53b7-48c5-849c-a9d00dc024f8] Successfully created port: 159b1e72-9328-460d-b334-9a3ad5d73b44 {{(pid=61957) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 911.098126] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b878c356-08dc-414d-abc7-5ec864ec45e2 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.104074] env[61957]: DEBUG nova.compute.manager [None req-bb69c8c2-41e0-4fa7-aea3-e810870c1532 tempest-ServersNegativeTestMultiTenantJSON-2135399494 tempest-ServersNegativeTestMultiTenantJSON-2135399494-project-member] [instance: 19966b0a-53b7-48c5-849c-a9d00dc024f8] Start building block device mappings for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 911.113958] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c793522-5fce-44a2-89f8-0ea8fc19fbf7 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.163933] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5445684-00fc-406e-b1b8-9a846b375d4b {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.169714] env[61957]: DEBUG nova.network.neutron [None req-a7399dbc-0983-4a91-82d1-9c1b173d4f53 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 39b8b1df-efdd-4c22-9f31-85c85be4f0eb] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 911.178437] env[61957]: DEBUG oslo_vmware.api [None req-3c689a75-14e7-47a1-b918-cef3025704e3 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': task-1277672, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.118969} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.180330] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2106f985-57f1-46c1-b7b4-574292e64c89 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.186903] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c689a75-14e7-47a1-b918-cef3025704e3 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore1] dbeeb200-70b9-4cb4-b5a4-182389d21918/dbeeb200-70b9-4cb4-b5a4-182389d21918.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 911.187491] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-3c689a75-14e7-47a1-b918-cef3025704e3 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: dbeeb200-70b9-4cb4-b5a4-182389d21918] Extending root virtual disk to 1048576 {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 911.188092] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-de1ce17a-c3ef-490c-8529-ea0d81eb8f7c {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.208176] env[61957]: DEBUG nova.compute.provider_tree [None req-ba6997f6-9064-4c0a-859c-de0e4f8ea2ee tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 911.220757] env[61957]: DEBUG oslo_vmware.api [None req-3c689a75-14e7-47a1-b918-cef3025704e3 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Waiting for the task: (returnval){ [ 911.220757] env[61957]: value = "task-1277676" [ 911.220757] env[61957]: _type = "Task" [ 911.220757] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.222326] env[61957]: DEBUG nova.scheduler.client.report [None req-ba6997f6-9064-4c0a-859c-de0e4f8ea2ee tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 911.234489] env[61957]: DEBUG oslo_vmware.api [None req-35c8e096-fa6b-43c8-874d-91a51735ac8e tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Task: {'id': task-1277674, 'name': ReconfigVM_Task, 'duration_secs': 0.309348} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.237467] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-35c8e096-fa6b-43c8-874d-91a51735ac8e tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] [instance: a8ca1ad6-d636-4fa2-b0ac-53b020e392ce] Reconfigured VM instance instance-00000049 to attach disk [datastore1] a8ca1ad6-d636-4fa2-b0ac-53b020e392ce/a8ca1ad6-d636-4fa2-b0ac-53b020e392ce.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 911.237467] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-06cd03bc-2775-4626-bb01-ec502e9ced75 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.241479] env[61957]: DEBUG oslo_vmware.api [None req-3c689a75-14e7-47a1-b918-cef3025704e3 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': task-1277676, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.248586] env[61957]: DEBUG oslo_vmware.api [None req-35c8e096-fa6b-43c8-874d-91a51735ac8e tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Waiting for the task: (returnval){ [ 911.248586] env[61957]: value = "task-1277677" [ 911.248586] env[61957]: _type = "Task" [ 911.248586] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.258814] env[61957]: DEBUG oslo_vmware.api [None req-35c8e096-fa6b-43c8-874d-91a51735ac8e tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Task: {'id': task-1277677, 'name': Rename_Task} progress is 5%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.363138] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ca2fb2e2-e99f-422a-be65-983bd81b6e15 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Lock "736eabcf-70f8-4e1e-9830-270d9971fc9a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 83.951s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 911.373992] env[61957]: DEBUG oslo_vmware.api [None req-c02d7885-6c3b-4a1f-ab8a-8e683ed8e586 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Task: {'id': task-1277675, 'name': ReconfigVM_Task, 'duration_secs': 0.405355} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.374346] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-c02d7885-6c3b-4a1f-ab8a-8e683ed8e586 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] [instance: c23141ee-0cbb-4d1b-8390-c3073fe354f1] Reconfigured VM instance instance-0000003f to attach disk [datastore1] volume-734d094d-c382-4070-87c9-699fa1355b79/volume-734d094d-c382-4070-87c9-699fa1355b79.vmdk or device None with type thin {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 911.379392] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-af0e35c8-55f6-4439-b022-e0e4329bba79 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.396728] env[61957]: DEBUG oslo_vmware.api [None req-c02d7885-6c3b-4a1f-ab8a-8e683ed8e586 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Waiting for the task: (returnval){ [ 911.396728] env[61957]: value = "task-1277678" [ 911.396728] env[61957]: _type = "Task" [ 911.396728] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.406673] env[61957]: DEBUG oslo_vmware.api [None req-c02d7885-6c3b-4a1f-ab8a-8e683ed8e586 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Task: {'id': task-1277678, 'name': ReconfigVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.516979] env[61957]: DEBUG nova.network.neutron [None req-a7399dbc-0983-4a91-82d1-9c1b173d4f53 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 39b8b1df-efdd-4c22-9f31-85c85be4f0eb] Updating instance_info_cache with network_info: [{"id": "7e47ddcd-8c57-46be-8e68-62cec87e391b", "address": "fa:16:3e:95:ae:e3", "network": {"id": "efe08823-51fe-4620-a849-5345e262cc71", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1035436819-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e16f6dce3f0e44fb96de516f17d4c6f5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "667a2e97-c1be-421d-9941-6b84c2629b43", "external-id": "nsx-vlan-transportzone-484", "segmentation_id": 484, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7e47ddcd-8c", "ovs_interfaceid": "7e47ddcd-8c57-46be-8e68-62cec87e391b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 911.731518] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ba6997f6-9064-4c0a-859c-de0e4f8ea2ee tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.645s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 911.734380] env[61957]: DEBUG oslo_concurrency.lockutils [None req-5a91d738-be3a-4667-949d-376fc04acd42 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.806s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 911.736365] env[61957]: INFO nova.compute.claims [None req-5a91d738-be3a-4667-949d-376fc04acd42 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: d4674fbc-47b1-42d1-aaba-e86d46c51e8f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 911.750021] env[61957]: DEBUG oslo_vmware.api [None req-3c689a75-14e7-47a1-b918-cef3025704e3 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': task-1277676, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.108914} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.750021] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-3c689a75-14e7-47a1-b918-cef3025704e3 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: dbeeb200-70b9-4cb4-b5a4-182389d21918] Extended root virtual disk {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 911.750021] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-063cc048-29c5-4509-9196-5824131d7e58 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.783901] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-3c689a75-14e7-47a1-b918-cef3025704e3 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: dbeeb200-70b9-4cb4-b5a4-182389d21918] Reconfiguring VM instance instance-0000004b to attach disk [datastore1] dbeeb200-70b9-4cb4-b5a4-182389d21918/dbeeb200-70b9-4cb4-b5a4-182389d21918.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 911.785547] env[61957]: INFO nova.scheduler.client.report [None req-ba6997f6-9064-4c0a-859c-de0e4f8ea2ee tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Deleted allocations for instance 0adae8e1-8c2f-4110-805b-1f286debc833 [ 911.791436] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9a2266c3-b99e-4f88-9c90-95342c07c3b7 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.808665] env[61957]: DEBUG oslo_vmware.api [None req-35c8e096-fa6b-43c8-874d-91a51735ac8e tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Task: {'id': task-1277677, 'name': Rename_Task, 'duration_secs': 0.177428} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.811520] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-35c8e096-fa6b-43c8-874d-91a51735ac8e tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] [instance: a8ca1ad6-d636-4fa2-b0ac-53b020e392ce] Powering on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 911.812367] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-cc62503f-9685-4869-823e-df633776c523 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.818648] env[61957]: DEBUG oslo_vmware.api [None req-3c689a75-14e7-47a1-b918-cef3025704e3 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Waiting for the task: (returnval){ [ 911.818648] env[61957]: value = "task-1277679" [ 911.818648] env[61957]: _type = "Task" [ 911.818648] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.824148] env[61957]: DEBUG oslo_vmware.api [None req-35c8e096-fa6b-43c8-874d-91a51735ac8e tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Waiting for the task: (returnval){ [ 911.824148] env[61957]: value = "task-1277680" [ 911.824148] env[61957]: _type = "Task" [ 911.824148] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.832115] env[61957]: DEBUG oslo_vmware.api [None req-3c689a75-14e7-47a1-b918-cef3025704e3 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': task-1277679, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.835231] env[61957]: DEBUG oslo_vmware.api [None req-35c8e096-fa6b-43c8-874d-91a51735ac8e tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Task: {'id': task-1277680, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.908545] env[61957]: DEBUG oslo_vmware.api [None req-c02d7885-6c3b-4a1f-ab8a-8e683ed8e586 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Task: {'id': task-1277678, 'name': ReconfigVM_Task, 'duration_secs': 0.165114} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.909377] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-c02d7885-6c3b-4a1f-ab8a-8e683ed8e586 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] [instance: c23141ee-0cbb-4d1b-8390-c3073fe354f1] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-274545', 'volume_id': '734d094d-c382-4070-87c9-699fa1355b79', 'name': 'volume-734d094d-c382-4070-87c9-699fa1355b79', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'c23141ee-0cbb-4d1b-8390-c3073fe354f1', 'attached_at': '', 'detached_at': '', 'volume_id': '734d094d-c382-4070-87c9-699fa1355b79', 'serial': '734d094d-c382-4070-87c9-699fa1355b79'} {{(pid=61957) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 912.023248] env[61957]: DEBUG oslo_concurrency.lockutils [None req-a7399dbc-0983-4a91-82d1-9c1b173d4f53 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Releasing lock "refresh_cache-39b8b1df-efdd-4c22-9f31-85c85be4f0eb" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 912.023851] env[61957]: DEBUG nova.compute.manager [None req-a7399dbc-0983-4a91-82d1-9c1b173d4f53 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 39b8b1df-efdd-4c22-9f31-85c85be4f0eb] Instance network_info: |[{"id": "7e47ddcd-8c57-46be-8e68-62cec87e391b", "address": "fa:16:3e:95:ae:e3", "network": {"id": "efe08823-51fe-4620-a849-5345e262cc71", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1035436819-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e16f6dce3f0e44fb96de516f17d4c6f5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "667a2e97-c1be-421d-9941-6b84c2629b43", "external-id": "nsx-vlan-transportzone-484", "segmentation_id": 484, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7e47ddcd-8c", "ovs_interfaceid": "7e47ddcd-8c57-46be-8e68-62cec87e391b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 912.023998] env[61957]: DEBUG oslo_concurrency.lockutils [None req-9065790e-dc28-4275-a1c6-42f3b501d4d2 tempest-ServersV294TestFqdnHostnames-1426486100 tempest-ServersV294TestFqdnHostnames-1426486100-project-member] Acquiring lock "2e4a5344-600f-4b61-826e-c15f96b50af2" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 912.024251] env[61957]: DEBUG oslo_concurrency.lockutils [None req-9065790e-dc28-4275-a1c6-42f3b501d4d2 tempest-ServersV294TestFqdnHostnames-1426486100 tempest-ServersV294TestFqdnHostnames-1426486100-project-member] Lock "2e4a5344-600f-4b61-826e-c15f96b50af2" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 912.024460] env[61957]: DEBUG oslo_concurrency.lockutils [None req-9065790e-dc28-4275-a1c6-42f3b501d4d2 tempest-ServersV294TestFqdnHostnames-1426486100 tempest-ServersV294TestFqdnHostnames-1426486100-project-member] Acquiring lock "2e4a5344-600f-4b61-826e-c15f96b50af2-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 912.024643] env[61957]: DEBUG oslo_concurrency.lockutils [None req-9065790e-dc28-4275-a1c6-42f3b501d4d2 tempest-ServersV294TestFqdnHostnames-1426486100 tempest-ServersV294TestFqdnHostnames-1426486100-project-member] Lock "2e4a5344-600f-4b61-826e-c15f96b50af2-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 912.024810] env[61957]: DEBUG oslo_concurrency.lockutils [None req-9065790e-dc28-4275-a1c6-42f3b501d4d2 tempest-ServersV294TestFqdnHostnames-1426486100 tempest-ServersV294TestFqdnHostnames-1426486100-project-member] Lock "2e4a5344-600f-4b61-826e-c15f96b50af2-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 912.026694] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-a7399dbc-0983-4a91-82d1-9c1b173d4f53 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 39b8b1df-efdd-4c22-9f31-85c85be4f0eb] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:95:ae:e3', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '667a2e97-c1be-421d-9941-6b84c2629b43', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7e47ddcd-8c57-46be-8e68-62cec87e391b', 'vif_model': 'vmxnet3'}] {{(pid=61957) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 912.035148] env[61957]: DEBUG oslo.service.loopingcall [None req-a7399dbc-0983-4a91-82d1-9c1b173d4f53 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 912.035731] env[61957]: INFO nova.compute.manager [None req-9065790e-dc28-4275-a1c6-42f3b501d4d2 tempest-ServersV294TestFqdnHostnames-1426486100 tempest-ServersV294TestFqdnHostnames-1426486100-project-member] [instance: 2e4a5344-600f-4b61-826e-c15f96b50af2] Terminating instance [ 912.037225] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 39b8b1df-efdd-4c22-9f31-85c85be4f0eb] Creating VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 912.037790] env[61957]: DEBUG nova.compute.manager [None req-9065790e-dc28-4275-a1c6-42f3b501d4d2 tempest-ServersV294TestFqdnHostnames-1426486100 tempest-ServersV294TestFqdnHostnames-1426486100-project-member] [instance: 2e4a5344-600f-4b61-826e-c15f96b50af2] Start destroying the instance on the hypervisor. {{(pid=61957) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 912.037986] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-9065790e-dc28-4275-a1c6-42f3b501d4d2 tempest-ServersV294TestFqdnHostnames-1426486100 tempest-ServersV294TestFqdnHostnames-1426486100-project-member] [instance: 2e4a5344-600f-4b61-826e-c15f96b50af2] Destroying instance {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 912.038227] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2b96eb7f-223c-4b4d-871a-55fdc3208b52 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.054554] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee8072ce-d41a-47b9-84fe-552ad9daf90d {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.063587] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-9065790e-dc28-4275-a1c6-42f3b501d4d2 tempest-ServersV294TestFqdnHostnames-1426486100 tempest-ServersV294TestFqdnHostnames-1426486100-project-member] [instance: 2e4a5344-600f-4b61-826e-c15f96b50af2] Powering off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 912.063864] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7d8e9ab5-472a-4c23-a140-31898cdeddc1 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.066614] env[61957]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 912.066614] env[61957]: value = "task-1277681" [ 912.066614] env[61957]: _type = "Task" [ 912.066614] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.075220] env[61957]: DEBUG oslo_vmware.api [None req-9065790e-dc28-4275-a1c6-42f3b501d4d2 tempest-ServersV294TestFqdnHostnames-1426486100 tempest-ServersV294TestFqdnHostnames-1426486100-project-member] Waiting for the task: (returnval){ [ 912.075220] env[61957]: value = "task-1277682" [ 912.075220] env[61957]: _type = "Task" [ 912.075220] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.084024] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277681, 'name': CreateVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.088316] env[61957]: DEBUG oslo_vmware.api [None req-9065790e-dc28-4275-a1c6-42f3b501d4d2 tempest-ServersV294TestFqdnHostnames-1426486100 tempest-ServersV294TestFqdnHostnames-1426486100-project-member] Task: {'id': task-1277682, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.121934] env[61957]: DEBUG nova.compute.manager [None req-bb69c8c2-41e0-4fa7-aea3-e810870c1532 tempest-ServersNegativeTestMultiTenantJSON-2135399494 tempest-ServersNegativeTestMultiTenantJSON-2135399494-project-member] [instance: 19966b0a-53b7-48c5-849c-a9d00dc024f8] Start spawning the instance on the hypervisor. {{(pid=61957) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 912.160931] env[61957]: DEBUG nova.virt.hardware [None req-bb69c8c2-41e0-4fa7-aea3-e810870c1532 tempest-ServersNegativeTestMultiTenantJSON-2135399494 tempest-ServersNegativeTestMultiTenantJSON-2135399494-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T17:22:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T17:21:50Z,direct_url=,disk_format='vmdk',id=11c76a2c-f705-470a-ba9d-4657858bab38,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='81d3096d0f094373913fc3dc8f5c3c6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T17:21:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 912.161301] env[61957]: DEBUG nova.virt.hardware [None req-bb69c8c2-41e0-4fa7-aea3-e810870c1532 tempest-ServersNegativeTestMultiTenantJSON-2135399494 tempest-ServersNegativeTestMultiTenantJSON-2135399494-project-member] Flavor limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 912.161390] env[61957]: DEBUG nova.virt.hardware [None req-bb69c8c2-41e0-4fa7-aea3-e810870c1532 tempest-ServersNegativeTestMultiTenantJSON-2135399494 tempest-ServersNegativeTestMultiTenantJSON-2135399494-project-member] Image limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 912.161561] env[61957]: DEBUG nova.virt.hardware [None req-bb69c8c2-41e0-4fa7-aea3-e810870c1532 tempest-ServersNegativeTestMultiTenantJSON-2135399494 tempest-ServersNegativeTestMultiTenantJSON-2135399494-project-member] Flavor pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 912.161744] env[61957]: DEBUG nova.virt.hardware [None req-bb69c8c2-41e0-4fa7-aea3-e810870c1532 tempest-ServersNegativeTestMultiTenantJSON-2135399494 tempest-ServersNegativeTestMultiTenantJSON-2135399494-project-member] Image pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 912.162114] env[61957]: DEBUG nova.virt.hardware [None req-bb69c8c2-41e0-4fa7-aea3-e810870c1532 tempest-ServersNegativeTestMultiTenantJSON-2135399494 tempest-ServersNegativeTestMultiTenantJSON-2135399494-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 912.162308] env[61957]: DEBUG nova.virt.hardware [None req-bb69c8c2-41e0-4fa7-aea3-e810870c1532 tempest-ServersNegativeTestMultiTenantJSON-2135399494 tempest-ServersNegativeTestMultiTenantJSON-2135399494-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 912.163813] env[61957]: DEBUG nova.virt.hardware [None req-bb69c8c2-41e0-4fa7-aea3-e810870c1532 tempest-ServersNegativeTestMultiTenantJSON-2135399494 tempest-ServersNegativeTestMultiTenantJSON-2135399494-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 912.163813] env[61957]: DEBUG nova.virt.hardware [None req-bb69c8c2-41e0-4fa7-aea3-e810870c1532 tempest-ServersNegativeTestMultiTenantJSON-2135399494 tempest-ServersNegativeTestMultiTenantJSON-2135399494-project-member] Got 1 possible topologies {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 912.163813] env[61957]: DEBUG nova.virt.hardware [None req-bb69c8c2-41e0-4fa7-aea3-e810870c1532 tempest-ServersNegativeTestMultiTenantJSON-2135399494 tempest-ServersNegativeTestMultiTenantJSON-2135399494-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 912.163813] env[61957]: DEBUG nova.virt.hardware [None req-bb69c8c2-41e0-4fa7-aea3-e810870c1532 tempest-ServersNegativeTestMultiTenantJSON-2135399494 tempest-ServersNegativeTestMultiTenantJSON-2135399494-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 912.163813] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-248ceac3-90b0-4986-bfdf-46b850abe825 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.175602] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80810e29-de4f-4620-8b9b-89c3dc55b857 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.315814] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ba6997f6-9064-4c0a-859c-de0e4f8ea2ee tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Lock "0adae8e1-8c2f-4110-805b-1f286debc833" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 33.390s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 912.333055] env[61957]: DEBUG oslo_vmware.api [None req-3c689a75-14e7-47a1-b918-cef3025704e3 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': task-1277679, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.340036] env[61957]: DEBUG oslo_vmware.api [None req-35c8e096-fa6b-43c8-874d-91a51735ac8e tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Task: {'id': task-1277680, 'name': PowerOnVM_Task, 'duration_secs': 0.509045} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.340346] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-35c8e096-fa6b-43c8-874d-91a51735ac8e tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] [instance: a8ca1ad6-d636-4fa2-b0ac-53b020e392ce] Powered on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 912.340556] env[61957]: DEBUG nova.compute.manager [None req-35c8e096-fa6b-43c8-874d-91a51735ac8e tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] [instance: a8ca1ad6-d636-4fa2-b0ac-53b020e392ce] Checking state {{(pid=61957) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 912.341393] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5e95300-5f30-4de9-ab0a-bbae489fde7d {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.413551] env[61957]: DEBUG nova.compute.manager [req-7f7652be-17c1-4936-b888-5f34a8e683e2 req-7378c111-7c57-4c1f-b6c8-093141adc67b service nova] [instance: 39b8b1df-efdd-4c22-9f31-85c85be4f0eb] Received event network-changed-7e47ddcd-8c57-46be-8e68-62cec87e391b {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 912.413551] env[61957]: DEBUG nova.compute.manager [req-7f7652be-17c1-4936-b888-5f34a8e683e2 req-7378c111-7c57-4c1f-b6c8-093141adc67b service nova] [instance: 39b8b1df-efdd-4c22-9f31-85c85be4f0eb] Refreshing instance network info cache due to event network-changed-7e47ddcd-8c57-46be-8e68-62cec87e391b. {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 912.413918] env[61957]: DEBUG oslo_concurrency.lockutils [req-7f7652be-17c1-4936-b888-5f34a8e683e2 req-7378c111-7c57-4c1f-b6c8-093141adc67b service nova] Acquiring lock "refresh_cache-39b8b1df-efdd-4c22-9f31-85c85be4f0eb" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 912.413918] env[61957]: DEBUG oslo_concurrency.lockutils [req-7f7652be-17c1-4936-b888-5f34a8e683e2 req-7378c111-7c57-4c1f-b6c8-093141adc67b service nova] Acquired lock "refresh_cache-39b8b1df-efdd-4c22-9f31-85c85be4f0eb" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 912.414188] env[61957]: DEBUG nova.network.neutron [req-7f7652be-17c1-4936-b888-5f34a8e683e2 req-7378c111-7c57-4c1f-b6c8-093141adc67b service nova] [instance: 39b8b1df-efdd-4c22-9f31-85c85be4f0eb] Refreshing network info cache for port 7e47ddcd-8c57-46be-8e68-62cec87e391b {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 912.476024] env[61957]: DEBUG oslo_concurrency.lockutils [None req-68adc3d5-230e-4d31-9bdb-009ac1731c2d tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Acquiring lock "526ff179-62a6-4763-ab25-797617c4ed57" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 912.476024] env[61957]: DEBUG oslo_concurrency.lockutils [None req-68adc3d5-230e-4d31-9bdb-009ac1731c2d tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Lock "526ff179-62a6-4763-ab25-797617c4ed57" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 912.578224] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277681, 'name': CreateVM_Task, 'duration_secs': 0.417151} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.582029] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 39b8b1df-efdd-4c22-9f31-85c85be4f0eb] Created VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 912.582921] env[61957]: DEBUG oslo_concurrency.lockutils [None req-a7399dbc-0983-4a91-82d1-9c1b173d4f53 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 912.583130] env[61957]: DEBUG oslo_concurrency.lockutils [None req-a7399dbc-0983-4a91-82d1-9c1b173d4f53 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Acquired lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 912.583453] env[61957]: DEBUG oslo_concurrency.lockutils [None req-a7399dbc-0983-4a91-82d1-9c1b173d4f53 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 912.584087] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b6c70d44-7235-4728-9aeb-ff65919509d7 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.589585] env[61957]: DEBUG oslo_vmware.api [None req-9065790e-dc28-4275-a1c6-42f3b501d4d2 tempest-ServersV294TestFqdnHostnames-1426486100 tempest-ServersV294TestFqdnHostnames-1426486100-project-member] Task: {'id': task-1277682, 'name': PowerOffVM_Task, 'duration_secs': 0.196534} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.590768] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-9065790e-dc28-4275-a1c6-42f3b501d4d2 tempest-ServersV294TestFqdnHostnames-1426486100 tempest-ServersV294TestFqdnHostnames-1426486100-project-member] [instance: 2e4a5344-600f-4b61-826e-c15f96b50af2] Powered off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 912.590912] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-9065790e-dc28-4275-a1c6-42f3b501d4d2 tempest-ServersV294TestFqdnHostnames-1426486100 tempest-ServersV294TestFqdnHostnames-1426486100-project-member] [instance: 2e4a5344-600f-4b61-826e-c15f96b50af2] Unregistering the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 912.591681] env[61957]: DEBUG oslo_vmware.api [None req-a7399dbc-0983-4a91-82d1-9c1b173d4f53 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Waiting for the task: (returnval){ [ 912.591681] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]526f334a-3f75-b0bf-4fce-9efbacab3fe5" [ 912.591681] env[61957]: _type = "Task" [ 912.591681] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.591681] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4e6b0aca-2641-4117-a94a-e1831812a874 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.604322] env[61957]: DEBUG oslo_vmware.api [None req-a7399dbc-0983-4a91-82d1-9c1b173d4f53 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]526f334a-3f75-b0bf-4fce-9efbacab3fe5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.667211] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-9065790e-dc28-4275-a1c6-42f3b501d4d2 tempest-ServersV294TestFqdnHostnames-1426486100 tempest-ServersV294TestFqdnHostnames-1426486100-project-member] [instance: 2e4a5344-600f-4b61-826e-c15f96b50af2] Unregistered the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 912.667477] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-9065790e-dc28-4275-a1c6-42f3b501d4d2 tempest-ServersV294TestFqdnHostnames-1426486100 tempest-ServersV294TestFqdnHostnames-1426486100-project-member] [instance: 2e4a5344-600f-4b61-826e-c15f96b50af2] Deleting contents of the VM from datastore datastore1 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 912.667688] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-9065790e-dc28-4275-a1c6-42f3b501d4d2 tempest-ServersV294TestFqdnHostnames-1426486100 tempest-ServersV294TestFqdnHostnames-1426486100-project-member] Deleting the datastore file [datastore1] 2e4a5344-600f-4b61-826e-c15f96b50af2 {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 912.668430] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-975e9ca4-21d1-4d05-80a4-8abeedfd2bf9 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.675433] env[61957]: DEBUG oslo_vmware.api [None req-9065790e-dc28-4275-a1c6-42f3b501d4d2 tempest-ServersV294TestFqdnHostnames-1426486100 tempest-ServersV294TestFqdnHostnames-1426486100-project-member] Waiting for the task: (returnval){ [ 912.675433] env[61957]: value = "task-1277684" [ 912.675433] env[61957]: _type = "Task" [ 912.675433] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.684997] env[61957]: DEBUG oslo_vmware.api [None req-9065790e-dc28-4275-a1c6-42f3b501d4d2 tempest-ServersV294TestFqdnHostnames-1426486100 tempest-ServersV294TestFqdnHostnames-1426486100-project-member] Task: {'id': task-1277684, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.830550] env[61957]: DEBUG oslo_vmware.api [None req-3c689a75-14e7-47a1-b918-cef3025704e3 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': task-1277679, 'name': ReconfigVM_Task, 'duration_secs': 0.530714} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.830867] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-3c689a75-14e7-47a1-b918-cef3025704e3 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: dbeeb200-70b9-4cb4-b5a4-182389d21918] Reconfigured VM instance instance-0000004b to attach disk [datastore1] dbeeb200-70b9-4cb4-b5a4-182389d21918/dbeeb200-70b9-4cb4-b5a4-182389d21918.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 912.831539] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-fffd95fe-fcd5-4f40-9d53-d51125444b99 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.839594] env[61957]: DEBUG oslo_vmware.api [None req-3c689a75-14e7-47a1-b918-cef3025704e3 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Waiting for the task: (returnval){ [ 912.839594] env[61957]: value = "task-1277685" [ 912.839594] env[61957]: _type = "Task" [ 912.839594] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.851907] env[61957]: DEBUG oslo_vmware.api [None req-3c689a75-14e7-47a1-b918-cef3025704e3 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': task-1277685, 'name': Rename_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.851907] env[61957]: DEBUG nova.network.neutron [None req-bb69c8c2-41e0-4fa7-aea3-e810870c1532 tempest-ServersNegativeTestMultiTenantJSON-2135399494 tempest-ServersNegativeTestMultiTenantJSON-2135399494-project-member] [instance: 19966b0a-53b7-48c5-849c-a9d00dc024f8] Successfully updated port: 159b1e72-9328-460d-b334-9a3ad5d73b44 {{(pid=61957) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 912.861125] env[61957]: DEBUG oslo_concurrency.lockutils [None req-35c8e096-fa6b-43c8-874d-91a51735ac8e tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 912.955286] env[61957]: DEBUG nova.objects.instance [None req-c02d7885-6c3b-4a1f-ab8a-8e683ed8e586 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Lazy-loading 'flavor' on Instance uuid c23141ee-0cbb-4d1b-8390-c3073fe354f1 {{(pid=61957) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 912.978285] env[61957]: DEBUG nova.compute.utils [None req-68adc3d5-230e-4d31-9bdb-009ac1731c2d tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Using /dev/sd instead of None {{(pid=61957) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 913.076855] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bea620f-4ccb-4c01-bba8-866943ba349c {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.085653] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55164f47-fc66-4e4c-9c34-27f0b006fd3b {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.121355] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73278d2c-5257-4697-bd5e-33a43b6efc30 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.131574] env[61957]: DEBUG oslo_vmware.api [None req-a7399dbc-0983-4a91-82d1-9c1b173d4f53 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]526f334a-3f75-b0bf-4fce-9efbacab3fe5, 'name': SearchDatastore_Task, 'duration_secs': 0.014632} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.134006] env[61957]: DEBUG oslo_concurrency.lockutils [None req-a7399dbc-0983-4a91-82d1-9c1b173d4f53 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Releasing lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 913.134273] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-a7399dbc-0983-4a91-82d1-9c1b173d4f53 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 39b8b1df-efdd-4c22-9f31-85c85be4f0eb] Processing image 11c76a2c-f705-470a-ba9d-4657858bab38 {{(pid=61957) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 913.134515] env[61957]: DEBUG oslo_concurrency.lockutils [None req-a7399dbc-0983-4a91-82d1-9c1b173d4f53 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 913.134665] env[61957]: DEBUG oslo_concurrency.lockutils [None req-a7399dbc-0983-4a91-82d1-9c1b173d4f53 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Acquired lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 913.134841] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-a7399dbc-0983-4a91-82d1-9c1b173d4f53 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 913.135176] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-82d53c6e-7126-4f37-b2ef-7d3f91f9fea5 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.138028] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be629d1e-29d9-4625-97ca-8ec2884328a2 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.153982] env[61957]: DEBUG nova.compute.provider_tree [None req-5a91d738-be3a-4667-949d-376fc04acd42 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 913.156415] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-a7399dbc-0983-4a91-82d1-9c1b173d4f53 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 913.156585] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-a7399dbc-0983-4a91-82d1-9c1b173d4f53 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61957) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 913.157522] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d52dc6a3-37df-492d-9851-12a860fba6aa {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.163910] env[61957]: DEBUG oslo_vmware.api [None req-a7399dbc-0983-4a91-82d1-9c1b173d4f53 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Waiting for the task: (returnval){ [ 913.163910] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]525435ec-2c4c-a854-1d65-7b0cb8d2a4e9" [ 913.163910] env[61957]: _type = "Task" [ 913.163910] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.169432] env[61957]: DEBUG nova.network.neutron [req-7f7652be-17c1-4936-b888-5f34a8e683e2 req-7378c111-7c57-4c1f-b6c8-093141adc67b service nova] [instance: 39b8b1df-efdd-4c22-9f31-85c85be4f0eb] Updated VIF entry in instance network info cache for port 7e47ddcd-8c57-46be-8e68-62cec87e391b. {{(pid=61957) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 913.169770] env[61957]: DEBUG nova.network.neutron [req-7f7652be-17c1-4936-b888-5f34a8e683e2 req-7378c111-7c57-4c1f-b6c8-093141adc67b service nova] [instance: 39b8b1df-efdd-4c22-9f31-85c85be4f0eb] Updating instance_info_cache with network_info: [{"id": "7e47ddcd-8c57-46be-8e68-62cec87e391b", "address": "fa:16:3e:95:ae:e3", "network": {"id": "efe08823-51fe-4620-a849-5345e262cc71", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1035436819-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e16f6dce3f0e44fb96de516f17d4c6f5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "667a2e97-c1be-421d-9941-6b84c2629b43", "external-id": "nsx-vlan-transportzone-484", "segmentation_id": 484, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7e47ddcd-8c", "ovs_interfaceid": "7e47ddcd-8c57-46be-8e68-62cec87e391b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 913.174270] env[61957]: DEBUG oslo_vmware.api [None req-a7399dbc-0983-4a91-82d1-9c1b173d4f53 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]525435ec-2c4c-a854-1d65-7b0cb8d2a4e9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.185463] env[61957]: DEBUG oslo_vmware.api [None req-9065790e-dc28-4275-a1c6-42f3b501d4d2 tempest-ServersV294TestFqdnHostnames-1426486100 tempest-ServersV294TestFqdnHostnames-1426486100-project-member] Task: {'id': task-1277684, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.208856} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.185716] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-9065790e-dc28-4275-a1c6-42f3b501d4d2 tempest-ServersV294TestFqdnHostnames-1426486100 tempest-ServersV294TestFqdnHostnames-1426486100-project-member] Deleted the datastore file {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 913.185900] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-9065790e-dc28-4275-a1c6-42f3b501d4d2 tempest-ServersV294TestFqdnHostnames-1426486100 tempest-ServersV294TestFqdnHostnames-1426486100-project-member] [instance: 2e4a5344-600f-4b61-826e-c15f96b50af2] Deleted contents of the VM from datastore datastore1 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 913.186092] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-9065790e-dc28-4275-a1c6-42f3b501d4d2 tempest-ServersV294TestFqdnHostnames-1426486100 tempest-ServersV294TestFqdnHostnames-1426486100-project-member] [instance: 2e4a5344-600f-4b61-826e-c15f96b50af2] Instance destroyed {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 913.186272] env[61957]: INFO nova.compute.manager [None req-9065790e-dc28-4275-a1c6-42f3b501d4d2 tempest-ServersV294TestFqdnHostnames-1426486100 tempest-ServersV294TestFqdnHostnames-1426486100-project-member] [instance: 2e4a5344-600f-4b61-826e-c15f96b50af2] Took 1.15 seconds to destroy the instance on the hypervisor. [ 913.186516] env[61957]: DEBUG oslo.service.loopingcall [None req-9065790e-dc28-4275-a1c6-42f3b501d4d2 tempest-ServersV294TestFqdnHostnames-1426486100 tempest-ServersV294TestFqdnHostnames-1426486100-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 913.186710] env[61957]: DEBUG nova.compute.manager [-] [instance: 2e4a5344-600f-4b61-826e-c15f96b50af2] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 913.186806] env[61957]: DEBUG nova.network.neutron [-] [instance: 2e4a5344-600f-4b61-826e-c15f96b50af2] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 913.356739] env[61957]: DEBUG oslo_vmware.api [None req-3c689a75-14e7-47a1-b918-cef3025704e3 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': task-1277685, 'name': Rename_Task, 'duration_secs': 0.270888} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.357699] env[61957]: DEBUG oslo_concurrency.lockutils [None req-bb69c8c2-41e0-4fa7-aea3-e810870c1532 tempest-ServersNegativeTestMultiTenantJSON-2135399494 tempest-ServersNegativeTestMultiTenantJSON-2135399494-project-member] Acquiring lock "refresh_cache-19966b0a-53b7-48c5-849c-a9d00dc024f8" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 913.357840] env[61957]: DEBUG oslo_concurrency.lockutils [None req-bb69c8c2-41e0-4fa7-aea3-e810870c1532 tempest-ServersNegativeTestMultiTenantJSON-2135399494 tempest-ServersNegativeTestMultiTenantJSON-2135399494-project-member] Acquired lock "refresh_cache-19966b0a-53b7-48c5-849c-a9d00dc024f8" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 913.357983] env[61957]: DEBUG nova.network.neutron [None req-bb69c8c2-41e0-4fa7-aea3-e810870c1532 tempest-ServersNegativeTestMultiTenantJSON-2135399494 tempest-ServersNegativeTestMultiTenantJSON-2135399494-project-member] [instance: 19966b0a-53b7-48c5-849c-a9d00dc024f8] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 913.359121] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c689a75-14e7-47a1-b918-cef3025704e3 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: dbeeb200-70b9-4cb4-b5a4-182389d21918] Powering on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 913.359560] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0bc0e29e-4963-44b6-972f-e0d9b4b1c2cb {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.367648] env[61957]: DEBUG oslo_vmware.api [None req-3c689a75-14e7-47a1-b918-cef3025704e3 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Waiting for the task: (returnval){ [ 913.367648] env[61957]: value = "task-1277686" [ 913.367648] env[61957]: _type = "Task" [ 913.367648] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.378432] env[61957]: DEBUG oslo_vmware.api [None req-3c689a75-14e7-47a1-b918-cef3025704e3 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': task-1277686, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.463285] env[61957]: DEBUG oslo_concurrency.lockutils [None req-c02d7885-6c3b-4a1f-ab8a-8e683ed8e586 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Lock "c23141ee-0cbb-4d1b-8390-c3073fe354f1" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.312s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 913.480897] env[61957]: DEBUG oslo_concurrency.lockutils [None req-68adc3d5-230e-4d31-9bdb-009ac1731c2d tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Lock "526ff179-62a6-4763-ab25-797617c4ed57" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.005s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 913.658873] env[61957]: DEBUG nova.scheduler.client.report [None req-5a91d738-be3a-4667-949d-376fc04acd42 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 913.675318] env[61957]: DEBUG oslo_vmware.api [None req-a7399dbc-0983-4a91-82d1-9c1b173d4f53 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]525435ec-2c4c-a854-1d65-7b0cb8d2a4e9, 'name': SearchDatastore_Task, 'duration_secs': 0.022267} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.675743] env[61957]: DEBUG oslo_concurrency.lockutils [req-7f7652be-17c1-4936-b888-5f34a8e683e2 req-7378c111-7c57-4c1f-b6c8-093141adc67b service nova] Releasing lock "refresh_cache-39b8b1df-efdd-4c22-9f31-85c85be4f0eb" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 913.675980] env[61957]: DEBUG nova.compute.manager [req-7f7652be-17c1-4936-b888-5f34a8e683e2 req-7378c111-7c57-4c1f-b6c8-093141adc67b service nova] [instance: 736eabcf-70f8-4e1e-9830-270d9971fc9a] Received event network-changed-95763a65-6383-409f-a9d6-1f37aaf44d34 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 913.676171] env[61957]: DEBUG nova.compute.manager [req-7f7652be-17c1-4936-b888-5f34a8e683e2 req-7378c111-7c57-4c1f-b6c8-093141adc67b service nova] [instance: 736eabcf-70f8-4e1e-9830-270d9971fc9a] Refreshing instance network info cache due to event network-changed-95763a65-6383-409f-a9d6-1f37aaf44d34. {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 913.676380] env[61957]: DEBUG oslo_concurrency.lockutils [req-7f7652be-17c1-4936-b888-5f34a8e683e2 req-7378c111-7c57-4c1f-b6c8-093141adc67b service nova] Acquiring lock "refresh_cache-736eabcf-70f8-4e1e-9830-270d9971fc9a" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 913.676525] env[61957]: DEBUG oslo_concurrency.lockutils [req-7f7652be-17c1-4936-b888-5f34a8e683e2 req-7378c111-7c57-4c1f-b6c8-093141adc67b service nova] Acquired lock "refresh_cache-736eabcf-70f8-4e1e-9830-270d9971fc9a" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 913.676687] env[61957]: DEBUG nova.network.neutron [req-7f7652be-17c1-4936-b888-5f34a8e683e2 req-7378c111-7c57-4c1f-b6c8-093141adc67b service nova] [instance: 736eabcf-70f8-4e1e-9830-270d9971fc9a] Refreshing network info cache for port 95763a65-6383-409f-a9d6-1f37aaf44d34 {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 913.678956] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-49b5b23a-b489-4260-b9c3-d5cc20f670d1 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.699092] env[61957]: DEBUG oslo_vmware.api [None req-a7399dbc-0983-4a91-82d1-9c1b173d4f53 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Waiting for the task: (returnval){ [ 913.699092] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52162003-ce28-85a5-3c11-e6c210eca4cd" [ 913.699092] env[61957]: _type = "Task" [ 913.699092] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.709109] env[61957]: DEBUG oslo_vmware.api [None req-a7399dbc-0983-4a91-82d1-9c1b173d4f53 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52162003-ce28-85a5-3c11-e6c210eca4cd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.880588] env[61957]: DEBUG oslo_vmware.api [None req-3c689a75-14e7-47a1-b918-cef3025704e3 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': task-1277686, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.887780] env[61957]: DEBUG nova.network.neutron [None req-bb69c8c2-41e0-4fa7-aea3-e810870c1532 tempest-ServersNegativeTestMultiTenantJSON-2135399494 tempest-ServersNegativeTestMultiTenantJSON-2135399494-project-member] [instance: 19966b0a-53b7-48c5-849c-a9d00dc024f8] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 914.021392] env[61957]: DEBUG nova.network.neutron [None req-bb69c8c2-41e0-4fa7-aea3-e810870c1532 tempest-ServersNegativeTestMultiTenantJSON-2135399494 tempest-ServersNegativeTestMultiTenantJSON-2135399494-project-member] [instance: 19966b0a-53b7-48c5-849c-a9d00dc024f8] Updating instance_info_cache with network_info: [{"id": "159b1e72-9328-460d-b334-9a3ad5d73b44", "address": "fa:16:3e:69:8b:7f", "network": {"id": "6a4b928d-7db0-4a1f-ba48-5af87be6c39a", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-1923984009-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "219263b31ffc43a0904d8e620e48e085", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c4349e30-c086-4c24-9e0e-83996d808a1b", "external-id": "nsx-vlan-transportzone-266", "segmentation_id": 266, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap159b1e72-93", "ovs_interfaceid": "159b1e72-9328-460d-b334-9a3ad5d73b44", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 914.145706] env[61957]: DEBUG nova.network.neutron [-] [instance: 2e4a5344-600f-4b61-826e-c15f96b50af2] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 914.164826] env[61957]: DEBUG oslo_concurrency.lockutils [None req-5a91d738-be3a-4667-949d-376fc04acd42 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.430s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 914.165575] env[61957]: DEBUG nova.compute.manager [None req-5a91d738-be3a-4667-949d-376fc04acd42 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: d4674fbc-47b1-42d1-aaba-e86d46c51e8f] Start building networks asynchronously for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 914.170905] env[61957]: DEBUG oslo_concurrency.lockutils [None req-9513c200-94de-43f4-aaa5-418ecbddf25f tempest-ServerTagsTestJSON-969152625 tempest-ServerTagsTestJSON-969152625-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 28.712s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 914.171877] env[61957]: DEBUG nova.objects.instance [None req-9513c200-94de-43f4-aaa5-418ecbddf25f tempest-ServerTagsTestJSON-969152625 tempest-ServerTagsTestJSON-969152625-project-member] Lazy-loading 'resources' on Instance uuid 8e04a098-25de-4d57-9e3c-ea44d234d57e {{(pid=61957) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 914.212206] env[61957]: DEBUG oslo_vmware.api [None req-a7399dbc-0983-4a91-82d1-9c1b173d4f53 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52162003-ce28-85a5-3c11-e6c210eca4cd, 'name': SearchDatastore_Task, 'duration_secs': 0.027584} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.216218] env[61957]: DEBUG oslo_concurrency.lockutils [None req-a7399dbc-0983-4a91-82d1-9c1b173d4f53 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Releasing lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 914.216218] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-a7399dbc-0983-4a91-82d1-9c1b173d4f53 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore1] 39b8b1df-efdd-4c22-9f31-85c85be4f0eb/39b8b1df-efdd-4c22-9f31-85c85be4f0eb.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 914.216218] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0edfb334-93a4-4487-9e18-ae6679836347 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.225112] env[61957]: DEBUG oslo_vmware.api [None req-a7399dbc-0983-4a91-82d1-9c1b173d4f53 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Waiting for the task: (returnval){ [ 914.225112] env[61957]: value = "task-1277687" [ 914.225112] env[61957]: _type = "Task" [ 914.225112] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.235861] env[61957]: DEBUG oslo_vmware.api [None req-a7399dbc-0983-4a91-82d1-9c1b173d4f53 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1277687, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.383398] env[61957]: DEBUG oslo_vmware.api [None req-3c689a75-14e7-47a1-b918-cef3025704e3 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': task-1277686, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.453207] env[61957]: DEBUG nova.network.neutron [req-7f7652be-17c1-4936-b888-5f34a8e683e2 req-7378c111-7c57-4c1f-b6c8-093141adc67b service nova] [instance: 736eabcf-70f8-4e1e-9830-270d9971fc9a] Updated VIF entry in instance network info cache for port 95763a65-6383-409f-a9d6-1f37aaf44d34. {{(pid=61957) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 914.454030] env[61957]: DEBUG nova.network.neutron [req-7f7652be-17c1-4936-b888-5f34a8e683e2 req-7378c111-7c57-4c1f-b6c8-093141adc67b service nova] [instance: 736eabcf-70f8-4e1e-9830-270d9971fc9a] Updating instance_info_cache with network_info: [{"id": "95763a65-6383-409f-a9d6-1f37aaf44d34", "address": "fa:16:3e:64:07:83", "network": {"id": "b303ad01-ef9b-4119-9722-3be27ffb30bd", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-96064402-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.145", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "55742b1d4fbb401ea61f990ecaa2def6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a8b99a46-3e7f-4ef1-9e45-58e6cd17f210", "external-id": "nsx-vlan-transportzone-704", "segmentation_id": 704, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap95763a65-63", "ovs_interfaceid": "95763a65-6383-409f-a9d6-1f37aaf44d34", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 914.458989] env[61957]: DEBUG oslo_concurrency.lockutils [None req-24cac752-c949-48ff-b582-7f3e67bd6cce tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Acquiring lock "c23141ee-0cbb-4d1b-8390-c3073fe354f1" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 914.459271] env[61957]: DEBUG oslo_concurrency.lockutils [None req-24cac752-c949-48ff-b582-7f3e67bd6cce tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Lock "c23141ee-0cbb-4d1b-8390-c3073fe354f1" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 914.477220] env[61957]: DEBUG nova.compute.manager [req-b4defd35-5ba1-4901-aec9-0a4f56ed7b79 req-25205183-5134-4679-bdb0-0658bc4ff8c4 service nova] [instance: 19966b0a-53b7-48c5-849c-a9d00dc024f8] Received event network-vif-plugged-159b1e72-9328-460d-b334-9a3ad5d73b44 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 914.477220] env[61957]: DEBUG oslo_concurrency.lockutils [req-b4defd35-5ba1-4901-aec9-0a4f56ed7b79 req-25205183-5134-4679-bdb0-0658bc4ff8c4 service nova] Acquiring lock "19966b0a-53b7-48c5-849c-a9d00dc024f8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 914.477220] env[61957]: DEBUG oslo_concurrency.lockutils [req-b4defd35-5ba1-4901-aec9-0a4f56ed7b79 req-25205183-5134-4679-bdb0-0658bc4ff8c4 service nova] Lock "19966b0a-53b7-48c5-849c-a9d00dc024f8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 914.477220] env[61957]: DEBUG oslo_concurrency.lockutils [req-b4defd35-5ba1-4901-aec9-0a4f56ed7b79 req-25205183-5134-4679-bdb0-0658bc4ff8c4 service nova] Lock "19966b0a-53b7-48c5-849c-a9d00dc024f8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 914.477220] env[61957]: DEBUG nova.compute.manager [req-b4defd35-5ba1-4901-aec9-0a4f56ed7b79 req-25205183-5134-4679-bdb0-0658bc4ff8c4 service nova] [instance: 19966b0a-53b7-48c5-849c-a9d00dc024f8] No waiting events found dispatching network-vif-plugged-159b1e72-9328-460d-b334-9a3ad5d73b44 {{(pid=61957) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 914.477220] env[61957]: WARNING nova.compute.manager [req-b4defd35-5ba1-4901-aec9-0a4f56ed7b79 req-25205183-5134-4679-bdb0-0658bc4ff8c4 service nova] [instance: 19966b0a-53b7-48c5-849c-a9d00dc024f8] Received unexpected event network-vif-plugged-159b1e72-9328-460d-b334-9a3ad5d73b44 for instance with vm_state building and task_state spawning. [ 914.477220] env[61957]: DEBUG nova.compute.manager [req-b4defd35-5ba1-4901-aec9-0a4f56ed7b79 req-25205183-5134-4679-bdb0-0658bc4ff8c4 service nova] [instance: 19966b0a-53b7-48c5-849c-a9d00dc024f8] Received event network-changed-159b1e72-9328-460d-b334-9a3ad5d73b44 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 914.477220] env[61957]: DEBUG nova.compute.manager [req-b4defd35-5ba1-4901-aec9-0a4f56ed7b79 req-25205183-5134-4679-bdb0-0658bc4ff8c4 service nova] [instance: 19966b0a-53b7-48c5-849c-a9d00dc024f8] Refreshing instance network info cache due to event network-changed-159b1e72-9328-460d-b334-9a3ad5d73b44. {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 914.477220] env[61957]: DEBUG oslo_concurrency.lockutils [req-b4defd35-5ba1-4901-aec9-0a4f56ed7b79 req-25205183-5134-4679-bdb0-0658bc4ff8c4 service nova] Acquiring lock "refresh_cache-19966b0a-53b7-48c5-849c-a9d00dc024f8" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 914.524743] env[61957]: DEBUG oslo_concurrency.lockutils [None req-bb69c8c2-41e0-4fa7-aea3-e810870c1532 tempest-ServersNegativeTestMultiTenantJSON-2135399494 tempest-ServersNegativeTestMultiTenantJSON-2135399494-project-member] Releasing lock "refresh_cache-19966b0a-53b7-48c5-849c-a9d00dc024f8" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 914.525765] env[61957]: DEBUG nova.compute.manager [None req-bb69c8c2-41e0-4fa7-aea3-e810870c1532 tempest-ServersNegativeTestMultiTenantJSON-2135399494 tempest-ServersNegativeTestMultiTenantJSON-2135399494-project-member] [instance: 19966b0a-53b7-48c5-849c-a9d00dc024f8] Instance network_info: |[{"id": "159b1e72-9328-460d-b334-9a3ad5d73b44", "address": "fa:16:3e:69:8b:7f", "network": {"id": "6a4b928d-7db0-4a1f-ba48-5af87be6c39a", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-1923984009-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "219263b31ffc43a0904d8e620e48e085", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c4349e30-c086-4c24-9e0e-83996d808a1b", "external-id": "nsx-vlan-transportzone-266", "segmentation_id": 266, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap159b1e72-93", "ovs_interfaceid": "159b1e72-9328-460d-b334-9a3ad5d73b44", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 914.525765] env[61957]: DEBUG oslo_concurrency.lockutils [req-b4defd35-5ba1-4901-aec9-0a4f56ed7b79 req-25205183-5134-4679-bdb0-0658bc4ff8c4 service nova] Acquired lock "refresh_cache-19966b0a-53b7-48c5-849c-a9d00dc024f8" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 914.525765] env[61957]: DEBUG nova.network.neutron [req-b4defd35-5ba1-4901-aec9-0a4f56ed7b79 req-25205183-5134-4679-bdb0-0658bc4ff8c4 service nova] [instance: 19966b0a-53b7-48c5-849c-a9d00dc024f8] Refreshing network info cache for port 159b1e72-9328-460d-b334-9a3ad5d73b44 {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 914.526930] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-bb69c8c2-41e0-4fa7-aea3-e810870c1532 tempest-ServersNegativeTestMultiTenantJSON-2135399494 tempest-ServersNegativeTestMultiTenantJSON-2135399494-project-member] [instance: 19966b0a-53b7-48c5-849c-a9d00dc024f8] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:69:8b:7f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c4349e30-c086-4c24-9e0e-83996d808a1b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '159b1e72-9328-460d-b334-9a3ad5d73b44', 'vif_model': 'vmxnet3'}] {{(pid=61957) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 914.534878] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-bb69c8c2-41e0-4fa7-aea3-e810870c1532 tempest-ServersNegativeTestMultiTenantJSON-2135399494 tempest-ServersNegativeTestMultiTenantJSON-2135399494-project-member] Creating folder: Project (219263b31ffc43a0904d8e620e48e085). Parent ref: group-v274445. {{(pid=61957) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 914.536025] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a614ab3c-3d2c-4e6a-8f5e-ebef3f928a14 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.549104] env[61957]: DEBUG oslo_concurrency.lockutils [None req-68adc3d5-230e-4d31-9bdb-009ac1731c2d tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Acquiring lock "526ff179-62a6-4763-ab25-797617c4ed57" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 914.549404] env[61957]: DEBUG oslo_concurrency.lockutils [None req-68adc3d5-230e-4d31-9bdb-009ac1731c2d tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Lock "526ff179-62a6-4763-ab25-797617c4ed57" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 914.549645] env[61957]: INFO nova.compute.manager [None req-68adc3d5-230e-4d31-9bdb-009ac1731c2d tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] [instance: 526ff179-62a6-4763-ab25-797617c4ed57] Attaching volume 6c0df615-4d1a-45c6-9f48-ab1813e515b1 to /dev/sdb [ 914.552963] env[61957]: INFO nova.virt.vmwareapi.vm_util [None req-bb69c8c2-41e0-4fa7-aea3-e810870c1532 tempest-ServersNegativeTestMultiTenantJSON-2135399494 tempest-ServersNegativeTestMultiTenantJSON-2135399494-project-member] Created folder: Project (219263b31ffc43a0904d8e620e48e085) in parent group-v274445. [ 914.553134] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-bb69c8c2-41e0-4fa7-aea3-e810870c1532 tempest-ServersNegativeTestMultiTenantJSON-2135399494 tempest-ServersNegativeTestMultiTenantJSON-2135399494-project-member] Creating folder: Instances. Parent ref: group-v274549. {{(pid=61957) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 914.555621] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-21ec8163-84f7-45b4-94ec-9c48c33d860a {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.571797] env[61957]: INFO nova.virt.vmwareapi.vm_util [None req-bb69c8c2-41e0-4fa7-aea3-e810870c1532 tempest-ServersNegativeTestMultiTenantJSON-2135399494 tempest-ServersNegativeTestMultiTenantJSON-2135399494-project-member] Created folder: Instances in parent group-v274549. [ 914.571797] env[61957]: DEBUG oslo.service.loopingcall [None req-bb69c8c2-41e0-4fa7-aea3-e810870c1532 tempest-ServersNegativeTestMultiTenantJSON-2135399494 tempest-ServersNegativeTestMultiTenantJSON-2135399494-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 914.571983] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 19966b0a-53b7-48c5-849c-a9d00dc024f8] Creating VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 914.572230] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3e5772d9-e9e9-490a-9917-030c4e039005 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.593986] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-712bffc7-4ffa-4723-8b91-f735de0cfcba {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.598144] env[61957]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 914.598144] env[61957]: value = "task-1277690" [ 914.598144] env[61957]: _type = "Task" [ 914.598144] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.605026] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95bfd6e2-60a8-43d3-a468-a90e78bcf295 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.611529] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277690, 'name': CreateVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.625411] env[61957]: DEBUG nova.virt.block_device [None req-68adc3d5-230e-4d31-9bdb-009ac1731c2d tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] [instance: 526ff179-62a6-4763-ab25-797617c4ed57] Updating existing volume attachment record: dd64e422-e884-4609-8232-c03593fe184e {{(pid=61957) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 914.653534] env[61957]: INFO nova.compute.manager [-] [instance: 2e4a5344-600f-4b61-826e-c15f96b50af2] Took 1.47 seconds to deallocate network for instance. [ 914.676072] env[61957]: DEBUG nova.compute.utils [None req-5a91d738-be3a-4667-949d-376fc04acd42 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Using /dev/sd instead of None {{(pid=61957) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 914.687763] env[61957]: DEBUG nova.compute.manager [None req-5a91d738-be3a-4667-949d-376fc04acd42 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: d4674fbc-47b1-42d1-aaba-e86d46c51e8f] Allocating IP information in the background. {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 914.687763] env[61957]: DEBUG nova.network.neutron [None req-5a91d738-be3a-4667-949d-376fc04acd42 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: d4674fbc-47b1-42d1-aaba-e86d46c51e8f] allocate_for_instance() {{(pid=61957) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 914.743115] env[61957]: DEBUG oslo_vmware.api [None req-a7399dbc-0983-4a91-82d1-9c1b173d4f53 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1277687, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.745232] env[61957]: DEBUG nova.policy [None req-5a91d738-be3a-4667-949d-376fc04acd42 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '76e1754f96ed438ea0ed91ad337419f1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '975189ebb3cc4cdb9391880f0c9ba6ba', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61957) authorize /opt/stack/nova/nova/policy.py:203}} [ 914.884273] env[61957]: DEBUG oslo_vmware.api [None req-3c689a75-14e7-47a1-b918-cef3025704e3 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': task-1277686, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.888948] env[61957]: DEBUG oslo_concurrency.lockutils [None req-63a396d5-1afb-44c8-a210-9949507fdc49 tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Acquiring lock "a8ca1ad6-d636-4fa2-b0ac-53b020e392ce" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 914.889264] env[61957]: DEBUG oslo_concurrency.lockutils [None req-63a396d5-1afb-44c8-a210-9949507fdc49 tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Lock "a8ca1ad6-d636-4fa2-b0ac-53b020e392ce" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 914.889502] env[61957]: DEBUG oslo_concurrency.lockutils [None req-63a396d5-1afb-44c8-a210-9949507fdc49 tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Acquiring lock "a8ca1ad6-d636-4fa2-b0ac-53b020e392ce-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 914.889685] env[61957]: DEBUG oslo_concurrency.lockutils [None req-63a396d5-1afb-44c8-a210-9949507fdc49 tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Lock "a8ca1ad6-d636-4fa2-b0ac-53b020e392ce-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 914.889970] env[61957]: DEBUG oslo_concurrency.lockutils [None req-63a396d5-1afb-44c8-a210-9949507fdc49 tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Lock "a8ca1ad6-d636-4fa2-b0ac-53b020e392ce-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 914.894660] env[61957]: INFO nova.compute.manager [None req-63a396d5-1afb-44c8-a210-9949507fdc49 tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] [instance: a8ca1ad6-d636-4fa2-b0ac-53b020e392ce] Terminating instance [ 914.896718] env[61957]: DEBUG oslo_concurrency.lockutils [None req-63a396d5-1afb-44c8-a210-9949507fdc49 tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Acquiring lock "refresh_cache-a8ca1ad6-d636-4fa2-b0ac-53b020e392ce" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 914.896846] env[61957]: DEBUG oslo_concurrency.lockutils [None req-63a396d5-1afb-44c8-a210-9949507fdc49 tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Acquired lock "refresh_cache-a8ca1ad6-d636-4fa2-b0ac-53b020e392ce" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 914.897046] env[61957]: DEBUG nova.network.neutron [None req-63a396d5-1afb-44c8-a210-9949507fdc49 tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] [instance: a8ca1ad6-d636-4fa2-b0ac-53b020e392ce] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 914.956830] env[61957]: DEBUG oslo_concurrency.lockutils [req-7f7652be-17c1-4936-b888-5f34a8e683e2 req-7378c111-7c57-4c1f-b6c8-093141adc67b service nova] Releasing lock "refresh_cache-736eabcf-70f8-4e1e-9830-270d9971fc9a" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 914.963167] env[61957]: INFO nova.compute.manager [None req-24cac752-c949-48ff-b582-7f3e67bd6cce tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] [instance: c23141ee-0cbb-4d1b-8390-c3073fe354f1] Detaching volume 734d094d-c382-4070-87c9-699fa1355b79 [ 914.983614] env[61957]: DEBUG oslo_concurrency.lockutils [None req-94184bcc-7df6-4480-913f-5d000fd6dd4e tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Acquiring lock "d5d10a31-0e78-4ed7-b944-9208138a4861" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 914.983614] env[61957]: DEBUG oslo_concurrency.lockutils [None req-94184bcc-7df6-4480-913f-5d000fd6dd4e tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Lock "d5d10a31-0e78-4ed7-b944-9208138a4861" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 914.983614] env[61957]: DEBUG oslo_concurrency.lockutils [None req-94184bcc-7df6-4480-913f-5d000fd6dd4e tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Acquiring lock "d5d10a31-0e78-4ed7-b944-9208138a4861-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 914.983614] env[61957]: DEBUG oslo_concurrency.lockutils [None req-94184bcc-7df6-4480-913f-5d000fd6dd4e tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Lock "d5d10a31-0e78-4ed7-b944-9208138a4861-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 914.983614] env[61957]: DEBUG oslo_concurrency.lockutils [None req-94184bcc-7df6-4480-913f-5d000fd6dd4e tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Lock "d5d10a31-0e78-4ed7-b944-9208138a4861-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 914.984516] env[61957]: INFO nova.compute.manager [None req-94184bcc-7df6-4480-913f-5d000fd6dd4e tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] [instance: d5d10a31-0e78-4ed7-b944-9208138a4861] Terminating instance [ 914.990676] env[61957]: DEBUG nova.compute.manager [None req-94184bcc-7df6-4480-913f-5d000fd6dd4e tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] [instance: d5d10a31-0e78-4ed7-b944-9208138a4861] Start destroying the instance on the hypervisor. {{(pid=61957) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 914.990892] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-94184bcc-7df6-4480-913f-5d000fd6dd4e tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] [instance: d5d10a31-0e78-4ed7-b944-9208138a4861] Destroying instance {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 914.992152] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bda0342-f687-4d6e-ba68-24934c56ee93 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.000137] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-94184bcc-7df6-4480-913f-5d000fd6dd4e tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] [instance: d5d10a31-0e78-4ed7-b944-9208138a4861] Powering off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 915.001299] env[61957]: INFO nova.virt.block_device [None req-24cac752-c949-48ff-b582-7f3e67bd6cce tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] [instance: c23141ee-0cbb-4d1b-8390-c3073fe354f1] Attempting to driver detach volume 734d094d-c382-4070-87c9-699fa1355b79 from mountpoint /dev/sdb [ 915.001507] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-24cac752-c949-48ff-b582-7f3e67bd6cce tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] [instance: c23141ee-0cbb-4d1b-8390-c3073fe354f1] Volume detach. Driver type: vmdk {{(pid=61957) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 915.001844] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-24cac752-c949-48ff-b582-7f3e67bd6cce tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] [instance: c23141ee-0cbb-4d1b-8390-c3073fe354f1] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-274545', 'volume_id': '734d094d-c382-4070-87c9-699fa1355b79', 'name': 'volume-734d094d-c382-4070-87c9-699fa1355b79', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'c23141ee-0cbb-4d1b-8390-c3073fe354f1', 'attached_at': '', 'detached_at': '', 'volume_id': '734d094d-c382-4070-87c9-699fa1355b79', 'serial': '734d094d-c382-4070-87c9-699fa1355b79'} {{(pid=61957) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 915.002094] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-63e3752d-b498-4c1f-9900-9a8dc43580a1 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.004478] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-827a10ce-fd0e-48c2-a29f-375dbb1e57f7 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.039384] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-390fce7e-74e7-4f32-8985-523901887b24 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.044107] env[61957]: DEBUG oslo_vmware.api [None req-94184bcc-7df6-4480-913f-5d000fd6dd4e tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Waiting for the task: (returnval){ [ 915.044107] env[61957]: value = "task-1277692" [ 915.044107] env[61957]: _type = "Task" [ 915.044107] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.052823] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6817669f-3ebb-455b-b302-bb2cdc880a29 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.060019] env[61957]: DEBUG oslo_vmware.api [None req-94184bcc-7df6-4480-913f-5d000fd6dd4e tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Task: {'id': task-1277692, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.088192] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bf69844-66ec-4fb5-838a-32c4e25251b0 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.109403] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-24cac752-c949-48ff-b582-7f3e67bd6cce tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] The volume has not been displaced from its original location: [datastore1] volume-734d094d-c382-4070-87c9-699fa1355b79/volume-734d094d-c382-4070-87c9-699fa1355b79.vmdk. No consolidation needed. {{(pid=61957) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 915.115303] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-24cac752-c949-48ff-b582-7f3e67bd6cce tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] [instance: c23141ee-0cbb-4d1b-8390-c3073fe354f1] Reconfiguring VM instance instance-0000003f to detach disk 2001 {{(pid=61957) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 915.117325] env[61957]: DEBUG nova.network.neutron [None req-5a91d738-be3a-4667-949d-376fc04acd42 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: d4674fbc-47b1-42d1-aaba-e86d46c51e8f] Successfully created port: 7e1da2f5-45f0-4991-bd1f-8d7115b1adbe {{(pid=61957) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 915.124542] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bb14f836-1d52-4942-817b-98366ad09f12 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.140022] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c992c95-c284-42c7-bf40-2718cda27b2f {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.153398] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277690, 'name': CreateVM_Task, 'duration_secs': 0.463902} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.154797] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 19966b0a-53b7-48c5-849c-a9d00dc024f8] Created VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 915.155859] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05287637-3888-46cf-bbba-dd65882a7aa6 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.159518] env[61957]: DEBUG oslo_vmware.api [None req-24cac752-c949-48ff-b582-7f3e67bd6cce tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Waiting for the task: (returnval){ [ 915.159518] env[61957]: value = "task-1277695" [ 915.159518] env[61957]: _type = "Task" [ 915.159518] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.160246] env[61957]: DEBUG oslo_concurrency.lockutils [None req-bb69c8c2-41e0-4fa7-aea3-e810870c1532 tempest-ServersNegativeTestMultiTenantJSON-2135399494 tempest-ServersNegativeTestMultiTenantJSON-2135399494-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 915.160419] env[61957]: DEBUG oslo_concurrency.lockutils [None req-bb69c8c2-41e0-4fa7-aea3-e810870c1532 tempest-ServersNegativeTestMultiTenantJSON-2135399494 tempest-ServersNegativeTestMultiTenantJSON-2135399494-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 915.160763] env[61957]: DEBUG oslo_concurrency.lockutils [None req-bb69c8c2-41e0-4fa7-aea3-e810870c1532 tempest-ServersNegativeTestMultiTenantJSON-2135399494 tempest-ServersNegativeTestMultiTenantJSON-2135399494-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 915.161865] env[61957]: DEBUG oslo_concurrency.lockutils [None req-9065790e-dc28-4275-a1c6-42f3b501d4d2 tempest-ServersV294TestFqdnHostnames-1426486100 tempest-ServersV294TestFqdnHostnames-1426486100-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 915.162123] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0645013d-90a3-4869-b2b8-e90d9c39c236 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.198261] env[61957]: DEBUG nova.compute.manager [None req-5a91d738-be3a-4667-949d-376fc04acd42 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: d4674fbc-47b1-42d1-aaba-e86d46c51e8f] Start building block device mappings for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 915.205747] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c37164c5-446d-4497-9064-c3035e308288 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.214520] env[61957]: DEBUG oslo_vmware.api [None req-24cac752-c949-48ff-b582-7f3e67bd6cce tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Task: {'id': task-1277695, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.214912] env[61957]: DEBUG oslo_vmware.api [None req-bb69c8c2-41e0-4fa7-aea3-e810870c1532 tempest-ServersNegativeTestMultiTenantJSON-2135399494 tempest-ServersNegativeTestMultiTenantJSON-2135399494-project-member] Waiting for the task: (returnval){ [ 915.214912] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]5296922b-ebc7-e06b-2c36-973ebe8c716b" [ 915.214912] env[61957]: _type = "Task" [ 915.214912] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.223187] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-013c315c-a045-4f0a-8fed-14bd63f89829 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.237309] env[61957]: DEBUG oslo_vmware.api [None req-bb69c8c2-41e0-4fa7-aea3-e810870c1532 tempest-ServersNegativeTestMultiTenantJSON-2135399494 tempest-ServersNegativeTestMultiTenantJSON-2135399494-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]5296922b-ebc7-e06b-2c36-973ebe8c716b, 'name': SearchDatastore_Task, 'duration_secs': 0.014614} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.241936] env[61957]: DEBUG oslo_concurrency.lockutils [None req-bb69c8c2-41e0-4fa7-aea3-e810870c1532 tempest-ServersNegativeTestMultiTenantJSON-2135399494 tempest-ServersNegativeTestMultiTenantJSON-2135399494-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 915.242291] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-bb69c8c2-41e0-4fa7-aea3-e810870c1532 tempest-ServersNegativeTestMultiTenantJSON-2135399494 tempest-ServersNegativeTestMultiTenantJSON-2135399494-project-member] [instance: 19966b0a-53b7-48c5-849c-a9d00dc024f8] Processing image 11c76a2c-f705-470a-ba9d-4657858bab38 {{(pid=61957) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 915.242566] env[61957]: DEBUG oslo_concurrency.lockutils [None req-bb69c8c2-41e0-4fa7-aea3-e810870c1532 tempest-ServersNegativeTestMultiTenantJSON-2135399494 tempest-ServersNegativeTestMultiTenantJSON-2135399494-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 915.242736] env[61957]: DEBUG oslo_concurrency.lockutils [None req-bb69c8c2-41e0-4fa7-aea3-e810870c1532 tempest-ServersNegativeTestMultiTenantJSON-2135399494 tempest-ServersNegativeTestMultiTenantJSON-2135399494-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 915.242976] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-bb69c8c2-41e0-4fa7-aea3-e810870c1532 tempest-ServersNegativeTestMultiTenantJSON-2135399494 tempest-ServersNegativeTestMultiTenantJSON-2135399494-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 915.244316] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-42999ebb-3603-4915-8447-7b6c0e6cfa6b {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.256485] env[61957]: DEBUG nova.compute.provider_tree [None req-9513c200-94de-43f4-aaa5-418ecbddf25f tempest-ServerTagsTestJSON-969152625 tempest-ServerTagsTestJSON-969152625-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 915.263270] env[61957]: DEBUG oslo_vmware.api [None req-a7399dbc-0983-4a91-82d1-9c1b173d4f53 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1277687, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.598625} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.263667] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-a7399dbc-0983-4a91-82d1-9c1b173d4f53 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore1] 39b8b1df-efdd-4c22-9f31-85c85be4f0eb/39b8b1df-efdd-4c22-9f31-85c85be4f0eb.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 915.263871] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-a7399dbc-0983-4a91-82d1-9c1b173d4f53 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 39b8b1df-efdd-4c22-9f31-85c85be4f0eb] Extending root virtual disk to 1048576 {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 915.264293] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c9920cbd-ff83-42e7-9890-6d17de50128f {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.268533] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-bb69c8c2-41e0-4fa7-aea3-e810870c1532 tempest-ServersNegativeTestMultiTenantJSON-2135399494 tempest-ServersNegativeTestMultiTenantJSON-2135399494-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 915.268728] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-bb69c8c2-41e0-4fa7-aea3-e810870c1532 tempest-ServersNegativeTestMultiTenantJSON-2135399494 tempest-ServersNegativeTestMultiTenantJSON-2135399494-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61957) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 915.269871] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e082b3b6-ce5e-45a8-b7bd-854c0639e8d7 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.275820] env[61957]: DEBUG oslo_vmware.api [None req-a7399dbc-0983-4a91-82d1-9c1b173d4f53 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Waiting for the task: (returnval){ [ 915.275820] env[61957]: value = "task-1277696" [ 915.275820] env[61957]: _type = "Task" [ 915.275820] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.278331] env[61957]: DEBUG oslo_vmware.api [None req-bb69c8c2-41e0-4fa7-aea3-e810870c1532 tempest-ServersNegativeTestMultiTenantJSON-2135399494 tempest-ServersNegativeTestMultiTenantJSON-2135399494-project-member] Waiting for the task: (returnval){ [ 915.278331] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]521f70a9-ad9c-76e0-5b7f-9babc09df82c" [ 915.278331] env[61957]: _type = "Task" [ 915.278331] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.297594] env[61957]: DEBUG oslo_vmware.api [None req-a7399dbc-0983-4a91-82d1-9c1b173d4f53 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1277696, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.302896] env[61957]: DEBUG oslo_vmware.api [None req-bb69c8c2-41e0-4fa7-aea3-e810870c1532 tempest-ServersNegativeTestMultiTenantJSON-2135399494 tempest-ServersNegativeTestMultiTenantJSON-2135399494-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]521f70a9-ad9c-76e0-5b7f-9babc09df82c, 'name': SearchDatastore_Task, 'duration_secs': 0.015686} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.303456] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ef4a9442-69be-4451-a824-5445a6437ea5 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.313339] env[61957]: DEBUG oslo_vmware.api [None req-bb69c8c2-41e0-4fa7-aea3-e810870c1532 tempest-ServersNegativeTestMultiTenantJSON-2135399494 tempest-ServersNegativeTestMultiTenantJSON-2135399494-project-member] Waiting for the task: (returnval){ [ 915.313339] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]527feba5-d967-35d6-722f-81c1394df984" [ 915.313339] env[61957]: _type = "Task" [ 915.313339] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.324122] env[61957]: DEBUG oslo_vmware.api [None req-bb69c8c2-41e0-4fa7-aea3-e810870c1532 tempest-ServersNegativeTestMultiTenantJSON-2135399494 tempest-ServersNegativeTestMultiTenantJSON-2135399494-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]527feba5-d967-35d6-722f-81c1394df984, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.383266] env[61957]: DEBUG oslo_vmware.api [None req-3c689a75-14e7-47a1-b918-cef3025704e3 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': task-1277686, 'name': PowerOnVM_Task, 'duration_secs': 1.931151} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.383266] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c689a75-14e7-47a1-b918-cef3025704e3 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: dbeeb200-70b9-4cb4-b5a4-182389d21918] Powered on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 915.383461] env[61957]: INFO nova.compute.manager [None req-3c689a75-14e7-47a1-b918-cef3025704e3 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: dbeeb200-70b9-4cb4-b5a4-182389d21918] Took 10.05 seconds to spawn the instance on the hypervisor. [ 915.383607] env[61957]: DEBUG nova.compute.manager [None req-3c689a75-14e7-47a1-b918-cef3025704e3 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: dbeeb200-70b9-4cb4-b5a4-182389d21918] Checking state {{(pid=61957) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 915.384511] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94ea74c1-fa2f-40c0-b5cf-52cb325a327a {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.427338] env[61957]: DEBUG nova.network.neutron [None req-63a396d5-1afb-44c8-a210-9949507fdc49 tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] [instance: a8ca1ad6-d636-4fa2-b0ac-53b020e392ce] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 915.555861] env[61957]: DEBUG oslo_vmware.api [None req-94184bcc-7df6-4480-913f-5d000fd6dd4e tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Task: {'id': task-1277692, 'name': PowerOffVM_Task, 'duration_secs': 0.287471} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.556324] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-94184bcc-7df6-4480-913f-5d000fd6dd4e tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] [instance: d5d10a31-0e78-4ed7-b944-9208138a4861] Powered off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 915.556324] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-94184bcc-7df6-4480-913f-5d000fd6dd4e tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] [instance: d5d10a31-0e78-4ed7-b944-9208138a4861] Unregistering the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 915.556643] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3ab6fe0e-3b80-4190-bc7f-695b2476240a {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.657015] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-94184bcc-7df6-4480-913f-5d000fd6dd4e tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] [instance: d5d10a31-0e78-4ed7-b944-9208138a4861] Unregistered the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 915.657391] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-94184bcc-7df6-4480-913f-5d000fd6dd4e tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] [instance: d5d10a31-0e78-4ed7-b944-9208138a4861] Deleting contents of the VM from datastore datastore2 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 915.657682] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-94184bcc-7df6-4480-913f-5d000fd6dd4e tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Deleting the datastore file [datastore2] d5d10a31-0e78-4ed7-b944-9208138a4861 {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 915.659461] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2b660f29-e2aa-4e65-acd7-b7c1fddeb776 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.668479] env[61957]: DEBUG oslo_vmware.api [None req-94184bcc-7df6-4480-913f-5d000fd6dd4e tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Waiting for the task: (returnval){ [ 915.668479] env[61957]: value = "task-1277698" [ 915.668479] env[61957]: _type = "Task" [ 915.668479] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.683853] env[61957]: DEBUG oslo_vmware.api [None req-24cac752-c949-48ff-b582-7f3e67bd6cce tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Task: {'id': task-1277695, 'name': ReconfigVM_Task, 'duration_secs': 0.286199} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.689158] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-24cac752-c949-48ff-b582-7f3e67bd6cce tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] [instance: c23141ee-0cbb-4d1b-8390-c3073fe354f1] Reconfigured VM instance instance-0000003f to detach disk 2001 {{(pid=61957) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 915.696961] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3419f592-4a32-4a48-af8d-6ff4eb8ae5d4 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.725209] env[61957]: DEBUG oslo_vmware.api [None req-94184bcc-7df6-4480-913f-5d000fd6dd4e tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Task: {'id': task-1277698, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.732474] env[61957]: DEBUG oslo_vmware.api [None req-24cac752-c949-48ff-b582-7f3e67bd6cce tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Waiting for the task: (returnval){ [ 915.732474] env[61957]: value = "task-1277699" [ 915.732474] env[61957]: _type = "Task" [ 915.732474] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.733583] env[61957]: DEBUG nova.network.neutron [None req-63a396d5-1afb-44c8-a210-9949507fdc49 tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] [instance: a8ca1ad6-d636-4fa2-b0ac-53b020e392ce] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 915.746096] env[61957]: DEBUG oslo_vmware.api [None req-24cac752-c949-48ff-b582-7f3e67bd6cce tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Task: {'id': task-1277699, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.764321] env[61957]: DEBUG nova.scheduler.client.report [None req-9513c200-94de-43f4-aaa5-418ecbddf25f tempest-ServerTagsTestJSON-969152625 tempest-ServerTagsTestJSON-969152625-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 915.791029] env[61957]: DEBUG nova.network.neutron [req-b4defd35-5ba1-4901-aec9-0a4f56ed7b79 req-25205183-5134-4679-bdb0-0658bc4ff8c4 service nova] [instance: 19966b0a-53b7-48c5-849c-a9d00dc024f8] Updated VIF entry in instance network info cache for port 159b1e72-9328-460d-b334-9a3ad5d73b44. {{(pid=61957) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 915.791541] env[61957]: DEBUG nova.network.neutron [req-b4defd35-5ba1-4901-aec9-0a4f56ed7b79 req-25205183-5134-4679-bdb0-0658bc4ff8c4 service nova] [instance: 19966b0a-53b7-48c5-849c-a9d00dc024f8] Updating instance_info_cache with network_info: [{"id": "159b1e72-9328-460d-b334-9a3ad5d73b44", "address": "fa:16:3e:69:8b:7f", "network": {"id": "6a4b928d-7db0-4a1f-ba48-5af87be6c39a", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-1923984009-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "219263b31ffc43a0904d8e620e48e085", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c4349e30-c086-4c24-9e0e-83996d808a1b", "external-id": "nsx-vlan-transportzone-266", "segmentation_id": 266, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap159b1e72-93", "ovs_interfaceid": "159b1e72-9328-460d-b334-9a3ad5d73b44", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 915.796019] env[61957]: DEBUG oslo_vmware.api [None req-a7399dbc-0983-4a91-82d1-9c1b173d4f53 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1277696, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.081132} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.796019] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-a7399dbc-0983-4a91-82d1-9c1b173d4f53 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 39b8b1df-efdd-4c22-9f31-85c85be4f0eb] Extended root virtual disk {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 915.796019] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e8961d0-a621-4b07-b1e4-8893e04e74a0 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.823855] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-a7399dbc-0983-4a91-82d1-9c1b173d4f53 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 39b8b1df-efdd-4c22-9f31-85c85be4f0eb] Reconfiguring VM instance instance-0000004c to attach disk [datastore1] 39b8b1df-efdd-4c22-9f31-85c85be4f0eb/39b8b1df-efdd-4c22-9f31-85c85be4f0eb.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 915.823855] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2d5ef59b-ffdd-4d64-9105-9b647eea4761 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.849086] env[61957]: DEBUG oslo_vmware.api [None req-bb69c8c2-41e0-4fa7-aea3-e810870c1532 tempest-ServersNegativeTestMultiTenantJSON-2135399494 tempest-ServersNegativeTestMultiTenantJSON-2135399494-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]527feba5-d967-35d6-722f-81c1394df984, 'name': SearchDatastore_Task, 'duration_secs': 0.013811} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.850441] env[61957]: DEBUG oslo_concurrency.lockutils [None req-bb69c8c2-41e0-4fa7-aea3-e810870c1532 tempest-ServersNegativeTestMultiTenantJSON-2135399494 tempest-ServersNegativeTestMultiTenantJSON-2135399494-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 915.850709] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-bb69c8c2-41e0-4fa7-aea3-e810870c1532 tempest-ServersNegativeTestMultiTenantJSON-2135399494 tempest-ServersNegativeTestMultiTenantJSON-2135399494-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore2] 19966b0a-53b7-48c5-849c-a9d00dc024f8/19966b0a-53b7-48c5-849c-a9d00dc024f8.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 915.851058] env[61957]: DEBUG oslo_vmware.api [None req-a7399dbc-0983-4a91-82d1-9c1b173d4f53 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Waiting for the task: (returnval){ [ 915.851058] env[61957]: value = "task-1277700" [ 915.851058] env[61957]: _type = "Task" [ 915.851058] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.851285] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ea33eed9-2995-4967-a6b0-f2a31538f3fe {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.861902] env[61957]: DEBUG oslo_vmware.api [None req-a7399dbc-0983-4a91-82d1-9c1b173d4f53 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1277700, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.863289] env[61957]: DEBUG oslo_vmware.api [None req-bb69c8c2-41e0-4fa7-aea3-e810870c1532 tempest-ServersNegativeTestMultiTenantJSON-2135399494 tempest-ServersNegativeTestMultiTenantJSON-2135399494-project-member] Waiting for the task: (returnval){ [ 915.863289] env[61957]: value = "task-1277701" [ 915.863289] env[61957]: _type = "Task" [ 915.863289] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.874345] env[61957]: DEBUG oslo_vmware.api [None req-bb69c8c2-41e0-4fa7-aea3-e810870c1532 tempest-ServersNegativeTestMultiTenantJSON-2135399494 tempest-ServersNegativeTestMultiTenantJSON-2135399494-project-member] Task: {'id': task-1277701, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.905959] env[61957]: INFO nova.compute.manager [None req-3c689a75-14e7-47a1-b918-cef3025704e3 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: dbeeb200-70b9-4cb4-b5a4-182389d21918] Took 40.65 seconds to build instance. [ 916.181858] env[61957]: DEBUG oslo_vmware.api [None req-94184bcc-7df6-4480-913f-5d000fd6dd4e tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Task: {'id': task-1277698, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.342434} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.182400] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-94184bcc-7df6-4480-913f-5d000fd6dd4e tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Deleted the datastore file {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 916.182400] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-94184bcc-7df6-4480-913f-5d000fd6dd4e tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] [instance: d5d10a31-0e78-4ed7-b944-9208138a4861] Deleted contents of the VM from datastore datastore2 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 916.182642] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-94184bcc-7df6-4480-913f-5d000fd6dd4e tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] [instance: d5d10a31-0e78-4ed7-b944-9208138a4861] Instance destroyed {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 916.182949] env[61957]: INFO nova.compute.manager [None req-94184bcc-7df6-4480-913f-5d000fd6dd4e tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] [instance: d5d10a31-0e78-4ed7-b944-9208138a4861] Took 1.19 seconds to destroy the instance on the hypervisor. [ 916.183202] env[61957]: DEBUG oslo.service.loopingcall [None req-94184bcc-7df6-4480-913f-5d000fd6dd4e tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 916.183462] env[61957]: DEBUG nova.compute.manager [-] [instance: d5d10a31-0e78-4ed7-b944-9208138a4861] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 916.183580] env[61957]: DEBUG nova.network.neutron [-] [instance: d5d10a31-0e78-4ed7-b944-9208138a4861] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 916.228916] env[61957]: DEBUG nova.compute.manager [None req-5a91d738-be3a-4667-949d-376fc04acd42 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: d4674fbc-47b1-42d1-aaba-e86d46c51e8f] Start spawning the instance on the hypervisor. {{(pid=61957) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 916.240362] env[61957]: DEBUG oslo_concurrency.lockutils [None req-63a396d5-1afb-44c8-a210-9949507fdc49 tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Releasing lock "refresh_cache-a8ca1ad6-d636-4fa2-b0ac-53b020e392ce" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 916.240954] env[61957]: DEBUG nova.compute.manager [None req-63a396d5-1afb-44c8-a210-9949507fdc49 tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] [instance: a8ca1ad6-d636-4fa2-b0ac-53b020e392ce] Start destroying the instance on the hypervisor. {{(pid=61957) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 916.241255] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-63a396d5-1afb-44c8-a210-9949507fdc49 tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] [instance: a8ca1ad6-d636-4fa2-b0ac-53b020e392ce] Destroying instance {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 916.242278] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efd78940-4dc6-4cbf-97ff-3edf5b018fae {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.252085] env[61957]: DEBUG oslo_vmware.api [None req-24cac752-c949-48ff-b582-7f3e67bd6cce tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Task: {'id': task-1277699, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.258297] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-63a396d5-1afb-44c8-a210-9949507fdc49 tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] [instance: a8ca1ad6-d636-4fa2-b0ac-53b020e392ce] Powering off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 916.258873] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9e985ad5-0fd9-4027-863f-e95086c9e884 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.266921] env[61957]: DEBUG nova.virt.hardware [None req-5a91d738-be3a-4667-949d-376fc04acd42 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T17:22:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T17:21:50Z,direct_url=,disk_format='vmdk',id=11c76a2c-f705-470a-ba9d-4657858bab38,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='81d3096d0f094373913fc3dc8f5c3c6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T17:21:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 916.267220] env[61957]: DEBUG nova.virt.hardware [None req-5a91d738-be3a-4667-949d-376fc04acd42 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Flavor limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 916.267381] env[61957]: DEBUG nova.virt.hardware [None req-5a91d738-be3a-4667-949d-376fc04acd42 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Image limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 916.267599] env[61957]: DEBUG nova.virt.hardware [None req-5a91d738-be3a-4667-949d-376fc04acd42 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Flavor pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 916.267773] env[61957]: DEBUG nova.virt.hardware [None req-5a91d738-be3a-4667-949d-376fc04acd42 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Image pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 916.267881] env[61957]: DEBUG nova.virt.hardware [None req-5a91d738-be3a-4667-949d-376fc04acd42 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 916.268113] env[61957]: DEBUG nova.virt.hardware [None req-5a91d738-be3a-4667-949d-376fc04acd42 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 916.268281] env[61957]: DEBUG nova.virt.hardware [None req-5a91d738-be3a-4667-949d-376fc04acd42 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 916.268453] env[61957]: DEBUG nova.virt.hardware [None req-5a91d738-be3a-4667-949d-376fc04acd42 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Got 1 possible topologies {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 916.268622] env[61957]: DEBUG nova.virt.hardware [None req-5a91d738-be3a-4667-949d-376fc04acd42 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 916.268806] env[61957]: DEBUG nova.virt.hardware [None req-5a91d738-be3a-4667-949d-376fc04acd42 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 916.269648] env[61957]: DEBUG oslo_concurrency.lockutils [None req-9513c200-94de-43f4-aaa5-418ecbddf25f tempest-ServerTagsTestJSON-969152625 tempest-ServerTagsTestJSON-969152625-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.099s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 916.272495] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f1acba2-66fe-48ff-8ce6-8df989cec8e4 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.276771] env[61957]: DEBUG oslo_concurrency.lockutils [None req-a3e467c9-b0de-4881-ac92-2329baed7e90 tempest-ServersTestJSON-1093633677 tempest-ServersTestJSON-1093633677-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 29.531s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 916.277045] env[61957]: DEBUG nova.objects.instance [None req-a3e467c9-b0de-4881-ac92-2329baed7e90 tempest-ServersTestJSON-1093633677 tempest-ServersTestJSON-1093633677-project-member] Lazy-loading 'resources' on Instance uuid ead6aae1-36b5-4f57-9129-3bb02cf103ce {{(pid=61957) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 916.278388] env[61957]: DEBUG oslo_vmware.api [None req-63a396d5-1afb-44c8-a210-9949507fdc49 tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Waiting for the task: (returnval){ [ 916.278388] env[61957]: value = "task-1277702" [ 916.278388] env[61957]: _type = "Task" [ 916.278388] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.286743] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24151123-5b01-4eb7-a3f0-f645cdd1ffa5 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.297211] env[61957]: DEBUG oslo_concurrency.lockutils [req-b4defd35-5ba1-4901-aec9-0a4f56ed7b79 req-25205183-5134-4679-bdb0-0658bc4ff8c4 service nova] Releasing lock "refresh_cache-19966b0a-53b7-48c5-849c-a9d00dc024f8" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 916.297211] env[61957]: DEBUG nova.compute.manager [req-b4defd35-5ba1-4901-aec9-0a4f56ed7b79 req-25205183-5134-4679-bdb0-0658bc4ff8c4 service nova] [instance: 2e4a5344-600f-4b61-826e-c15f96b50af2] Received event network-vif-deleted-a1026b09-e6b6-47f4-b5b2-567a15abfa41 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 916.297211] env[61957]: DEBUG oslo_vmware.api [None req-63a396d5-1afb-44c8-a210-9949507fdc49 tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Task: {'id': task-1277702, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.310182] env[61957]: INFO nova.scheduler.client.report [None req-9513c200-94de-43f4-aaa5-418ecbddf25f tempest-ServerTagsTestJSON-969152625 tempest-ServerTagsTestJSON-969152625-project-member] Deleted allocations for instance 8e04a098-25de-4d57-9e3c-ea44d234d57e [ 916.371025] env[61957]: DEBUG oslo_vmware.api [None req-a7399dbc-0983-4a91-82d1-9c1b173d4f53 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1277700, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.380124] env[61957]: DEBUG oslo_vmware.api [None req-bb69c8c2-41e0-4fa7-aea3-e810870c1532 tempest-ServersNegativeTestMultiTenantJSON-2135399494 tempest-ServersNegativeTestMultiTenantJSON-2135399494-project-member] Task: {'id': task-1277701, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.409268] env[61957]: DEBUG oslo_concurrency.lockutils [None req-3c689a75-14e7-47a1-b918-cef3025704e3 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Lock "dbeeb200-70b9-4cb4-b5a4-182389d21918" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 57.757s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 916.648626] env[61957]: DEBUG nova.compute.manager [req-af9be644-d299-495b-a6c2-f11fde55aaa8 req-04550a8b-2d3a-4bdf-9675-a2beba4a2d38 service nova] [instance: d5d10a31-0e78-4ed7-b944-9208138a4861] Received event network-vif-deleted-572549f2-6ec4-411e-8905-0c33119b31c8 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 916.651117] env[61957]: INFO nova.compute.manager [req-af9be644-d299-495b-a6c2-f11fde55aaa8 req-04550a8b-2d3a-4bdf-9675-a2beba4a2d38 service nova] [instance: d5d10a31-0e78-4ed7-b944-9208138a4861] Neutron deleted interface 572549f2-6ec4-411e-8905-0c33119b31c8; detaching it from the instance and deleting it from the info cache [ 916.651117] env[61957]: DEBUG nova.network.neutron [req-af9be644-d299-495b-a6c2-f11fde55aaa8 req-04550a8b-2d3a-4bdf-9675-a2beba4a2d38 service nova] [instance: d5d10a31-0e78-4ed7-b944-9208138a4861] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 916.749994] env[61957]: DEBUG oslo_vmware.api [None req-24cac752-c949-48ff-b582-7f3e67bd6cce tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Task: {'id': task-1277699, 'name': ReconfigVM_Task, 'duration_secs': 0.816649} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.749994] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-24cac752-c949-48ff-b582-7f3e67bd6cce tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] [instance: c23141ee-0cbb-4d1b-8390-c3073fe354f1] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-274545', 'volume_id': '734d094d-c382-4070-87c9-699fa1355b79', 'name': 'volume-734d094d-c382-4070-87c9-699fa1355b79', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'c23141ee-0cbb-4d1b-8390-c3073fe354f1', 'attached_at': '', 'detached_at': '', 'volume_id': '734d094d-c382-4070-87c9-699fa1355b79', 'serial': '734d094d-c382-4070-87c9-699fa1355b79'} {{(pid=61957) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 916.792510] env[61957]: DEBUG oslo_vmware.api [None req-63a396d5-1afb-44c8-a210-9949507fdc49 tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Task: {'id': task-1277702, 'name': PowerOffVM_Task, 'duration_secs': 0.289729} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.793577] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-63a396d5-1afb-44c8-a210-9949507fdc49 tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] [instance: a8ca1ad6-d636-4fa2-b0ac-53b020e392ce] Powered off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 916.794016] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-63a396d5-1afb-44c8-a210-9949507fdc49 tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] [instance: a8ca1ad6-d636-4fa2-b0ac-53b020e392ce] Unregistering the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 916.794572] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7b75c1bf-fef6-4c71-b333-085f4e8a8346 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.823022] env[61957]: DEBUG oslo_concurrency.lockutils [None req-9513c200-94de-43f4-aaa5-418ecbddf25f tempest-ServerTagsTestJSON-969152625 tempest-ServerTagsTestJSON-969152625-project-member] Lock "8e04a098-25de-4d57-9e3c-ea44d234d57e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 34.370s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 916.823022] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-63a396d5-1afb-44c8-a210-9949507fdc49 tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] [instance: a8ca1ad6-d636-4fa2-b0ac-53b020e392ce] Unregistered the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 916.823022] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-63a396d5-1afb-44c8-a210-9949507fdc49 tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] [instance: a8ca1ad6-d636-4fa2-b0ac-53b020e392ce] Deleting contents of the VM from datastore datastore1 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 916.823022] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-63a396d5-1afb-44c8-a210-9949507fdc49 tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Deleting the datastore file [datastore1] a8ca1ad6-d636-4fa2-b0ac-53b020e392ce {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 916.823022] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-98932695-653f-4903-8814-606443f0e975 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.831173] env[61957]: DEBUG oslo_vmware.api [None req-63a396d5-1afb-44c8-a210-9949507fdc49 tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Waiting for the task: (returnval){ [ 916.831173] env[61957]: value = "task-1277704" [ 916.831173] env[61957]: _type = "Task" [ 916.831173] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.841911] env[61957]: DEBUG oslo_vmware.api [None req-63a396d5-1afb-44c8-a210-9949507fdc49 tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Task: {'id': task-1277704, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.866248] env[61957]: DEBUG oslo_vmware.api [None req-a7399dbc-0983-4a91-82d1-9c1b173d4f53 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1277700, 'name': ReconfigVM_Task, 'duration_secs': 0.760155} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.866569] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-a7399dbc-0983-4a91-82d1-9c1b173d4f53 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 39b8b1df-efdd-4c22-9f31-85c85be4f0eb] Reconfigured VM instance instance-0000004c to attach disk [datastore1] 39b8b1df-efdd-4c22-9f31-85c85be4f0eb/39b8b1df-efdd-4c22-9f31-85c85be4f0eb.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 916.867376] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-af848e2f-1c97-4044-84c0-785e3dc8213f {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.882111] env[61957]: DEBUG oslo_vmware.api [None req-bb69c8c2-41e0-4fa7-aea3-e810870c1532 tempest-ServersNegativeTestMultiTenantJSON-2135399494 tempest-ServersNegativeTestMultiTenantJSON-2135399494-project-member] Task: {'id': task-1277701, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.717898} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.886081] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-bb69c8c2-41e0-4fa7-aea3-e810870c1532 tempest-ServersNegativeTestMultiTenantJSON-2135399494 tempest-ServersNegativeTestMultiTenantJSON-2135399494-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore2] 19966b0a-53b7-48c5-849c-a9d00dc024f8/19966b0a-53b7-48c5-849c-a9d00dc024f8.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 916.887477] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-bb69c8c2-41e0-4fa7-aea3-e810870c1532 tempest-ServersNegativeTestMultiTenantJSON-2135399494 tempest-ServersNegativeTestMultiTenantJSON-2135399494-project-member] [instance: 19966b0a-53b7-48c5-849c-a9d00dc024f8] Extending root virtual disk to 1048576 {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 916.887477] env[61957]: DEBUG oslo_vmware.api [None req-a7399dbc-0983-4a91-82d1-9c1b173d4f53 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Waiting for the task: (returnval){ [ 916.887477] env[61957]: value = "task-1277705" [ 916.887477] env[61957]: _type = "Task" [ 916.887477] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.887477] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-584e2a0d-b224-4d08-9357-4ff2a10a584b {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.898567] env[61957]: DEBUG oslo_vmware.api [None req-a7399dbc-0983-4a91-82d1-9c1b173d4f53 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1277705, 'name': Rename_Task} progress is 6%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.900395] env[61957]: DEBUG oslo_vmware.api [None req-bb69c8c2-41e0-4fa7-aea3-e810870c1532 tempest-ServersNegativeTestMultiTenantJSON-2135399494 tempest-ServersNegativeTestMultiTenantJSON-2135399494-project-member] Waiting for the task: (returnval){ [ 916.900395] env[61957]: value = "task-1277706" [ 916.900395] env[61957]: _type = "Task" [ 916.900395] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.915901] env[61957]: DEBUG oslo_vmware.api [None req-bb69c8c2-41e0-4fa7-aea3-e810870c1532 tempest-ServersNegativeTestMultiTenantJSON-2135399494 tempest-ServersNegativeTestMultiTenantJSON-2135399494-project-member] Task: {'id': task-1277706, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.027137] env[61957]: DEBUG nova.network.neutron [-] [instance: d5d10a31-0e78-4ed7-b944-9208138a4861] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 917.155471] env[61957]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9b79a6d3-22eb-4252-abf4-60051f42b139 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.158701] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f60c98c-2593-42b6-af4c-fe48a29da4f5 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.166524] env[61957]: DEBUG nova.network.neutron [None req-5a91d738-be3a-4667-949d-376fc04acd42 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: d4674fbc-47b1-42d1-aaba-e86d46c51e8f] Successfully updated port: 7e1da2f5-45f0-4991-bd1f-8d7115b1adbe {{(pid=61957) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 917.173635] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7cc8b87-5a48-4cb9-aa6d-b07fa79c5f04 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.181076] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74430dd7-55fc-4f75-8662-09cb63d31690 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.229744] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7a6406a-10da-4048-a8ef-6747f13ffb15 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.232194] env[61957]: DEBUG nova.compute.manager [req-af9be644-d299-495b-a6c2-f11fde55aaa8 req-04550a8b-2d3a-4bdf-9675-a2beba4a2d38 service nova] [instance: d5d10a31-0e78-4ed7-b944-9208138a4861] Detach interface failed, port_id=572549f2-6ec4-411e-8905-0c33119b31c8, reason: Instance d5d10a31-0e78-4ed7-b944-9208138a4861 could not be found. {{(pid=61957) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 917.238664] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d9971c1-dd72-428d-8e1f-ab7643b223ee {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.253208] env[61957]: DEBUG nova.compute.provider_tree [None req-a3e467c9-b0de-4881-ac92-2329baed7e90 tempest-ServersTestJSON-1093633677 tempest-ServersTestJSON-1093633677-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 917.304864] env[61957]: DEBUG nova.objects.instance [None req-24cac752-c949-48ff-b582-7f3e67bd6cce tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Lazy-loading 'flavor' on Instance uuid c23141ee-0cbb-4d1b-8390-c3073fe354f1 {{(pid=61957) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 917.343509] env[61957]: DEBUG oslo_vmware.api [None req-63a396d5-1afb-44c8-a210-9949507fdc49 tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Task: {'id': task-1277704, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.275165} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.343786] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-63a396d5-1afb-44c8-a210-9949507fdc49 tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Deleted the datastore file {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 917.343972] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-63a396d5-1afb-44c8-a210-9949507fdc49 tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] [instance: a8ca1ad6-d636-4fa2-b0ac-53b020e392ce] Deleted contents of the VM from datastore datastore1 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 917.344162] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-63a396d5-1afb-44c8-a210-9949507fdc49 tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] [instance: a8ca1ad6-d636-4fa2-b0ac-53b020e392ce] Instance destroyed {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 917.344332] env[61957]: INFO nova.compute.manager [None req-63a396d5-1afb-44c8-a210-9949507fdc49 tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] [instance: a8ca1ad6-d636-4fa2-b0ac-53b020e392ce] Took 1.10 seconds to destroy the instance on the hypervisor. [ 917.344593] env[61957]: DEBUG oslo.service.loopingcall [None req-63a396d5-1afb-44c8-a210-9949507fdc49 tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 917.344752] env[61957]: DEBUG nova.compute.manager [-] [instance: a8ca1ad6-d636-4fa2-b0ac-53b020e392ce] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 917.344848] env[61957]: DEBUG nova.network.neutron [-] [instance: a8ca1ad6-d636-4fa2-b0ac-53b020e392ce] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 917.360686] env[61957]: DEBUG nova.network.neutron [-] [instance: a8ca1ad6-d636-4fa2-b0ac-53b020e392ce] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 917.400308] env[61957]: DEBUG oslo_vmware.api [None req-a7399dbc-0983-4a91-82d1-9c1b173d4f53 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1277705, 'name': Rename_Task, 'duration_secs': 0.437741} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.400607] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-a7399dbc-0983-4a91-82d1-9c1b173d4f53 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 39b8b1df-efdd-4c22-9f31-85c85be4f0eb] Powering on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 917.401241] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-beaa5812-1c2c-4ce5-a224-1e37b49786d6 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.408847] env[61957]: DEBUG oslo_vmware.api [None req-a7399dbc-0983-4a91-82d1-9c1b173d4f53 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Waiting for the task: (returnval){ [ 917.408847] env[61957]: value = "task-1277707" [ 917.408847] env[61957]: _type = "Task" [ 917.408847] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.414509] env[61957]: DEBUG oslo_vmware.api [None req-bb69c8c2-41e0-4fa7-aea3-e810870c1532 tempest-ServersNegativeTestMultiTenantJSON-2135399494 tempest-ServersNegativeTestMultiTenantJSON-2135399494-project-member] Task: {'id': task-1277706, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072901} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.415169] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-bb69c8c2-41e0-4fa7-aea3-e810870c1532 tempest-ServersNegativeTestMultiTenantJSON-2135399494 tempest-ServersNegativeTestMultiTenantJSON-2135399494-project-member] [instance: 19966b0a-53b7-48c5-849c-a9d00dc024f8] Extended root virtual disk {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 917.415958] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-531a6280-1beb-4f5d-8c63-5d2990266eb5 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.421373] env[61957]: DEBUG oslo_vmware.api [None req-a7399dbc-0983-4a91-82d1-9c1b173d4f53 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1277707, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.442284] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-bb69c8c2-41e0-4fa7-aea3-e810870c1532 tempest-ServersNegativeTestMultiTenantJSON-2135399494 tempest-ServersNegativeTestMultiTenantJSON-2135399494-project-member] [instance: 19966b0a-53b7-48c5-849c-a9d00dc024f8] Reconfiguring VM instance instance-0000004d to attach disk [datastore2] 19966b0a-53b7-48c5-849c-a9d00dc024f8/19966b0a-53b7-48c5-849c-a9d00dc024f8.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 917.442284] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0c42c847-6a78-409e-8f55-30ec21635a02 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.462685] env[61957]: DEBUG oslo_vmware.api [None req-bb69c8c2-41e0-4fa7-aea3-e810870c1532 tempest-ServersNegativeTestMultiTenantJSON-2135399494 tempest-ServersNegativeTestMultiTenantJSON-2135399494-project-member] Waiting for the task: (returnval){ [ 917.462685] env[61957]: value = "task-1277708" [ 917.462685] env[61957]: _type = "Task" [ 917.462685] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.473340] env[61957]: DEBUG oslo_vmware.api [None req-bb69c8c2-41e0-4fa7-aea3-e810870c1532 tempest-ServersNegativeTestMultiTenantJSON-2135399494 tempest-ServersNegativeTestMultiTenantJSON-2135399494-project-member] Task: {'id': task-1277708, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.532863] env[61957]: INFO nova.compute.manager [-] [instance: d5d10a31-0e78-4ed7-b944-9208138a4861] Took 1.35 seconds to deallocate network for instance. [ 917.674112] env[61957]: DEBUG oslo_concurrency.lockutils [None req-5a91d738-be3a-4667-949d-376fc04acd42 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Acquiring lock "refresh_cache-d4674fbc-47b1-42d1-aaba-e86d46c51e8f" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 917.674423] env[61957]: DEBUG oslo_concurrency.lockutils [None req-5a91d738-be3a-4667-949d-376fc04acd42 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Acquired lock "refresh_cache-d4674fbc-47b1-42d1-aaba-e86d46c51e8f" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 917.674534] env[61957]: DEBUG nova.network.neutron [None req-5a91d738-be3a-4667-949d-376fc04acd42 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: d4674fbc-47b1-42d1-aaba-e86d46c51e8f] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 917.756896] env[61957]: DEBUG nova.scheduler.client.report [None req-a3e467c9-b0de-4881-ac92-2329baed7e90 tempest-ServersTestJSON-1093633677 tempest-ServersTestJSON-1093633677-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 917.863234] env[61957]: DEBUG nova.network.neutron [-] [instance: a8ca1ad6-d636-4fa2-b0ac-53b020e392ce] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 917.920259] env[61957]: DEBUG oslo_vmware.api [None req-a7399dbc-0983-4a91-82d1-9c1b173d4f53 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1277707, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.973352] env[61957]: DEBUG oslo_vmware.api [None req-bb69c8c2-41e0-4fa7-aea3-e810870c1532 tempest-ServersNegativeTestMultiTenantJSON-2135399494 tempest-ServersNegativeTestMultiTenantJSON-2135399494-project-member] Task: {'id': task-1277708, 'name': ReconfigVM_Task, 'duration_secs': 0.407784} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.973706] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-bb69c8c2-41e0-4fa7-aea3-e810870c1532 tempest-ServersNegativeTestMultiTenantJSON-2135399494 tempest-ServersNegativeTestMultiTenantJSON-2135399494-project-member] [instance: 19966b0a-53b7-48c5-849c-a9d00dc024f8] Reconfigured VM instance instance-0000004d to attach disk [datastore2] 19966b0a-53b7-48c5-849c-a9d00dc024f8/19966b0a-53b7-48c5-849c-a9d00dc024f8.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 917.974411] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1b861ba0-ea2b-410e-b94e-936f1b1c18b5 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.982787] env[61957]: DEBUG oslo_vmware.api [None req-bb69c8c2-41e0-4fa7-aea3-e810870c1532 tempest-ServersNegativeTestMultiTenantJSON-2135399494 tempest-ServersNegativeTestMultiTenantJSON-2135399494-project-member] Waiting for the task: (returnval){ [ 917.982787] env[61957]: value = "task-1277709" [ 917.982787] env[61957]: _type = "Task" [ 917.982787] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.992800] env[61957]: DEBUG oslo_vmware.api [None req-bb69c8c2-41e0-4fa7-aea3-e810870c1532 tempest-ServersNegativeTestMultiTenantJSON-2135399494 tempest-ServersNegativeTestMultiTenantJSON-2135399494-project-member] Task: {'id': task-1277709, 'name': Rename_Task} progress is 5%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.040198] env[61957]: DEBUG oslo_concurrency.lockutils [None req-94184bcc-7df6-4480-913f-5d000fd6dd4e tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 918.232783] env[61957]: DEBUG nova.network.neutron [None req-5a91d738-be3a-4667-949d-376fc04acd42 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: d4674fbc-47b1-42d1-aaba-e86d46c51e8f] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 918.265082] env[61957]: DEBUG oslo_concurrency.lockutils [None req-a3e467c9-b0de-4881-ac92-2329baed7e90 tempest-ServersTestJSON-1093633677 tempest-ServersTestJSON-1093633677-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.987s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 918.267853] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d61e8152-d9da-413c-b7fd-8eed0b01a69d tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.743s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 918.269602] env[61957]: INFO nova.compute.claims [None req-d61e8152-d9da-413c-b7fd-8eed0b01a69d tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: b598b8bb-919e-4404-b264-7b76161b0f79] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 918.297276] env[61957]: INFO nova.scheduler.client.report [None req-a3e467c9-b0de-4881-ac92-2329baed7e90 tempest-ServersTestJSON-1093633677 tempest-ServersTestJSON-1093633677-project-member] Deleted allocations for instance ead6aae1-36b5-4f57-9129-3bb02cf103ce [ 918.313303] env[61957]: DEBUG oslo_concurrency.lockutils [None req-24cac752-c949-48ff-b582-7f3e67bd6cce tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Lock "c23141ee-0cbb-4d1b-8390-c3073fe354f1" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.854s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 918.356849] env[61957]: DEBUG nova.compute.manager [None req-23521104-b947-45c5-a62a-77ad796d9410 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: dbeeb200-70b9-4cb4-b5a4-182389d21918] Stashing vm_state: active {{(pid=61957) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 918.365650] env[61957]: INFO nova.compute.manager [-] [instance: a8ca1ad6-d636-4fa2-b0ac-53b020e392ce] Took 1.02 seconds to deallocate network for instance. [ 918.407394] env[61957]: DEBUG nova.network.neutron [None req-5a91d738-be3a-4667-949d-376fc04acd42 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: d4674fbc-47b1-42d1-aaba-e86d46c51e8f] Updating instance_info_cache with network_info: [{"id": "7e1da2f5-45f0-4991-bd1f-8d7115b1adbe", "address": "fa:16:3e:eb:f1:35", "network": {"id": "dc7d4c3b-1f91-4937-a2b0-1084e5c478f0", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1637370260-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "975189ebb3cc4cdb9391880f0c9ba6ba", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b71230ae-e879-4384-88ce-fe64c86fce22", "external-id": "nsx-vlan-transportzone-473", "segmentation_id": 473, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7e1da2f5-45", "ovs_interfaceid": "7e1da2f5-45f0-4991-bd1f-8d7115b1adbe", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 918.421268] env[61957]: DEBUG oslo_vmware.api [None req-a7399dbc-0983-4a91-82d1-9c1b173d4f53 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1277707, 'name': PowerOnVM_Task, 'duration_secs': 0.579489} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.421542] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-a7399dbc-0983-4a91-82d1-9c1b173d4f53 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 39b8b1df-efdd-4c22-9f31-85c85be4f0eb] Powered on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 918.421762] env[61957]: INFO nova.compute.manager [None req-a7399dbc-0983-4a91-82d1-9c1b173d4f53 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 39b8b1df-efdd-4c22-9f31-85c85be4f0eb] Took 8.74 seconds to spawn the instance on the hypervisor. [ 918.421941] env[61957]: DEBUG nova.compute.manager [None req-a7399dbc-0983-4a91-82d1-9c1b173d4f53 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 39b8b1df-efdd-4c22-9f31-85c85be4f0eb] Checking state {{(pid=61957) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 918.422805] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1eef58f9-8469-460b-ac31-ebe61693078d {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.493505] env[61957]: DEBUG oslo_vmware.api [None req-bb69c8c2-41e0-4fa7-aea3-e810870c1532 tempest-ServersNegativeTestMultiTenantJSON-2135399494 tempest-ServersNegativeTestMultiTenantJSON-2135399494-project-member] Task: {'id': task-1277709, 'name': Rename_Task, 'duration_secs': 0.21785} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.493890] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-bb69c8c2-41e0-4fa7-aea3-e810870c1532 tempest-ServersNegativeTestMultiTenantJSON-2135399494 tempest-ServersNegativeTestMultiTenantJSON-2135399494-project-member] [instance: 19966b0a-53b7-48c5-849c-a9d00dc024f8] Powering on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 918.494080] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ae355d40-8e23-4bdf-a486-63b858e62211 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.502427] env[61957]: DEBUG oslo_vmware.api [None req-bb69c8c2-41e0-4fa7-aea3-e810870c1532 tempest-ServersNegativeTestMultiTenantJSON-2135399494 tempest-ServersNegativeTestMultiTenantJSON-2135399494-project-member] Waiting for the task: (returnval){ [ 918.502427] env[61957]: value = "task-1277710" [ 918.502427] env[61957]: _type = "Task" [ 918.502427] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.511080] env[61957]: DEBUG oslo_vmware.api [None req-bb69c8c2-41e0-4fa7-aea3-e810870c1532 tempest-ServersNegativeTestMultiTenantJSON-2135399494 tempest-ServersNegativeTestMultiTenantJSON-2135399494-project-member] Task: {'id': task-1277710, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.807295] env[61957]: DEBUG oslo_concurrency.lockutils [None req-a3e467c9-b0de-4881-ac92-2329baed7e90 tempest-ServersTestJSON-1093633677 tempest-ServersTestJSON-1093633677-project-member] Lock "ead6aae1-36b5-4f57-9129-3bb02cf103ce" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 36.164s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 918.875868] env[61957]: DEBUG nova.compute.manager [req-25573613-106b-4d57-8ee8-2383f38ee4b8 req-23216ff4-759f-4da9-b130-1ade81db735c service nova] [instance: d4674fbc-47b1-42d1-aaba-e86d46c51e8f] Received event network-vif-plugged-7e1da2f5-45f0-4991-bd1f-8d7115b1adbe {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 918.875868] env[61957]: DEBUG oslo_concurrency.lockutils [req-25573613-106b-4d57-8ee8-2383f38ee4b8 req-23216ff4-759f-4da9-b130-1ade81db735c service nova] Acquiring lock "d4674fbc-47b1-42d1-aaba-e86d46c51e8f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 918.875868] env[61957]: DEBUG oslo_concurrency.lockutils [req-25573613-106b-4d57-8ee8-2383f38ee4b8 req-23216ff4-759f-4da9-b130-1ade81db735c service nova] Lock "d4674fbc-47b1-42d1-aaba-e86d46c51e8f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 918.876009] env[61957]: DEBUG oslo_concurrency.lockutils [req-25573613-106b-4d57-8ee8-2383f38ee4b8 req-23216ff4-759f-4da9-b130-1ade81db735c service nova] Lock "d4674fbc-47b1-42d1-aaba-e86d46c51e8f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 918.876097] env[61957]: DEBUG nova.compute.manager [req-25573613-106b-4d57-8ee8-2383f38ee4b8 req-23216ff4-759f-4da9-b130-1ade81db735c service nova] [instance: d4674fbc-47b1-42d1-aaba-e86d46c51e8f] No waiting events found dispatching network-vif-plugged-7e1da2f5-45f0-4991-bd1f-8d7115b1adbe {{(pid=61957) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 918.876273] env[61957]: WARNING nova.compute.manager [req-25573613-106b-4d57-8ee8-2383f38ee4b8 req-23216ff4-759f-4da9-b130-1ade81db735c service nova] [instance: d4674fbc-47b1-42d1-aaba-e86d46c51e8f] Received unexpected event network-vif-plugged-7e1da2f5-45f0-4991-bd1f-8d7115b1adbe for instance with vm_state building and task_state spawning. [ 918.876726] env[61957]: DEBUG nova.compute.manager [req-25573613-106b-4d57-8ee8-2383f38ee4b8 req-23216ff4-759f-4da9-b130-1ade81db735c service nova] [instance: d4674fbc-47b1-42d1-aaba-e86d46c51e8f] Received event network-changed-7e1da2f5-45f0-4991-bd1f-8d7115b1adbe {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 918.876726] env[61957]: DEBUG nova.compute.manager [req-25573613-106b-4d57-8ee8-2383f38ee4b8 req-23216ff4-759f-4da9-b130-1ade81db735c service nova] [instance: d4674fbc-47b1-42d1-aaba-e86d46c51e8f] Refreshing instance network info cache due to event network-changed-7e1da2f5-45f0-4991-bd1f-8d7115b1adbe. {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 918.876726] env[61957]: DEBUG oslo_concurrency.lockutils [req-25573613-106b-4d57-8ee8-2383f38ee4b8 req-23216ff4-759f-4da9-b130-1ade81db735c service nova] Acquiring lock "refresh_cache-d4674fbc-47b1-42d1-aaba-e86d46c51e8f" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 918.877564] env[61957]: DEBUG oslo_concurrency.lockutils [None req-63a396d5-1afb-44c8-a210-9949507fdc49 tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 918.880836] env[61957]: DEBUG oslo_concurrency.lockutils [None req-23521104-b947-45c5-a62a-77ad796d9410 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 918.910082] env[61957]: DEBUG oslo_concurrency.lockutils [None req-5a91d738-be3a-4667-949d-376fc04acd42 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Releasing lock "refresh_cache-d4674fbc-47b1-42d1-aaba-e86d46c51e8f" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 918.910464] env[61957]: DEBUG nova.compute.manager [None req-5a91d738-be3a-4667-949d-376fc04acd42 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: d4674fbc-47b1-42d1-aaba-e86d46c51e8f] Instance network_info: |[{"id": "7e1da2f5-45f0-4991-bd1f-8d7115b1adbe", "address": "fa:16:3e:eb:f1:35", "network": {"id": "dc7d4c3b-1f91-4937-a2b0-1084e5c478f0", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1637370260-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "975189ebb3cc4cdb9391880f0c9ba6ba", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b71230ae-e879-4384-88ce-fe64c86fce22", "external-id": "nsx-vlan-transportzone-473", "segmentation_id": 473, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7e1da2f5-45", "ovs_interfaceid": "7e1da2f5-45f0-4991-bd1f-8d7115b1adbe", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 918.910931] env[61957]: DEBUG oslo_concurrency.lockutils [req-25573613-106b-4d57-8ee8-2383f38ee4b8 req-23216ff4-759f-4da9-b130-1ade81db735c service nova] Acquired lock "refresh_cache-d4674fbc-47b1-42d1-aaba-e86d46c51e8f" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 918.911047] env[61957]: DEBUG nova.network.neutron [req-25573613-106b-4d57-8ee8-2383f38ee4b8 req-23216ff4-759f-4da9-b130-1ade81db735c service nova] [instance: d4674fbc-47b1-42d1-aaba-e86d46c51e8f] Refreshing network info cache for port 7e1da2f5-45f0-4991-bd1f-8d7115b1adbe {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 918.912884] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-5a91d738-be3a-4667-949d-376fc04acd42 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: d4674fbc-47b1-42d1-aaba-e86d46c51e8f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:eb:f1:35', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'b71230ae-e879-4384-88ce-fe64c86fce22', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7e1da2f5-45f0-4991-bd1f-8d7115b1adbe', 'vif_model': 'vmxnet3'}] {{(pid=61957) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 918.921886] env[61957]: DEBUG oslo.service.loopingcall [None req-5a91d738-be3a-4667-949d-376fc04acd42 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 918.924935] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d4674fbc-47b1-42d1-aaba-e86d46c51e8f] Creating VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 918.925303] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f44a5cfa-98f6-4940-a656-74f3444a6bec {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.949294] env[61957]: INFO nova.compute.manager [None req-a7399dbc-0983-4a91-82d1-9c1b173d4f53 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 39b8b1df-efdd-4c22-9f31-85c85be4f0eb] Took 41.10 seconds to build instance. [ 918.953098] env[61957]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 918.953098] env[61957]: value = "task-1277711" [ 918.953098] env[61957]: _type = "Task" [ 918.953098] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.963683] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277711, 'name': CreateVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.016165] env[61957]: DEBUG oslo_vmware.api [None req-bb69c8c2-41e0-4fa7-aea3-e810870c1532 tempest-ServersNegativeTestMultiTenantJSON-2135399494 tempest-ServersNegativeTestMultiTenantJSON-2135399494-project-member] Task: {'id': task-1277710, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.106576] env[61957]: DEBUG oslo_concurrency.lockutils [None req-3e0cd30b-989c-4521-a40d-70a59bfcf72a tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Acquiring lock "c23141ee-0cbb-4d1b-8390-c3073fe354f1" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 919.106849] env[61957]: DEBUG oslo_concurrency.lockutils [None req-3e0cd30b-989c-4521-a40d-70a59bfcf72a tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Lock "c23141ee-0cbb-4d1b-8390-c3073fe354f1" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 919.107084] env[61957]: DEBUG oslo_concurrency.lockutils [None req-3e0cd30b-989c-4521-a40d-70a59bfcf72a tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Acquiring lock "c23141ee-0cbb-4d1b-8390-c3073fe354f1-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 919.107271] env[61957]: DEBUG oslo_concurrency.lockutils [None req-3e0cd30b-989c-4521-a40d-70a59bfcf72a tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Lock "c23141ee-0cbb-4d1b-8390-c3073fe354f1-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 919.107440] env[61957]: DEBUG oslo_concurrency.lockutils [None req-3e0cd30b-989c-4521-a40d-70a59bfcf72a tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Lock "c23141ee-0cbb-4d1b-8390-c3073fe354f1-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 919.109690] env[61957]: INFO nova.compute.manager [None req-3e0cd30b-989c-4521-a40d-70a59bfcf72a tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] [instance: c23141ee-0cbb-4d1b-8390-c3073fe354f1] Terminating instance [ 919.113328] env[61957]: DEBUG nova.compute.manager [None req-3e0cd30b-989c-4521-a40d-70a59bfcf72a tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] [instance: c23141ee-0cbb-4d1b-8390-c3073fe354f1] Start destroying the instance on the hypervisor. {{(pid=61957) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 919.113328] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-3e0cd30b-989c-4521-a40d-70a59bfcf72a tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] [instance: c23141ee-0cbb-4d1b-8390-c3073fe354f1] Destroying instance {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 919.113328] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53fd70cb-4fe2-4bb4-b163-a5b9b5bd97b1 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.121470] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-3e0cd30b-989c-4521-a40d-70a59bfcf72a tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] [instance: c23141ee-0cbb-4d1b-8390-c3073fe354f1] Powering off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 919.121769] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c5c74749-7ced-42bf-a7ab-cb7edddaf5d2 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.132072] env[61957]: DEBUG oslo_vmware.api [None req-3e0cd30b-989c-4521-a40d-70a59bfcf72a tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Waiting for the task: (returnval){ [ 919.132072] env[61957]: value = "task-1277712" [ 919.132072] env[61957]: _type = "Task" [ 919.132072] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.140725] env[61957]: DEBUG oslo_vmware.api [None req-3e0cd30b-989c-4521-a40d-70a59bfcf72a tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Task: {'id': task-1277712, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.170699] env[61957]: DEBUG nova.network.neutron [req-25573613-106b-4d57-8ee8-2383f38ee4b8 req-23216ff4-759f-4da9-b130-1ade81db735c service nova] [instance: d4674fbc-47b1-42d1-aaba-e86d46c51e8f] Updated VIF entry in instance network info cache for port 7e1da2f5-45f0-4991-bd1f-8d7115b1adbe. {{(pid=61957) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 919.171427] env[61957]: DEBUG nova.network.neutron [req-25573613-106b-4d57-8ee8-2383f38ee4b8 req-23216ff4-759f-4da9-b130-1ade81db735c service nova] [instance: d4674fbc-47b1-42d1-aaba-e86d46c51e8f] Updating instance_info_cache with network_info: [{"id": "7e1da2f5-45f0-4991-bd1f-8d7115b1adbe", "address": "fa:16:3e:eb:f1:35", "network": {"id": "dc7d4c3b-1f91-4937-a2b0-1084e5c478f0", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1637370260-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "975189ebb3cc4cdb9391880f0c9ba6ba", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b71230ae-e879-4384-88ce-fe64c86fce22", "external-id": "nsx-vlan-transportzone-473", "segmentation_id": 473, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7e1da2f5-45", "ovs_interfaceid": "7e1da2f5-45f0-4991-bd1f-8d7115b1adbe", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 919.451625] env[61957]: DEBUG oslo_concurrency.lockutils [None req-a7399dbc-0983-4a91-82d1-9c1b173d4f53 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Lock "39b8b1df-efdd-4c22-9f31-85c85be4f0eb" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 57.060s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 919.469334] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277711, 'name': CreateVM_Task, 'duration_secs': 0.509853} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.469524] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d4674fbc-47b1-42d1-aaba-e86d46c51e8f] Created VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 919.470249] env[61957]: DEBUG oslo_concurrency.lockutils [None req-5a91d738-be3a-4667-949d-376fc04acd42 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 919.470423] env[61957]: DEBUG oslo_concurrency.lockutils [None req-5a91d738-be3a-4667-949d-376fc04acd42 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 919.470797] env[61957]: DEBUG oslo_concurrency.lockutils [None req-5a91d738-be3a-4667-949d-376fc04acd42 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 919.471144] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c129548f-a18c-4c61-89f6-76f18862b929 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.476642] env[61957]: DEBUG oslo_vmware.api [None req-5a91d738-be3a-4667-949d-376fc04acd42 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Waiting for the task: (returnval){ [ 919.476642] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]5250908c-854a-75b3-c12b-26f942e94341" [ 919.476642] env[61957]: _type = "Task" [ 919.476642] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.491409] env[61957]: DEBUG oslo_vmware.api [None req-5a91d738-be3a-4667-949d-376fc04acd42 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]5250908c-854a-75b3-c12b-26f942e94341, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.515060] env[61957]: DEBUG oslo_vmware.api [None req-bb69c8c2-41e0-4fa7-aea3-e810870c1532 tempest-ServersNegativeTestMultiTenantJSON-2135399494 tempest-ServersNegativeTestMultiTenantJSON-2135399494-project-member] Task: {'id': task-1277710, 'name': PowerOnVM_Task, 'duration_secs': 0.77979} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.515060] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-bb69c8c2-41e0-4fa7-aea3-e810870c1532 tempest-ServersNegativeTestMultiTenantJSON-2135399494 tempest-ServersNegativeTestMultiTenantJSON-2135399494-project-member] [instance: 19966b0a-53b7-48c5-849c-a9d00dc024f8] Powered on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 919.515060] env[61957]: INFO nova.compute.manager [None req-bb69c8c2-41e0-4fa7-aea3-e810870c1532 tempest-ServersNegativeTestMultiTenantJSON-2135399494 tempest-ServersNegativeTestMultiTenantJSON-2135399494-project-member] [instance: 19966b0a-53b7-48c5-849c-a9d00dc024f8] Took 7.39 seconds to spawn the instance on the hypervisor. [ 919.515060] env[61957]: DEBUG nova.compute.manager [None req-bb69c8c2-41e0-4fa7-aea3-e810870c1532 tempest-ServersNegativeTestMultiTenantJSON-2135399494 tempest-ServersNegativeTestMultiTenantJSON-2135399494-project-member] [instance: 19966b0a-53b7-48c5-849c-a9d00dc024f8] Checking state {{(pid=61957) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 919.515338] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a12b62d-d418-411d-b993-cd7453e01fbb {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.633358] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-689c8419-263e-4bf6-bfd4-776456add52b {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.648901] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a705453d-090e-4c12-9069-f768597c6c04 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.651992] env[61957]: DEBUG oslo_vmware.api [None req-3e0cd30b-989c-4521-a40d-70a59bfcf72a tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Task: {'id': task-1277712, 'name': PowerOffVM_Task, 'duration_secs': 0.21172} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.652284] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-3e0cd30b-989c-4521-a40d-70a59bfcf72a tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] [instance: c23141ee-0cbb-4d1b-8390-c3073fe354f1] Powered off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 919.652626] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-3e0cd30b-989c-4521-a40d-70a59bfcf72a tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] [instance: c23141ee-0cbb-4d1b-8390-c3073fe354f1] Unregistering the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 919.653201] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-639ad717-0393-4ff8-aedb-1dcffb44cb6b {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.683844] env[61957]: DEBUG oslo_concurrency.lockutils [req-25573613-106b-4d57-8ee8-2383f38ee4b8 req-23216ff4-759f-4da9-b130-1ade81db735c service nova] Releasing lock "refresh_cache-d4674fbc-47b1-42d1-aaba-e86d46c51e8f" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 919.684984] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca9a9a1d-9382-441b-a058-ea0392d17725 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.700308] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cace696a-ecb7-4419-a170-fd89384d116b {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.714636] env[61957]: DEBUG nova.compute.provider_tree [None req-d61e8152-d9da-413c-b7fd-8eed0b01a69d tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 919.728497] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-3e0cd30b-989c-4521-a40d-70a59bfcf72a tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] [instance: c23141ee-0cbb-4d1b-8390-c3073fe354f1] Unregistered the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 919.728758] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-3e0cd30b-989c-4521-a40d-70a59bfcf72a tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] [instance: c23141ee-0cbb-4d1b-8390-c3073fe354f1] Deleting contents of the VM from datastore datastore1 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 919.729063] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-3e0cd30b-989c-4521-a40d-70a59bfcf72a tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Deleting the datastore file [datastore1] c23141ee-0cbb-4d1b-8390-c3073fe354f1 {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 919.729342] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-44dce0f0-1b7f-4221-b647-861956c89856 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.736052] env[61957]: DEBUG oslo_vmware.api [None req-3e0cd30b-989c-4521-a40d-70a59bfcf72a tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Waiting for the task: (returnval){ [ 919.736052] env[61957]: value = "task-1277715" [ 919.736052] env[61957]: _type = "Task" [ 919.736052] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.744971] env[61957]: DEBUG oslo_vmware.api [None req-3e0cd30b-989c-4521-a40d-70a59bfcf72a tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Task: {'id': task-1277715, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.988583] env[61957]: DEBUG oslo_vmware.api [None req-5a91d738-be3a-4667-949d-376fc04acd42 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]5250908c-854a-75b3-c12b-26f942e94341, 'name': SearchDatastore_Task, 'duration_secs': 0.026284} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.989089] env[61957]: DEBUG oslo_concurrency.lockutils [None req-5a91d738-be3a-4667-949d-376fc04acd42 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 919.989335] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-5a91d738-be3a-4667-949d-376fc04acd42 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: d4674fbc-47b1-42d1-aaba-e86d46c51e8f] Processing image 11c76a2c-f705-470a-ba9d-4657858bab38 {{(pid=61957) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 919.989570] env[61957]: DEBUG oslo_concurrency.lockutils [None req-5a91d738-be3a-4667-949d-376fc04acd42 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 919.989717] env[61957]: DEBUG oslo_concurrency.lockutils [None req-5a91d738-be3a-4667-949d-376fc04acd42 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 919.989910] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-5a91d738-be3a-4667-949d-376fc04acd42 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 919.990187] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-271a8403-59f4-4c73-890e-0163aa63f6fd {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.999658] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-5a91d738-be3a-4667-949d-376fc04acd42 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 919.999846] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-5a91d738-be3a-4667-949d-376fc04acd42 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61957) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 920.000602] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1efa2790-8cc5-465a-aaa6-4de9bc3a3af8 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.006078] env[61957]: DEBUG oslo_vmware.api [None req-5a91d738-be3a-4667-949d-376fc04acd42 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Waiting for the task: (returnval){ [ 920.006078] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]526ca0e7-149b-94c0-f3fd-40f8c9181bde" [ 920.006078] env[61957]: _type = "Task" [ 920.006078] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.014352] env[61957]: DEBUG oslo_vmware.api [None req-5a91d738-be3a-4667-949d-376fc04acd42 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]526ca0e7-149b-94c0-f3fd-40f8c9181bde, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.040991] env[61957]: INFO nova.compute.manager [None req-bb69c8c2-41e0-4fa7-aea3-e810870c1532 tempest-ServersNegativeTestMultiTenantJSON-2135399494 tempest-ServersNegativeTestMultiTenantJSON-2135399494-project-member] [instance: 19966b0a-53b7-48c5-849c-a9d00dc024f8] Took 39.86 seconds to build instance. [ 920.218554] env[61957]: DEBUG nova.scheduler.client.report [None req-d61e8152-d9da-413c-b7fd-8eed0b01a69d tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 920.248205] env[61957]: DEBUG oslo_vmware.api [None req-3e0cd30b-989c-4521-a40d-70a59bfcf72a tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Task: {'id': task-1277715, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.46694} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.248205] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-3e0cd30b-989c-4521-a40d-70a59bfcf72a tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Deleted the datastore file {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 920.248308] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-3e0cd30b-989c-4521-a40d-70a59bfcf72a tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] [instance: c23141ee-0cbb-4d1b-8390-c3073fe354f1] Deleted contents of the VM from datastore datastore1 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 920.248469] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-3e0cd30b-989c-4521-a40d-70a59bfcf72a tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] [instance: c23141ee-0cbb-4d1b-8390-c3073fe354f1] Instance destroyed {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 920.248657] env[61957]: INFO nova.compute.manager [None req-3e0cd30b-989c-4521-a40d-70a59bfcf72a tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] [instance: c23141ee-0cbb-4d1b-8390-c3073fe354f1] Took 1.14 seconds to destroy the instance on the hypervisor. [ 920.248953] env[61957]: DEBUG oslo.service.loopingcall [None req-3e0cd30b-989c-4521-a40d-70a59bfcf72a tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 920.249176] env[61957]: DEBUG nova.compute.manager [-] [instance: c23141ee-0cbb-4d1b-8390-c3073fe354f1] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 920.249271] env[61957]: DEBUG nova.network.neutron [-] [instance: c23141ee-0cbb-4d1b-8390-c3073fe354f1] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 920.517475] env[61957]: DEBUG oslo_vmware.api [None req-5a91d738-be3a-4667-949d-376fc04acd42 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]526ca0e7-149b-94c0-f3fd-40f8c9181bde, 'name': SearchDatastore_Task, 'duration_secs': 0.011681} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.518365] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a59d7d8b-30e4-4119-93ec-fa31403b9eaa {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.524782] env[61957]: DEBUG oslo_vmware.api [None req-5a91d738-be3a-4667-949d-376fc04acd42 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Waiting for the task: (returnval){ [ 920.524782] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]5263fde6-d586-1d3e-879e-e6287cc5d15e" [ 920.524782] env[61957]: _type = "Task" [ 920.524782] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.537322] env[61957]: DEBUG oslo_vmware.api [None req-5a91d738-be3a-4667-949d-376fc04acd42 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]5263fde6-d586-1d3e-879e-e6287cc5d15e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.543137] env[61957]: DEBUG oslo_concurrency.lockutils [None req-bb69c8c2-41e0-4fa7-aea3-e810870c1532 tempest-ServersNegativeTestMultiTenantJSON-2135399494 tempest-ServersNegativeTestMultiTenantJSON-2135399494-project-member] Lock "19966b0a-53b7-48c5-849c-a9d00dc024f8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 49.719s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 920.725817] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d61e8152-d9da-413c-b7fd-8eed0b01a69d tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.458s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 920.726384] env[61957]: DEBUG nova.compute.manager [None req-d61e8152-d9da-413c-b7fd-8eed0b01a69d tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: b598b8bb-919e-4404-b264-7b76161b0f79] Start building networks asynchronously for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 920.730556] env[61957]: DEBUG oslo_concurrency.lockutils [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 28.468s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 920.993749] env[61957]: DEBUG nova.compute.manager [req-5a11457b-e31e-4f00-b113-9943a12387bf req-15a59068-2532-4200-8e0e-893c76353532 service nova] [instance: c23141ee-0cbb-4d1b-8390-c3073fe354f1] Received event network-vif-deleted-a9884dda-2fb3-4e26-84b1-877dd0e0459c {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 920.994066] env[61957]: INFO nova.compute.manager [req-5a11457b-e31e-4f00-b113-9943a12387bf req-15a59068-2532-4200-8e0e-893c76353532 service nova] [instance: c23141ee-0cbb-4d1b-8390-c3073fe354f1] Neutron deleted interface a9884dda-2fb3-4e26-84b1-877dd0e0459c; detaching it from the instance and deleting it from the info cache [ 920.994198] env[61957]: DEBUG nova.network.neutron [req-5a11457b-e31e-4f00-b113-9943a12387bf req-15a59068-2532-4200-8e0e-893c76353532 service nova] [instance: c23141ee-0cbb-4d1b-8390-c3073fe354f1] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 921.047395] env[61957]: DEBUG oslo_vmware.api [None req-5a91d738-be3a-4667-949d-376fc04acd42 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]5263fde6-d586-1d3e-879e-e6287cc5d15e, 'name': SearchDatastore_Task, 'duration_secs': 0.017117} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.047756] env[61957]: DEBUG oslo_concurrency.lockutils [None req-5a91d738-be3a-4667-949d-376fc04acd42 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 921.050019] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a91d738-be3a-4667-949d-376fc04acd42 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore2] d4674fbc-47b1-42d1-aaba-e86d46c51e8f/d4674fbc-47b1-42d1-aaba-e86d46c51e8f.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 921.050019] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-02b9300d-f21c-44b6-afd2-d51670a7bee2 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.057117] env[61957]: DEBUG oslo_vmware.api [None req-5a91d738-be3a-4667-949d-376fc04acd42 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Waiting for the task: (returnval){ [ 921.057117] env[61957]: value = "task-1277716" [ 921.057117] env[61957]: _type = "Task" [ 921.057117] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.067928] env[61957]: DEBUG oslo_vmware.api [None req-5a91d738-be3a-4667-949d-376fc04acd42 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': task-1277716, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.232374] env[61957]: DEBUG nova.compute.utils [None req-d61e8152-d9da-413c-b7fd-8eed0b01a69d tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Using /dev/sd instead of None {{(pid=61957) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 921.239330] env[61957]: DEBUG nova.compute.manager [None req-d61e8152-d9da-413c-b7fd-8eed0b01a69d tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: b598b8bb-919e-4404-b264-7b76161b0f79] Allocating IP information in the background. {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 921.239641] env[61957]: DEBUG nova.network.neutron [None req-d61e8152-d9da-413c-b7fd-8eed0b01a69d tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: b598b8bb-919e-4404-b264-7b76161b0f79] allocate_for_instance() {{(pid=61957) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 921.254579] env[61957]: DEBUG nova.network.neutron [-] [instance: c23141ee-0cbb-4d1b-8390-c3073fe354f1] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 921.309822] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d2e4cd3a-a43a-4a0c-a755-6e27eb102a3e tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Acquiring lock "39b8b1df-efdd-4c22-9f31-85c85be4f0eb" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 921.310176] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d2e4cd3a-a43a-4a0c-a755-6e27eb102a3e tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Lock "39b8b1df-efdd-4c22-9f31-85c85be4f0eb" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 921.347461] env[61957]: DEBUG nova.policy [None req-d61e8152-d9da-413c-b7fd-8eed0b01a69d tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e2f13689624b483d903ce12ef290db8b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '65fd71bcd1bf41238a9cc3a5d6dd4924', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61957) authorize /opt/stack/nova/nova/policy.py:203}} [ 921.499239] env[61957]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0bd3743b-0707-4fd9-83dd-ac2b5147df38 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.510083] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d69797a-8ac7-48cc-af2b-8792c240b488 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.539118] env[61957]: DEBUG nova.compute.manager [req-5a11457b-e31e-4f00-b113-9943a12387bf req-15a59068-2532-4200-8e0e-893c76353532 service nova] [instance: c23141ee-0cbb-4d1b-8390-c3073fe354f1] Detach interface failed, port_id=a9884dda-2fb3-4e26-84b1-877dd0e0459c, reason: Instance c23141ee-0cbb-4d1b-8390-c3073fe354f1 could not be found. {{(pid=61957) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 921.567187] env[61957]: DEBUG oslo_vmware.api [None req-5a91d738-be3a-4667-949d-376fc04acd42 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': task-1277716, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.683359] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-68adc3d5-230e-4d31-9bdb-009ac1731c2d tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] [instance: 526ff179-62a6-4763-ab25-797617c4ed57] Volume attach. Driver type: vmdk {{(pid=61957) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 921.683359] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-68adc3d5-230e-4d31-9bdb-009ac1731c2d tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] [instance: 526ff179-62a6-4763-ab25-797617c4ed57] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-274553', 'volume_id': '6c0df615-4d1a-45c6-9f48-ab1813e515b1', 'name': 'volume-6c0df615-4d1a-45c6-9f48-ab1813e515b1', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '526ff179-62a6-4763-ab25-797617c4ed57', 'attached_at': '', 'detached_at': '', 'volume_id': '6c0df615-4d1a-45c6-9f48-ab1813e515b1', 'serial': '6c0df615-4d1a-45c6-9f48-ab1813e515b1'} {{(pid=61957) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 921.683359] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b60a165e-7cd5-42c6-acad-663d3e7cbb75 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.704768] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3ab81ac-6b19-4e56-b190-7d3e36b8cf20 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.737825] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-68adc3d5-230e-4d31-9bdb-009ac1731c2d tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] [instance: 526ff179-62a6-4763-ab25-797617c4ed57] Reconfiguring VM instance instance-00000041 to attach disk [datastore1] volume-6c0df615-4d1a-45c6-9f48-ab1813e515b1/volume-6c0df615-4d1a-45c6-9f48-ab1813e515b1.vmdk or device None with type thin {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 921.738954] env[61957]: DEBUG nova.network.neutron [None req-d61e8152-d9da-413c-b7fd-8eed0b01a69d tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: b598b8bb-919e-4404-b264-7b76161b0f79] Successfully created port: be3f3b45-e91d-4c38-b746-a10c838decf7 {{(pid=61957) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 921.741068] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-97471c9b-ad34-4cc0-a4be-5b6f3623fe22 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.755897] env[61957]: DEBUG nova.compute.manager [None req-d61e8152-d9da-413c-b7fd-8eed0b01a69d tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: b598b8bb-919e-4404-b264-7b76161b0f79] Start building block device mappings for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 921.763040] env[61957]: INFO nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] [instance: dbeeb200-70b9-4cb4-b5a4-182389d21918] Updating resource usage from migration f4269166-72d4-48ca-b806-e1e29520eb19 [ 921.765240] env[61957]: INFO nova.compute.manager [-] [instance: c23141ee-0cbb-4d1b-8390-c3073fe354f1] Took 1.52 seconds to deallocate network for instance. [ 921.777603] env[61957]: DEBUG oslo_vmware.api [None req-68adc3d5-230e-4d31-9bdb-009ac1731c2d tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Waiting for the task: (returnval){ [ 921.777603] env[61957]: value = "task-1277717" [ 921.777603] env[61957]: _type = "Task" [ 921.777603] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.791126] env[61957]: DEBUG oslo_vmware.api [None req-68adc3d5-230e-4d31-9bdb-009ac1731c2d tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Task: {'id': task-1277717, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.796328] env[61957]: WARNING nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Instance 4951132e-7247-4772-8f88-3664c6a7e61e is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 921.796488] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Instance c23141ee-0cbb-4d1b-8390-c3073fe354f1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61957) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 921.796677] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Instance 526ff179-62a6-4763-ab25-797617c4ed57 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61957) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 921.796844] env[61957]: WARNING nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Instance 2e4a5344-600f-4b61-826e-c15f96b50af2 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 921.796969] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Instance e40fa112-4648-428f-a403-b3e3b8319ea8 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61957) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 921.797098] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Instance 1a0327c2-5671-4970-9db7-c7cc912d8678 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61957) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 921.797212] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Instance 28ab7d23-487f-4ae1-8fe4-58db55b59918 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61957) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 921.797690] env[61957]: WARNING nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Instance d5d10a31-0e78-4ed7-b944-9208138a4861 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 921.797690] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Instance 0d0f2d34-de35-4e80-8d9f-12693add0786 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61957) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 921.797690] env[61957]: WARNING nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Instance a8ca1ad6-d636-4fa2-b0ac-53b020e392ce is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 921.797690] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Instance 736eabcf-70f8-4e1e-9830-270d9971fc9a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61957) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 921.797997] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Instance 39b8b1df-efdd-4c22-9f31-85c85be4f0eb actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61957) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 921.797997] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Instance 19966b0a-53b7-48c5-849c-a9d00dc024f8 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61957) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 921.797997] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Instance d4674fbc-47b1-42d1-aaba-e86d46c51e8f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61957) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 921.798255] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Instance b598b8bb-919e-4404-b264-7b76161b0f79 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61957) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 921.814401] env[61957]: DEBUG nova.compute.utils [None req-d2e4cd3a-a43a-4a0c-a755-6e27eb102a3e tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Using /dev/sd instead of None {{(pid=61957) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 922.068807] env[61957]: DEBUG oslo_vmware.api [None req-5a91d738-be3a-4667-949d-376fc04acd42 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': task-1277716, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.76749} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.069169] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a91d738-be3a-4667-949d-376fc04acd42 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore2] d4674fbc-47b1-42d1-aaba-e86d46c51e8f/d4674fbc-47b1-42d1-aaba-e86d46c51e8f.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 922.069439] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-5a91d738-be3a-4667-949d-376fc04acd42 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: d4674fbc-47b1-42d1-aaba-e86d46c51e8f] Extending root virtual disk to 1048576 {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 922.069727] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b9906fd1-ebd8-4be3-8e82-05cf711d3b76 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.076165] env[61957]: DEBUG oslo_vmware.api [None req-5a91d738-be3a-4667-949d-376fc04acd42 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Waiting for the task: (returnval){ [ 922.076165] env[61957]: value = "task-1277718" [ 922.076165] env[61957]: _type = "Task" [ 922.076165] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.085315] env[61957]: DEBUG oslo_vmware.api [None req-5a91d738-be3a-4667-949d-376fc04acd42 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': task-1277718, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.272293] env[61957]: DEBUG oslo_concurrency.lockutils [None req-9213a8e4-435d-4ff1-a82b-fdbd92c25d20 tempest-ServersNegativeTestMultiTenantJSON-2135399494 tempest-ServersNegativeTestMultiTenantJSON-2135399494-project-member] Acquiring lock "19966b0a-53b7-48c5-849c-a9d00dc024f8" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 922.272293] env[61957]: DEBUG oslo_concurrency.lockutils [None req-9213a8e4-435d-4ff1-a82b-fdbd92c25d20 tempest-ServersNegativeTestMultiTenantJSON-2135399494 tempest-ServersNegativeTestMultiTenantJSON-2135399494-project-member] Lock "19966b0a-53b7-48c5-849c-a9d00dc024f8" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 922.272527] env[61957]: DEBUG oslo_concurrency.lockutils [None req-9213a8e4-435d-4ff1-a82b-fdbd92c25d20 tempest-ServersNegativeTestMultiTenantJSON-2135399494 tempest-ServersNegativeTestMultiTenantJSON-2135399494-project-member] Acquiring lock "19966b0a-53b7-48c5-849c-a9d00dc024f8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 922.272726] env[61957]: DEBUG oslo_concurrency.lockutils [None req-9213a8e4-435d-4ff1-a82b-fdbd92c25d20 tempest-ServersNegativeTestMultiTenantJSON-2135399494 tempest-ServersNegativeTestMultiTenantJSON-2135399494-project-member] Lock "19966b0a-53b7-48c5-849c-a9d00dc024f8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 922.272953] env[61957]: DEBUG oslo_concurrency.lockutils [None req-9213a8e4-435d-4ff1-a82b-fdbd92c25d20 tempest-ServersNegativeTestMultiTenantJSON-2135399494 tempest-ServersNegativeTestMultiTenantJSON-2135399494-project-member] Lock "19966b0a-53b7-48c5-849c-a9d00dc024f8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 922.276645] env[61957]: INFO nova.compute.manager [None req-9213a8e4-435d-4ff1-a82b-fdbd92c25d20 tempest-ServersNegativeTestMultiTenantJSON-2135399494 tempest-ServersNegativeTestMultiTenantJSON-2135399494-project-member] [instance: 19966b0a-53b7-48c5-849c-a9d00dc024f8] Terminating instance [ 922.282897] env[61957]: DEBUG nova.compute.manager [None req-9213a8e4-435d-4ff1-a82b-fdbd92c25d20 tempest-ServersNegativeTestMultiTenantJSON-2135399494 tempest-ServersNegativeTestMultiTenantJSON-2135399494-project-member] [instance: 19966b0a-53b7-48c5-849c-a9d00dc024f8] Start destroying the instance on the hypervisor. {{(pid=61957) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 922.283043] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-9213a8e4-435d-4ff1-a82b-fdbd92c25d20 tempest-ServersNegativeTestMultiTenantJSON-2135399494 tempest-ServersNegativeTestMultiTenantJSON-2135399494-project-member] [instance: 19966b0a-53b7-48c5-849c-a9d00dc024f8] Destroying instance {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 922.283835] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-151a5880-3e4c-437f-8551-33892822c615 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.288057] env[61957]: DEBUG oslo_concurrency.lockutils [None req-3e0cd30b-989c-4521-a40d-70a59bfcf72a tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 922.292909] env[61957]: DEBUG oslo_vmware.api [None req-68adc3d5-230e-4d31-9bdb-009ac1731c2d tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Task: {'id': task-1277717, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.296092] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-9213a8e4-435d-4ff1-a82b-fdbd92c25d20 tempest-ServersNegativeTestMultiTenantJSON-2135399494 tempest-ServersNegativeTestMultiTenantJSON-2135399494-project-member] [instance: 19966b0a-53b7-48c5-849c-a9d00dc024f8] Powering off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 922.296092] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c3eb98b4-1033-47ff-8900-cf43a8dfbeea {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.300798] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Instance 7e2db5f6-1419-415e-b957-66cf2e3ec19d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61957) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 922.304018] env[61957]: DEBUG oslo_vmware.api [None req-9213a8e4-435d-4ff1-a82b-fdbd92c25d20 tempest-ServersNegativeTestMultiTenantJSON-2135399494 tempest-ServersNegativeTestMultiTenantJSON-2135399494-project-member] Waiting for the task: (returnval){ [ 922.304018] env[61957]: value = "task-1277719" [ 922.304018] env[61957]: _type = "Task" [ 922.304018] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.313752] env[61957]: DEBUG oslo_vmware.api [None req-9213a8e4-435d-4ff1-a82b-fdbd92c25d20 tempest-ServersNegativeTestMultiTenantJSON-2135399494 tempest-ServersNegativeTestMultiTenantJSON-2135399494-project-member] Task: {'id': task-1277719, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.317279] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d2e4cd3a-a43a-4a0c-a755-6e27eb102a3e tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Lock "39b8b1df-efdd-4c22-9f31-85c85be4f0eb" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.007s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 922.586440] env[61957]: DEBUG oslo_vmware.api [None req-5a91d738-be3a-4667-949d-376fc04acd42 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': task-1277718, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063411} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.586723] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-5a91d738-be3a-4667-949d-376fc04acd42 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: d4674fbc-47b1-42d1-aaba-e86d46c51e8f] Extended root virtual disk {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 922.587502] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6cfae0b-e704-40df-8c78-cfa6d2f58e8f {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.610961] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-5a91d738-be3a-4667-949d-376fc04acd42 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: d4674fbc-47b1-42d1-aaba-e86d46c51e8f] Reconfiguring VM instance instance-0000004e to attach disk [datastore2] d4674fbc-47b1-42d1-aaba-e86d46c51e8f/d4674fbc-47b1-42d1-aaba-e86d46c51e8f.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 922.611252] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3d16ae08-d5b4-4681-9acd-c3848a4f5c64 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.631606] env[61957]: DEBUG oslo_vmware.api [None req-5a91d738-be3a-4667-949d-376fc04acd42 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Waiting for the task: (returnval){ [ 922.631606] env[61957]: value = "task-1277720" [ 922.631606] env[61957]: _type = "Task" [ 922.631606] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.640914] env[61957]: DEBUG oslo_vmware.api [None req-5a91d738-be3a-4667-949d-376fc04acd42 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': task-1277720, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.765985] env[61957]: DEBUG nova.compute.manager [None req-d61e8152-d9da-413c-b7fd-8eed0b01a69d tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: b598b8bb-919e-4404-b264-7b76161b0f79] Start spawning the instance on the hypervisor. {{(pid=61957) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 922.791046] env[61957]: DEBUG oslo_vmware.api [None req-68adc3d5-230e-4d31-9bdb-009ac1731c2d tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Task: {'id': task-1277717, 'name': ReconfigVM_Task, 'duration_secs': 0.586812} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.791046] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-68adc3d5-230e-4d31-9bdb-009ac1731c2d tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] [instance: 526ff179-62a6-4763-ab25-797617c4ed57] Reconfigured VM instance instance-00000041 to attach disk [datastore1] volume-6c0df615-4d1a-45c6-9f48-ab1813e515b1/volume-6c0df615-4d1a-45c6-9f48-ab1813e515b1.vmdk or device None with type thin {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 922.794631] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3090f007-cd3f-4e89-aab8-5dec5601d7ee {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.806305] env[61957]: DEBUG nova.virt.hardware [None req-d61e8152-d9da-413c-b7fd-8eed0b01a69d tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T17:22:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T17:21:50Z,direct_url=,disk_format='vmdk',id=11c76a2c-f705-470a-ba9d-4657858bab38,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='81d3096d0f094373913fc3dc8f5c3c6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T17:21:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 922.806576] env[61957]: DEBUG nova.virt.hardware [None req-d61e8152-d9da-413c-b7fd-8eed0b01a69d tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Flavor limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 922.806737] env[61957]: DEBUG nova.virt.hardware [None req-d61e8152-d9da-413c-b7fd-8eed0b01a69d tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Image limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 922.806924] env[61957]: DEBUG nova.virt.hardware [None req-d61e8152-d9da-413c-b7fd-8eed0b01a69d tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Flavor pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 922.807089] env[61957]: DEBUG nova.virt.hardware [None req-d61e8152-d9da-413c-b7fd-8eed0b01a69d tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Image pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 922.807242] env[61957]: DEBUG nova.virt.hardware [None req-d61e8152-d9da-413c-b7fd-8eed0b01a69d tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 922.807440] env[61957]: DEBUG nova.virt.hardware [None req-d61e8152-d9da-413c-b7fd-8eed0b01a69d tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 922.807594] env[61957]: DEBUG nova.virt.hardware [None req-d61e8152-d9da-413c-b7fd-8eed0b01a69d tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 922.807771] env[61957]: DEBUG nova.virt.hardware [None req-d61e8152-d9da-413c-b7fd-8eed0b01a69d tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Got 1 possible topologies {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 922.807914] env[61957]: DEBUG nova.virt.hardware [None req-d61e8152-d9da-413c-b7fd-8eed0b01a69d tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 922.808102] env[61957]: DEBUG nova.virt.hardware [None req-d61e8152-d9da-413c-b7fd-8eed0b01a69d tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 922.808850] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Instance 9b0053fe-420e-4b92-86a8-ed44f6a6ec49 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61957) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 922.811032] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e96422f2-7c88-4959-9f74-c78f204a4885 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.819643] env[61957]: DEBUG oslo_vmware.api [None req-68adc3d5-230e-4d31-9bdb-009ac1731c2d tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Waiting for the task: (returnval){ [ 922.819643] env[61957]: value = "task-1277721" [ 922.819643] env[61957]: _type = "Task" [ 922.819643] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.824531] env[61957]: DEBUG oslo_vmware.api [None req-9213a8e4-435d-4ff1-a82b-fdbd92c25d20 tempest-ServersNegativeTestMultiTenantJSON-2135399494 tempest-ServersNegativeTestMultiTenantJSON-2135399494-project-member] Task: {'id': task-1277719, 'name': PowerOffVM_Task, 'duration_secs': 0.191401} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.828974] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4eddc597-b169-478f-934e-b6e2b8aec43a {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.833123] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-9213a8e4-435d-4ff1-a82b-fdbd92c25d20 tempest-ServersNegativeTestMultiTenantJSON-2135399494 tempest-ServersNegativeTestMultiTenantJSON-2135399494-project-member] [instance: 19966b0a-53b7-48c5-849c-a9d00dc024f8] Powered off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 922.833703] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-9213a8e4-435d-4ff1-a82b-fdbd92c25d20 tempest-ServersNegativeTestMultiTenantJSON-2135399494 tempest-ServersNegativeTestMultiTenantJSON-2135399494-project-member] [instance: 19966b0a-53b7-48c5-849c-a9d00dc024f8] Unregistering the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 922.833853] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-81a4693a-0035-4a66-b73d-03c251ba626b {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.840964] env[61957]: DEBUG oslo_vmware.api [None req-68adc3d5-230e-4d31-9bdb-009ac1731c2d tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Task: {'id': task-1277721, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.905025] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-9213a8e4-435d-4ff1-a82b-fdbd92c25d20 tempest-ServersNegativeTestMultiTenantJSON-2135399494 tempest-ServersNegativeTestMultiTenantJSON-2135399494-project-member] [instance: 19966b0a-53b7-48c5-849c-a9d00dc024f8] Unregistered the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 922.905025] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-9213a8e4-435d-4ff1-a82b-fdbd92c25d20 tempest-ServersNegativeTestMultiTenantJSON-2135399494 tempest-ServersNegativeTestMultiTenantJSON-2135399494-project-member] [instance: 19966b0a-53b7-48c5-849c-a9d00dc024f8] Deleting contents of the VM from datastore datastore2 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 922.905025] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-9213a8e4-435d-4ff1-a82b-fdbd92c25d20 tempest-ServersNegativeTestMultiTenantJSON-2135399494 tempest-ServersNegativeTestMultiTenantJSON-2135399494-project-member] Deleting the datastore file [datastore2] 19966b0a-53b7-48c5-849c-a9d00dc024f8 {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 922.905025] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-db2a5f93-e70e-4bfa-8896-5a59f3a2d829 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.912777] env[61957]: DEBUG oslo_vmware.api [None req-9213a8e4-435d-4ff1-a82b-fdbd92c25d20 tempest-ServersNegativeTestMultiTenantJSON-2135399494 tempest-ServersNegativeTestMultiTenantJSON-2135399494-project-member] Waiting for the task: (returnval){ [ 922.912777] env[61957]: value = "task-1277723" [ 922.912777] env[61957]: _type = "Task" [ 922.912777] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.922769] env[61957]: DEBUG oslo_vmware.api [None req-9213a8e4-435d-4ff1-a82b-fdbd92c25d20 tempest-ServersNegativeTestMultiTenantJSON-2135399494 tempest-ServersNegativeTestMultiTenantJSON-2135399494-project-member] Task: {'id': task-1277723, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.142610] env[61957]: DEBUG oslo_vmware.api [None req-5a91d738-be3a-4667-949d-376fc04acd42 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': task-1277720, 'name': ReconfigVM_Task, 'duration_secs': 0.346546} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.142955] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-5a91d738-be3a-4667-949d-376fc04acd42 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: d4674fbc-47b1-42d1-aaba-e86d46c51e8f] Reconfigured VM instance instance-0000004e to attach disk [datastore2] d4674fbc-47b1-42d1-aaba-e86d46c51e8f/d4674fbc-47b1-42d1-aaba-e86d46c51e8f.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 923.143710] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-424606d2-530c-45c4-9882-adaa10f4dbbe {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.151513] env[61957]: DEBUG oslo_vmware.api [None req-5a91d738-be3a-4667-949d-376fc04acd42 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Waiting for the task: (returnval){ [ 923.151513] env[61957]: value = "task-1277724" [ 923.151513] env[61957]: _type = "Task" [ 923.151513] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.164077] env[61957]: DEBUG oslo_vmware.api [None req-5a91d738-be3a-4667-949d-376fc04acd42 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': task-1277724, 'name': Rename_Task} progress is 6%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.213786] env[61957]: DEBUG nova.compute.manager [req-f299a1db-51d3-4ef8-aaa5-9044fad27936 req-70101fae-667f-4f8f-9e98-9f27da2b9929 service nova] [instance: b598b8bb-919e-4404-b264-7b76161b0f79] Received event network-vif-plugged-be3f3b45-e91d-4c38-b746-a10c838decf7 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 923.214066] env[61957]: DEBUG oslo_concurrency.lockutils [req-f299a1db-51d3-4ef8-aaa5-9044fad27936 req-70101fae-667f-4f8f-9e98-9f27da2b9929 service nova] Acquiring lock "b598b8bb-919e-4404-b264-7b76161b0f79-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 923.214335] env[61957]: DEBUG oslo_concurrency.lockutils [req-f299a1db-51d3-4ef8-aaa5-9044fad27936 req-70101fae-667f-4f8f-9e98-9f27da2b9929 service nova] Lock "b598b8bb-919e-4404-b264-7b76161b0f79-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 923.214543] env[61957]: DEBUG oslo_concurrency.lockutils [req-f299a1db-51d3-4ef8-aaa5-9044fad27936 req-70101fae-667f-4f8f-9e98-9f27da2b9929 service nova] Lock "b598b8bb-919e-4404-b264-7b76161b0f79-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 923.214746] env[61957]: DEBUG nova.compute.manager [req-f299a1db-51d3-4ef8-aaa5-9044fad27936 req-70101fae-667f-4f8f-9e98-9f27da2b9929 service nova] [instance: b598b8bb-919e-4404-b264-7b76161b0f79] No waiting events found dispatching network-vif-plugged-be3f3b45-e91d-4c38-b746-a10c838decf7 {{(pid=61957) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 923.214945] env[61957]: WARNING nova.compute.manager [req-f299a1db-51d3-4ef8-aaa5-9044fad27936 req-70101fae-667f-4f8f-9e98-9f27da2b9929 service nova] [instance: b598b8bb-919e-4404-b264-7b76161b0f79] Received unexpected event network-vif-plugged-be3f3b45-e91d-4c38-b746-a10c838decf7 for instance with vm_state building and task_state spawning. [ 923.315570] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Instance 50e09db7-9f8f-452b-8232-bd7473e9e63a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61957) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 923.323446] env[61957]: DEBUG nova.network.neutron [None req-d61e8152-d9da-413c-b7fd-8eed0b01a69d tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: b598b8bb-919e-4404-b264-7b76161b0f79] Successfully updated port: be3f3b45-e91d-4c38-b746-a10c838decf7 {{(pid=61957) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 923.336982] env[61957]: DEBUG oslo_vmware.api [None req-68adc3d5-230e-4d31-9bdb-009ac1731c2d tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Task: {'id': task-1277721, 'name': ReconfigVM_Task, 'duration_secs': 0.199594} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.337307] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-68adc3d5-230e-4d31-9bdb-009ac1731c2d tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] [instance: 526ff179-62a6-4763-ab25-797617c4ed57] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-274553', 'volume_id': '6c0df615-4d1a-45c6-9f48-ab1813e515b1', 'name': 'volume-6c0df615-4d1a-45c6-9f48-ab1813e515b1', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '526ff179-62a6-4763-ab25-797617c4ed57', 'attached_at': '', 'detached_at': '', 'volume_id': '6c0df615-4d1a-45c6-9f48-ab1813e515b1', 'serial': '6c0df615-4d1a-45c6-9f48-ab1813e515b1'} {{(pid=61957) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 923.387514] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d2e4cd3a-a43a-4a0c-a755-6e27eb102a3e tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Acquiring lock "39b8b1df-efdd-4c22-9f31-85c85be4f0eb" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 923.387728] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d2e4cd3a-a43a-4a0c-a755-6e27eb102a3e tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Lock "39b8b1df-efdd-4c22-9f31-85c85be4f0eb" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 923.387984] env[61957]: INFO nova.compute.manager [None req-d2e4cd3a-a43a-4a0c-a755-6e27eb102a3e tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 39b8b1df-efdd-4c22-9f31-85c85be4f0eb] Attaching volume f5184167-7e0d-480b-b064-0361333ea716 to /dev/sdb [ 923.423138] env[61957]: DEBUG oslo_vmware.api [None req-9213a8e4-435d-4ff1-a82b-fdbd92c25d20 tempest-ServersNegativeTestMultiTenantJSON-2135399494 tempest-ServersNegativeTestMultiTenantJSON-2135399494-project-member] Task: {'id': task-1277723, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.416455} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.423904] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-9213a8e4-435d-4ff1-a82b-fdbd92c25d20 tempest-ServersNegativeTestMultiTenantJSON-2135399494 tempest-ServersNegativeTestMultiTenantJSON-2135399494-project-member] Deleted the datastore file {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 923.426088] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-9213a8e4-435d-4ff1-a82b-fdbd92c25d20 tempest-ServersNegativeTestMultiTenantJSON-2135399494 tempest-ServersNegativeTestMultiTenantJSON-2135399494-project-member] [instance: 19966b0a-53b7-48c5-849c-a9d00dc024f8] Deleted contents of the VM from datastore datastore2 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 923.426088] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-9213a8e4-435d-4ff1-a82b-fdbd92c25d20 tempest-ServersNegativeTestMultiTenantJSON-2135399494 tempest-ServersNegativeTestMultiTenantJSON-2135399494-project-member] [instance: 19966b0a-53b7-48c5-849c-a9d00dc024f8] Instance destroyed {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 923.426088] env[61957]: INFO nova.compute.manager [None req-9213a8e4-435d-4ff1-a82b-fdbd92c25d20 tempest-ServersNegativeTestMultiTenantJSON-2135399494 tempest-ServersNegativeTestMultiTenantJSON-2135399494-project-member] [instance: 19966b0a-53b7-48c5-849c-a9d00dc024f8] Took 1.14 seconds to destroy the instance on the hypervisor. [ 923.426088] env[61957]: DEBUG oslo.service.loopingcall [None req-9213a8e4-435d-4ff1-a82b-fdbd92c25d20 tempest-ServersNegativeTestMultiTenantJSON-2135399494 tempest-ServersNegativeTestMultiTenantJSON-2135399494-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 923.426088] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8272d5fa-cd0a-4a7f-ae9f-017bc89e5c92 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.428231] env[61957]: DEBUG nova.compute.manager [-] [instance: 19966b0a-53b7-48c5-849c-a9d00dc024f8] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 923.428467] env[61957]: DEBUG nova.network.neutron [-] [instance: 19966b0a-53b7-48c5-849c-a9d00dc024f8] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 923.434261] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19b8afa3-c061-47f5-961f-88747d338c7a {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.447900] env[61957]: DEBUG nova.virt.block_device [None req-d2e4cd3a-a43a-4a0c-a755-6e27eb102a3e tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 39b8b1df-efdd-4c22-9f31-85c85be4f0eb] Updating existing volume attachment record: 6ea3e493-9fb6-42bb-807e-30d63e64f785 {{(pid=61957) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 923.664777] env[61957]: DEBUG oslo_vmware.api [None req-5a91d738-be3a-4667-949d-376fc04acd42 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': task-1277724, 'name': Rename_Task, 'duration_secs': 0.166596} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.665208] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a91d738-be3a-4667-949d-376fc04acd42 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: d4674fbc-47b1-42d1-aaba-e86d46c51e8f] Powering on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 923.665987] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c2edcd28-110e-4a9c-9f49-68ea9d3c5fcd {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.676263] env[61957]: DEBUG oslo_vmware.api [None req-5a91d738-be3a-4667-949d-376fc04acd42 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Waiting for the task: (returnval){ [ 923.676263] env[61957]: value = "task-1277726" [ 923.676263] env[61957]: _type = "Task" [ 923.676263] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.685362] env[61957]: DEBUG oslo_vmware.api [None req-5a91d738-be3a-4667-949d-376fc04acd42 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': task-1277726, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.822941] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Instance 2e32aae5-edd7-402f-98ad-75a93d26f7a1 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61957) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 923.823284] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Migration f4269166-72d4-48ca-b806-e1e29520eb19 is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61957) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 923.823690] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Instance dbeeb200-70b9-4cb4-b5a4-182389d21918 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61957) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 923.824042] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Total usable vcpus: 48, total allocated vcpus: 12 {{(pid=61957) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 923.824227] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2816MB phys_disk=200GB used_disk=12GB total_vcpus=48 used_vcpus=12 pci_stats=[] {{(pid=61957) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 923.828129] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d61e8152-d9da-413c-b7fd-8eed0b01a69d tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Acquiring lock "refresh_cache-b598b8bb-919e-4404-b264-7b76161b0f79" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 923.828212] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d61e8152-d9da-413c-b7fd-8eed0b01a69d tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Acquired lock "refresh_cache-b598b8bb-919e-4404-b264-7b76161b0f79" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 923.828382] env[61957]: DEBUG nova.network.neutron [None req-d61e8152-d9da-413c-b7fd-8eed0b01a69d tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: b598b8bb-919e-4404-b264-7b76161b0f79] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 924.126152] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cf11ecb-d12e-42da-8836-db02e2e1c464 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.134514] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9e1aba9-bf74-46e3-9775-fefb3bb73165 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.169727] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19e950f2-500d-40e6-8aeb-559311d3542f {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.180889] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c63736c0-9d69-4483-bd07-d54dac0873f5 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.189214] env[61957]: DEBUG nova.network.neutron [-] [instance: 19966b0a-53b7-48c5-849c-a9d00dc024f8] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 924.190389] env[61957]: DEBUG oslo_vmware.api [None req-5a91d738-be3a-4667-949d-376fc04acd42 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': task-1277726, 'name': PowerOnVM_Task, 'duration_secs': 0.462578} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.198327] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a91d738-be3a-4667-949d-376fc04acd42 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: d4674fbc-47b1-42d1-aaba-e86d46c51e8f] Powered on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 924.198560] env[61957]: INFO nova.compute.manager [None req-5a91d738-be3a-4667-949d-376fc04acd42 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: d4674fbc-47b1-42d1-aaba-e86d46c51e8f] Took 7.97 seconds to spawn the instance on the hypervisor. [ 924.198758] env[61957]: DEBUG nova.compute.manager [None req-5a91d738-be3a-4667-949d-376fc04acd42 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: d4674fbc-47b1-42d1-aaba-e86d46c51e8f] Checking state {{(pid=61957) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 924.199256] env[61957]: DEBUG nova.compute.provider_tree [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 924.200969] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91811ac1-85bd-40d0-a118-79f5dec48c1c {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.367891] env[61957]: DEBUG nova.network.neutron [None req-d61e8152-d9da-413c-b7fd-8eed0b01a69d tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: b598b8bb-919e-4404-b264-7b76161b0f79] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 924.379756] env[61957]: DEBUG nova.objects.instance [None req-68adc3d5-230e-4d31-9bdb-009ac1731c2d tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Lazy-loading 'flavor' on Instance uuid 526ff179-62a6-4763-ab25-797617c4ed57 {{(pid=61957) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 924.522225] env[61957]: DEBUG nova.network.neutron [None req-d61e8152-d9da-413c-b7fd-8eed0b01a69d tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: b598b8bb-919e-4404-b264-7b76161b0f79] Updating instance_info_cache with network_info: [{"id": "be3f3b45-e91d-4c38-b746-a10c838decf7", "address": "fa:16:3e:6d:a5:2a", "network": {"id": "4f0e57be-8906-48b5-865f-59ff12250a8b", "bridge": "br-int", "label": "tempest-ServersTestJSON-1916182687-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "65fd71bcd1bf41238a9cc3a5d6dd4924", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "939c05b6-8f31-4f3a-95ac-6297e0bd243e", "external-id": "nsx-vlan-transportzone-825", "segmentation_id": 825, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbe3f3b45-e9", "ovs_interfaceid": "be3f3b45-e91d-4c38-b746-a10c838decf7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 924.699331] env[61957]: INFO nova.compute.manager [-] [instance: 19966b0a-53b7-48c5-849c-a9d00dc024f8] Took 1.27 seconds to deallocate network for instance. [ 924.709128] env[61957]: DEBUG nova.scheduler.client.report [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 924.723138] env[61957]: INFO nova.compute.manager [None req-5a91d738-be3a-4667-949d-376fc04acd42 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: d4674fbc-47b1-42d1-aaba-e86d46c51e8f] Took 39.82 seconds to build instance. [ 924.885347] env[61957]: DEBUG oslo_concurrency.lockutils [None req-68adc3d5-230e-4d31-9bdb-009ac1731c2d tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Lock "526ff179-62a6-4763-ab25-797617c4ed57" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 10.336s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 925.024696] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d61e8152-d9da-413c-b7fd-8eed0b01a69d tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Releasing lock "refresh_cache-b598b8bb-919e-4404-b264-7b76161b0f79" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 925.024997] env[61957]: DEBUG nova.compute.manager [None req-d61e8152-d9da-413c-b7fd-8eed0b01a69d tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: b598b8bb-919e-4404-b264-7b76161b0f79] Instance network_info: |[{"id": "be3f3b45-e91d-4c38-b746-a10c838decf7", "address": "fa:16:3e:6d:a5:2a", "network": {"id": "4f0e57be-8906-48b5-865f-59ff12250a8b", "bridge": "br-int", "label": "tempest-ServersTestJSON-1916182687-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "65fd71bcd1bf41238a9cc3a5d6dd4924", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "939c05b6-8f31-4f3a-95ac-6297e0bd243e", "external-id": "nsx-vlan-transportzone-825", "segmentation_id": 825, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbe3f3b45-e9", "ovs_interfaceid": "be3f3b45-e91d-4c38-b746-a10c838decf7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 925.025473] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-d61e8152-d9da-413c-b7fd-8eed0b01a69d tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: b598b8bb-919e-4404-b264-7b76161b0f79] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6d:a5:2a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '939c05b6-8f31-4f3a-95ac-6297e0bd243e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'be3f3b45-e91d-4c38-b746-a10c838decf7', 'vif_model': 'vmxnet3'}] {{(pid=61957) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 925.034531] env[61957]: DEBUG oslo.service.loopingcall [None req-d61e8152-d9da-413c-b7fd-8eed0b01a69d tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 925.037754] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b598b8bb-919e-4404-b264-7b76161b0f79] Creating VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 925.037754] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8eb301a2-ca14-40cf-8ccd-754303dfc0c9 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.056463] env[61957]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 925.056463] env[61957]: value = "task-1277729" [ 925.056463] env[61957]: _type = "Task" [ 925.056463] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.065826] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277729, 'name': CreateVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.213934] env[61957]: DEBUG oslo_concurrency.lockutils [None req-9213a8e4-435d-4ff1-a82b-fdbd92c25d20 tempest-ServersNegativeTestMultiTenantJSON-2135399494 tempest-ServersNegativeTestMultiTenantJSON-2135399494-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 925.219962] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61957) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 925.220482] env[61957]: DEBUG oslo_concurrency.lockutils [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 4.490s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 925.220482] env[61957]: DEBUG oslo_concurrency.lockutils [None req-9314189b-c158-4aed-834e-ffe55cc4ba51 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 30.574s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 925.220632] env[61957]: DEBUG oslo_concurrency.lockutils [None req-9314189b-c158-4aed-834e-ffe55cc4ba51 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 925.222781] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d28d4db3-4ba8-4b6b-b020-dcb0a978b0ab tempest-InstanceActionsNegativeTestJSON-259524954 tempest-InstanceActionsNegativeTestJSON-259524954-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 29.290s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 925.225992] env[61957]: INFO nova.compute.claims [None req-d28d4db3-4ba8-4b6b-b020-dcb0a978b0ab tempest-InstanceActionsNegativeTestJSON-259524954 tempest-InstanceActionsNegativeTestJSON-259524954-project-member] [instance: 7e2db5f6-1419-415e-b957-66cf2e3ec19d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 925.229516] env[61957]: DEBUG oslo_concurrency.lockutils [None req-5a91d738-be3a-4667-949d-376fc04acd42 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Lock "d4674fbc-47b1-42d1-aaba-e86d46c51e8f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 45.432s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 925.234680] env[61957]: DEBUG oslo_concurrency.lockutils [None req-afc58718-a61e-46a6-84fa-2df40ae107ee tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Acquiring lock "526ff179-62a6-4763-ab25-797617c4ed57" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 925.234766] env[61957]: DEBUG oslo_concurrency.lockutils [None req-afc58718-a61e-46a6-84fa-2df40ae107ee tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Lock "526ff179-62a6-4763-ab25-797617c4ed57" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 925.234917] env[61957]: DEBUG nova.compute.manager [None req-afc58718-a61e-46a6-84fa-2df40ae107ee tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] [instance: 526ff179-62a6-4763-ab25-797617c4ed57] Checking state {{(pid=61957) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 925.235824] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce88e971-ffee-4082-95a0-6793ac36f012 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.244034] env[61957]: DEBUG nova.compute.manager [None req-afc58718-a61e-46a6-84fa-2df40ae107ee tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] [instance: 526ff179-62a6-4763-ab25-797617c4ed57] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61957) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 925.244571] env[61957]: DEBUG nova.objects.instance [None req-afc58718-a61e-46a6-84fa-2df40ae107ee tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Lazy-loading 'flavor' on Instance uuid 526ff179-62a6-4763-ab25-797617c4ed57 {{(pid=61957) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 925.252297] env[61957]: INFO nova.scheduler.client.report [None req-9314189b-c158-4aed-834e-ffe55cc4ba51 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Deleted allocations for instance 4951132e-7247-4772-8f88-3664c6a7e61e [ 925.569490] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277729, 'name': CreateVM_Task, 'duration_secs': 0.425296} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.570372] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b598b8bb-919e-4404-b264-7b76161b0f79] Created VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 925.571122] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d61e8152-d9da-413c-b7fd-8eed0b01a69d tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 925.571323] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d61e8152-d9da-413c-b7fd-8eed0b01a69d tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Acquired lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 925.571649] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d61e8152-d9da-413c-b7fd-8eed0b01a69d tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 925.572151] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-81e27f20-a23f-4440-9d0e-1912681c1069 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.577120] env[61957]: DEBUG oslo_vmware.api [None req-d61e8152-d9da-413c-b7fd-8eed0b01a69d tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Waiting for the task: (returnval){ [ 925.577120] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]5260b694-2096-63b0-f3e1-b0d38d46a473" [ 925.577120] env[61957]: _type = "Task" [ 925.577120] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.584214] env[61957]: DEBUG nova.compute.manager [req-11b605d4-56ad-42d3-a0bc-80b780936788 req-f16114d4-4bcd-4726-b1a7-843ee136ee90 service nova] [instance: b598b8bb-919e-4404-b264-7b76161b0f79] Received event network-changed-be3f3b45-e91d-4c38-b746-a10c838decf7 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 925.584829] env[61957]: DEBUG nova.compute.manager [req-11b605d4-56ad-42d3-a0bc-80b780936788 req-f16114d4-4bcd-4726-b1a7-843ee136ee90 service nova] [instance: b598b8bb-919e-4404-b264-7b76161b0f79] Refreshing instance network info cache due to event network-changed-be3f3b45-e91d-4c38-b746-a10c838decf7. {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 925.584829] env[61957]: DEBUG oslo_concurrency.lockutils [req-11b605d4-56ad-42d3-a0bc-80b780936788 req-f16114d4-4bcd-4726-b1a7-843ee136ee90 service nova] Acquiring lock "refresh_cache-b598b8bb-919e-4404-b264-7b76161b0f79" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 925.584829] env[61957]: DEBUG oslo_concurrency.lockutils [req-11b605d4-56ad-42d3-a0bc-80b780936788 req-f16114d4-4bcd-4726-b1a7-843ee136ee90 service nova] Acquired lock "refresh_cache-b598b8bb-919e-4404-b264-7b76161b0f79" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 925.584998] env[61957]: DEBUG nova.network.neutron [req-11b605d4-56ad-42d3-a0bc-80b780936788 req-f16114d4-4bcd-4726-b1a7-843ee136ee90 service nova] [instance: b598b8bb-919e-4404-b264-7b76161b0f79] Refreshing network info cache for port be3f3b45-e91d-4c38-b746-a10c838decf7 {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 925.593158] env[61957]: DEBUG oslo_vmware.api [None req-d61e8152-d9da-413c-b7fd-8eed0b01a69d tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]5260b694-2096-63b0-f3e1-b0d38d46a473, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.750508] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-afc58718-a61e-46a6-84fa-2df40ae107ee tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] [instance: 526ff179-62a6-4763-ab25-797617c4ed57] Powering off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 925.753344] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1e97e106-eeb7-46c6-8fea-69b1bf17f5b7 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.760347] env[61957]: DEBUG oslo_concurrency.lockutils [None req-9314189b-c158-4aed-834e-ffe55cc4ba51 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Lock "4951132e-7247-4772-8f88-3664c6a7e61e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 34.307s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 925.763233] env[61957]: DEBUG oslo_vmware.api [None req-afc58718-a61e-46a6-84fa-2df40ae107ee tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Waiting for the task: (returnval){ [ 925.763233] env[61957]: value = "task-1277730" [ 925.763233] env[61957]: _type = "Task" [ 925.763233] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.773959] env[61957]: DEBUG oslo_vmware.api [None req-afc58718-a61e-46a6-84fa-2df40ae107ee tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Task: {'id': task-1277730, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.089875] env[61957]: DEBUG oslo_vmware.api [None req-d61e8152-d9da-413c-b7fd-8eed0b01a69d tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]5260b694-2096-63b0-f3e1-b0d38d46a473, 'name': SearchDatastore_Task, 'duration_secs': 0.015675} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.090214] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d61e8152-d9da-413c-b7fd-8eed0b01a69d tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Releasing lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 926.090440] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-d61e8152-d9da-413c-b7fd-8eed0b01a69d tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: b598b8bb-919e-4404-b264-7b76161b0f79] Processing image 11c76a2c-f705-470a-ba9d-4657858bab38 {{(pid=61957) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 926.091096] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d61e8152-d9da-413c-b7fd-8eed0b01a69d tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 926.091096] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d61e8152-d9da-413c-b7fd-8eed0b01a69d tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Acquired lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 926.091096] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-d61e8152-d9da-413c-b7fd-8eed0b01a69d tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 926.092252] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c9d62601-c1d7-4189-a9fe-1e29de1f3805 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.099996] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-d61e8152-d9da-413c-b7fd-8eed0b01a69d tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 926.100219] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-d61e8152-d9da-413c-b7fd-8eed0b01a69d tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61957) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 926.100939] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-03a2a7c9-2afd-43b5-a08b-9584705642b4 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.106847] env[61957]: DEBUG oslo_vmware.api [None req-d61e8152-d9da-413c-b7fd-8eed0b01a69d tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Waiting for the task: (returnval){ [ 926.106847] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52ba917b-6b61-8ccb-ce00-a2000f0c8db8" [ 926.106847] env[61957]: _type = "Task" [ 926.106847] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.118815] env[61957]: DEBUG oslo_vmware.api [None req-d61e8152-d9da-413c-b7fd-8eed0b01a69d tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52ba917b-6b61-8ccb-ce00-a2000f0c8db8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.248234] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d1519ec7-d973-47ee-af77-990ab0a419d4 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Acquiring lock "7793aa07-6aa2-459d-8a91-56c6b0412d3b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 926.248565] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d1519ec7-d973-47ee-af77-990ab0a419d4 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Lock "7793aa07-6aa2-459d-8a91-56c6b0412d3b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 926.291567] env[61957]: DEBUG oslo_vmware.api [None req-afc58718-a61e-46a6-84fa-2df40ae107ee tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Task: {'id': task-1277730, 'name': PowerOffVM_Task, 'duration_secs': 0.222699} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.292382] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-afc58718-a61e-46a6-84fa-2df40ae107ee tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] [instance: 526ff179-62a6-4763-ab25-797617c4ed57] Powered off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 926.294019] env[61957]: DEBUG nova.compute.manager [None req-afc58718-a61e-46a6-84fa-2df40ae107ee tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] [instance: 526ff179-62a6-4763-ab25-797617c4ed57] Checking state {{(pid=61957) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 926.294019] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df15e4c6-53c7-4665-ac43-cc3aab2f7e9b {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.374249] env[61957]: DEBUG nova.network.neutron [req-11b605d4-56ad-42d3-a0bc-80b780936788 req-f16114d4-4bcd-4726-b1a7-843ee136ee90 service nova] [instance: b598b8bb-919e-4404-b264-7b76161b0f79] Updated VIF entry in instance network info cache for port be3f3b45-e91d-4c38-b746-a10c838decf7. {{(pid=61957) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 926.374249] env[61957]: DEBUG nova.network.neutron [req-11b605d4-56ad-42d3-a0bc-80b780936788 req-f16114d4-4bcd-4726-b1a7-843ee136ee90 service nova] [instance: b598b8bb-919e-4404-b264-7b76161b0f79] Updating instance_info_cache with network_info: [{"id": "be3f3b45-e91d-4c38-b746-a10c838decf7", "address": "fa:16:3e:6d:a5:2a", "network": {"id": "4f0e57be-8906-48b5-865f-59ff12250a8b", "bridge": "br-int", "label": "tempest-ServersTestJSON-1916182687-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "65fd71bcd1bf41238a9cc3a5d6dd4924", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "939c05b6-8f31-4f3a-95ac-6297e0bd243e", "external-id": "nsx-vlan-transportzone-825", "segmentation_id": 825, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbe3f3b45-e9", "ovs_interfaceid": "be3f3b45-e91d-4c38-b746-a10c838decf7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 926.534187] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f82baec0-31f2-4ce0-8277-7e0e6aaf1311 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.542266] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9ff7328-84ca-40ab-a82d-fcb210fa9c2a {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.574259] env[61957]: DEBUG nova.compute.manager [None req-4cbf1ed6-4756-412e-9889-a64859683b60 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: d4674fbc-47b1-42d1-aaba-e86d46c51e8f] Checking state {{(pid=61957) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 926.575234] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98f1c0a9-a2d3-47aa-8f57-e317d5c3f7d9 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.578602] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bdfe402a-a98c-4d27-941b-8530e4b949e4 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.587832] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2325d55c-b303-41a2-94d8-7a8952c3c8c3 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.604510] env[61957]: DEBUG nova.compute.provider_tree [None req-d28d4db3-4ba8-4b6b-b020-dcb0a978b0ab tempest-InstanceActionsNegativeTestJSON-259524954 tempest-InstanceActionsNegativeTestJSON-259524954-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 926.617224] env[61957]: DEBUG oslo_vmware.api [None req-d61e8152-d9da-413c-b7fd-8eed0b01a69d tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52ba917b-6b61-8ccb-ce00-a2000f0c8db8, 'name': SearchDatastore_Task, 'duration_secs': 0.007735} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.618341] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e4506dab-0a98-46bf-81c2-d7cb3af5af52 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.625753] env[61957]: DEBUG oslo_vmware.api [None req-d61e8152-d9da-413c-b7fd-8eed0b01a69d tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Waiting for the task: (returnval){ [ 926.625753] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52fe53a0-8f62-a509-fb57-f144dc869ea2" [ 926.625753] env[61957]: _type = "Task" [ 926.625753] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.633893] env[61957]: DEBUG oslo_vmware.api [None req-d61e8152-d9da-413c-b7fd-8eed0b01a69d tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52fe53a0-8f62-a509-fb57-f144dc869ea2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.756660] env[61957]: DEBUG nova.compute.manager [None req-d1519ec7-d973-47ee-af77-990ab0a419d4 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] [instance: 7793aa07-6aa2-459d-8a91-56c6b0412d3b] Starting instance... {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 926.812448] env[61957]: DEBUG oslo_concurrency.lockutils [None req-afc58718-a61e-46a6-84fa-2df40ae107ee tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Lock "526ff179-62a6-4763-ab25-797617c4ed57" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.576s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 926.877094] env[61957]: DEBUG oslo_concurrency.lockutils [req-11b605d4-56ad-42d3-a0bc-80b780936788 req-f16114d4-4bcd-4726-b1a7-843ee136ee90 service nova] Releasing lock "refresh_cache-b598b8bb-919e-4404-b264-7b76161b0f79" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 926.877440] env[61957]: DEBUG nova.compute.manager [req-11b605d4-56ad-42d3-a0bc-80b780936788 req-f16114d4-4bcd-4726-b1a7-843ee136ee90 service nova] [instance: 19966b0a-53b7-48c5-849c-a9d00dc024f8] Received event network-vif-deleted-159b1e72-9328-460d-b334-9a3ad5d73b44 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 927.096927] env[61957]: INFO nova.compute.manager [None req-4cbf1ed6-4756-412e-9889-a64859683b60 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: d4674fbc-47b1-42d1-aaba-e86d46c51e8f] instance snapshotting [ 927.099710] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c459005-95e7-4e8e-9daa-e2a4a835a4e1 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.121664] env[61957]: DEBUG nova.scheduler.client.report [None req-d28d4db3-4ba8-4b6b-b020-dcb0a978b0ab tempest-InstanceActionsNegativeTestJSON-259524954 tempest-InstanceActionsNegativeTestJSON-259524954-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 927.125916] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eae32e05-b6e2-4d18-8b2a-4ce8436f82a3 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.146196] env[61957]: DEBUG oslo_vmware.api [None req-d61e8152-d9da-413c-b7fd-8eed0b01a69d tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52fe53a0-8f62-a509-fb57-f144dc869ea2, 'name': SearchDatastore_Task, 'duration_secs': 0.009566} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.146811] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d61e8152-d9da-413c-b7fd-8eed0b01a69d tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Releasing lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 927.146811] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-d61e8152-d9da-413c-b7fd-8eed0b01a69d tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore1] b598b8bb-919e-4404-b264-7b76161b0f79/b598b8bb-919e-4404-b264-7b76161b0f79.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 927.147083] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3f77e593-3e53-46d1-be00-04c9d9aa1883 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.155109] env[61957]: DEBUG oslo_vmware.api [None req-d61e8152-d9da-413c-b7fd-8eed0b01a69d tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Waiting for the task: (returnval){ [ 927.155109] env[61957]: value = "task-1277732" [ 927.155109] env[61957]: _type = "Task" [ 927.155109] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.163517] env[61957]: DEBUG oslo_vmware.api [None req-d61e8152-d9da-413c-b7fd-8eed0b01a69d tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': task-1277732, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.277363] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d1519ec7-d973-47ee-af77-990ab0a419d4 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 927.630882] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d28d4db3-4ba8-4b6b-b020-dcb0a978b0ab tempest-InstanceActionsNegativeTestJSON-259524954 tempest-InstanceActionsNegativeTestJSON-259524954-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.408s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 927.631622] env[61957]: DEBUG nova.compute.manager [None req-d28d4db3-4ba8-4b6b-b020-dcb0a978b0ab tempest-InstanceActionsNegativeTestJSON-259524954 tempest-InstanceActionsNegativeTestJSON-259524954-project-member] [instance: 7e2db5f6-1419-415e-b957-66cf2e3ec19d] Start building networks asynchronously for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 927.635777] env[61957]: DEBUG oslo_concurrency.lockutils [None req-1036c4ee-30e7-4291-a697-7231cf748564 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.519s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 927.637313] env[61957]: INFO nova.compute.claims [None req-1036c4ee-30e7-4291-a697-7231cf748564 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] [instance: 9b0053fe-420e-4b92-86a8-ed44f6a6ec49] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 927.642516] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-4cbf1ed6-4756-412e-9889-a64859683b60 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: d4674fbc-47b1-42d1-aaba-e86d46c51e8f] Creating Snapshot of the VM instance {{(pid=61957) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 927.642880] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-9de3aca9-1bf8-44f6-939c-77fcd17e6cff {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.652911] env[61957]: DEBUG oslo_vmware.api [None req-4cbf1ed6-4756-412e-9889-a64859683b60 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Waiting for the task: (returnval){ [ 927.652911] env[61957]: value = "task-1277733" [ 927.652911] env[61957]: _type = "Task" [ 927.652911] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.667456] env[61957]: DEBUG oslo_vmware.api [None req-d61e8152-d9da-413c-b7fd-8eed0b01a69d tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': task-1277732, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.671374] env[61957]: DEBUG oslo_vmware.api [None req-4cbf1ed6-4756-412e-9889-a64859683b60 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': task-1277733, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.908410] env[61957]: DEBUG nova.objects.instance [None req-2a26d3c1-8d56-45d2-a36f-db3655d7eccb tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Lazy-loading 'flavor' on Instance uuid 526ff179-62a6-4763-ab25-797617c4ed57 {{(pid=61957) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 928.011070] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-d2e4cd3a-a43a-4a0c-a755-6e27eb102a3e tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 39b8b1df-efdd-4c22-9f31-85c85be4f0eb] Volume attach. Driver type: vmdk {{(pid=61957) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 928.011470] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-d2e4cd3a-a43a-4a0c-a755-6e27eb102a3e tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 39b8b1df-efdd-4c22-9f31-85c85be4f0eb] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-274556', 'volume_id': 'f5184167-7e0d-480b-b064-0361333ea716', 'name': 'volume-f5184167-7e0d-480b-b064-0361333ea716', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '39b8b1df-efdd-4c22-9f31-85c85be4f0eb', 'attached_at': '', 'detached_at': '', 'volume_id': 'f5184167-7e0d-480b-b064-0361333ea716', 'serial': 'f5184167-7e0d-480b-b064-0361333ea716'} {{(pid=61957) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 928.012392] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5eca0e2-f1ae-4e84-b58f-9cc682315605 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.029283] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e89276c6-287d-4374-af08-b2a275276250 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.053898] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-d2e4cd3a-a43a-4a0c-a755-6e27eb102a3e tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 39b8b1df-efdd-4c22-9f31-85c85be4f0eb] Reconfiguring VM instance instance-0000004c to attach disk [datastore1] volume-f5184167-7e0d-480b-b064-0361333ea716/volume-f5184167-7e0d-480b-b064-0361333ea716.vmdk or device None with type thin {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 928.054529] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-53e60ac6-cb74-46c3-a3d7-128b860bd2c9 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.072700] env[61957]: DEBUG oslo_vmware.api [None req-d2e4cd3a-a43a-4a0c-a755-6e27eb102a3e tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Waiting for the task: (returnval){ [ 928.072700] env[61957]: value = "task-1277734" [ 928.072700] env[61957]: _type = "Task" [ 928.072700] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.080337] env[61957]: DEBUG oslo_vmware.api [None req-d2e4cd3a-a43a-4a0c-a755-6e27eb102a3e tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1277734, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.141971] env[61957]: DEBUG nova.compute.utils [None req-d28d4db3-4ba8-4b6b-b020-dcb0a978b0ab tempest-InstanceActionsNegativeTestJSON-259524954 tempest-InstanceActionsNegativeTestJSON-259524954-project-member] Using /dev/sd instead of None {{(pid=61957) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 928.145612] env[61957]: DEBUG nova.compute.manager [None req-d28d4db3-4ba8-4b6b-b020-dcb0a978b0ab tempest-InstanceActionsNegativeTestJSON-259524954 tempest-InstanceActionsNegativeTestJSON-259524954-project-member] [instance: 7e2db5f6-1419-415e-b957-66cf2e3ec19d] Allocating IP information in the background. {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 928.145612] env[61957]: DEBUG nova.network.neutron [None req-d28d4db3-4ba8-4b6b-b020-dcb0a978b0ab tempest-InstanceActionsNegativeTestJSON-259524954 tempest-InstanceActionsNegativeTestJSON-259524954-project-member] [instance: 7e2db5f6-1419-415e-b957-66cf2e3ec19d] allocate_for_instance() {{(pid=61957) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 928.163144] env[61957]: DEBUG oslo_vmware.api [None req-4cbf1ed6-4756-412e-9889-a64859683b60 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': task-1277733, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.168072] env[61957]: DEBUG oslo_vmware.api [None req-d61e8152-d9da-413c-b7fd-8eed0b01a69d tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': task-1277732, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.184486] env[61957]: DEBUG nova.policy [None req-d28d4db3-4ba8-4b6b-b020-dcb0a978b0ab tempest-InstanceActionsNegativeTestJSON-259524954 tempest-InstanceActionsNegativeTestJSON-259524954-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd65fe12b4a6949a1860134c173e393c7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '04f963c71c6e468abb28f1cd30dc0787', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61957) authorize /opt/stack/nova/nova/policy.py:203}} [ 928.273086] env[61957]: DEBUG oslo_concurrency.lockutils [None req-7678a123-82e2-420d-9405-a713c30b30cb tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Acquiring lock "08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 928.273418] env[61957]: DEBUG oslo_concurrency.lockutils [None req-7678a123-82e2-420d-9405-a713c30b30cb tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Lock "08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 928.413842] env[61957]: DEBUG oslo_concurrency.lockutils [None req-2a26d3c1-8d56-45d2-a36f-db3655d7eccb tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Acquiring lock "refresh_cache-526ff179-62a6-4763-ab25-797617c4ed57" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 928.414102] env[61957]: DEBUG oslo_concurrency.lockutils [None req-2a26d3c1-8d56-45d2-a36f-db3655d7eccb tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Acquired lock "refresh_cache-526ff179-62a6-4763-ab25-797617c4ed57" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 928.414211] env[61957]: DEBUG nova.network.neutron [None req-2a26d3c1-8d56-45d2-a36f-db3655d7eccb tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] [instance: 526ff179-62a6-4763-ab25-797617c4ed57] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 928.414392] env[61957]: DEBUG nova.objects.instance [None req-2a26d3c1-8d56-45d2-a36f-db3655d7eccb tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Lazy-loading 'info_cache' on Instance uuid 526ff179-62a6-4763-ab25-797617c4ed57 {{(pid=61957) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 928.459091] env[61957]: DEBUG nova.network.neutron [None req-d28d4db3-4ba8-4b6b-b020-dcb0a978b0ab tempest-InstanceActionsNegativeTestJSON-259524954 tempest-InstanceActionsNegativeTestJSON-259524954-project-member] [instance: 7e2db5f6-1419-415e-b957-66cf2e3ec19d] Successfully created port: e3abccf7-4ab1-4e6e-b0f4-8680d99be3bd {{(pid=61957) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 928.582970] env[61957]: DEBUG oslo_vmware.api [None req-d2e4cd3a-a43a-4a0c-a755-6e27eb102a3e tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1277734, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.646814] env[61957]: DEBUG nova.compute.manager [None req-d28d4db3-4ba8-4b6b-b020-dcb0a978b0ab tempest-InstanceActionsNegativeTestJSON-259524954 tempest-InstanceActionsNegativeTestJSON-259524954-project-member] [instance: 7e2db5f6-1419-415e-b957-66cf2e3ec19d] Start building block device mappings for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 928.667298] env[61957]: DEBUG oslo_vmware.api [None req-4cbf1ed6-4756-412e-9889-a64859683b60 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': task-1277733, 'name': CreateSnapshot_Task, 'duration_secs': 0.966904} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.668020] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-4cbf1ed6-4756-412e-9889-a64859683b60 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: d4674fbc-47b1-42d1-aaba-e86d46c51e8f] Created Snapshot of the VM instance {{(pid=61957) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 928.668738] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3321c2b4-4d6a-4817-aace-663c144e8410 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.674011] env[61957]: DEBUG oslo_vmware.api [None req-d61e8152-d9da-413c-b7fd-8eed0b01a69d tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': task-1277732, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.775991] env[61957]: DEBUG nova.compute.manager [None req-7678a123-82e2-420d-9405-a713c30b30cb tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2] Starting instance... {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 928.909321] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3108411e-5369-4f7f-bfb0-4960132442fb {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.916640] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-768e3c50-5720-4200-ba47-dcd1520f9273 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.920819] env[61957]: DEBUG nova.objects.base [None req-2a26d3c1-8d56-45d2-a36f-db3655d7eccb tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Object Instance<526ff179-62a6-4763-ab25-797617c4ed57> lazy-loaded attributes: flavor,info_cache {{(pid=61957) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 928.951800] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d6ea2af-712a-4c4e-bc0b-c8c72b9a40d1 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.959790] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7992302-66d4-4c44-8941-835002de8ea9 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.973776] env[61957]: DEBUG nova.compute.provider_tree [None req-1036c4ee-30e7-4291-a697-7231cf748564 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 929.082818] env[61957]: DEBUG oslo_vmware.api [None req-d2e4cd3a-a43a-4a0c-a755-6e27eb102a3e tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1277734, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.168812] env[61957]: DEBUG oslo_vmware.api [None req-d61e8152-d9da-413c-b7fd-8eed0b01a69d tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': task-1277732, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.761134} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.169125] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-d61e8152-d9da-413c-b7fd-8eed0b01a69d tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore1] b598b8bb-919e-4404-b264-7b76161b0f79/b598b8bb-919e-4404-b264-7b76161b0f79.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 929.169347] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-d61e8152-d9da-413c-b7fd-8eed0b01a69d tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: b598b8bb-919e-4404-b264-7b76161b0f79] Extending root virtual disk to 1048576 {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 929.169581] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-63906343-a695-4203-ba74-9c62e4a80ce8 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.176189] env[61957]: DEBUG oslo_vmware.api [None req-d61e8152-d9da-413c-b7fd-8eed0b01a69d tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Waiting for the task: (returnval){ [ 929.176189] env[61957]: value = "task-1277735" [ 929.176189] env[61957]: _type = "Task" [ 929.176189] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.838355] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-4cbf1ed6-4756-412e-9889-a64859683b60 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: d4674fbc-47b1-42d1-aaba-e86d46c51e8f] Creating linked-clone VM from snapshot {{(pid=61957) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 929.846018] env[61957]: DEBUG nova.scheduler.client.report [None req-1036c4ee-30e7-4291-a697-7231cf748564 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 929.847534] env[61957]: DEBUG nova.compute.manager [None req-d28d4db3-4ba8-4b6b-b020-dcb0a978b0ab tempest-InstanceActionsNegativeTestJSON-259524954 tempest-InstanceActionsNegativeTestJSON-259524954-project-member] [instance: 7e2db5f6-1419-415e-b957-66cf2e3ec19d] Start spawning the instance on the hypervisor. {{(pid=61957) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 929.849583] env[61957]: DEBUG oslo_vmware.api [None req-d61e8152-d9da-413c-b7fd-8eed0b01a69d tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': task-1277735, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.849768] env[61957]: WARNING oslo_vmware.common.loopingcall [None req-d61e8152-d9da-413c-b7fd-8eed0b01a69d tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] task run outlasted interval by 0.17337199999999997 sec [ 929.852225] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-727c39c2-5045-4845-bca1-fccf0f220abe {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.863124] env[61957]: DEBUG nova.compute.manager [req-61916fe8-1c6e-42bd-8a92-c94c5928b65b req-b7b99f8d-0ec4-43e3-8b91-7553f6316457 service nova] [instance: 7e2db5f6-1419-415e-b957-66cf2e3ec19d] Received event network-vif-plugged-e3abccf7-4ab1-4e6e-b0f4-8680d99be3bd {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 929.863550] env[61957]: DEBUG oslo_concurrency.lockutils [req-61916fe8-1c6e-42bd-8a92-c94c5928b65b req-b7b99f8d-0ec4-43e3-8b91-7553f6316457 service nova] Acquiring lock "7e2db5f6-1419-415e-b957-66cf2e3ec19d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 929.863550] env[61957]: DEBUG oslo_concurrency.lockutils [req-61916fe8-1c6e-42bd-8a92-c94c5928b65b req-b7b99f8d-0ec4-43e3-8b91-7553f6316457 service nova] Lock "7e2db5f6-1419-415e-b957-66cf2e3ec19d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 929.863676] env[61957]: DEBUG oslo_concurrency.lockutils [req-61916fe8-1c6e-42bd-8a92-c94c5928b65b req-b7b99f8d-0ec4-43e3-8b91-7553f6316457 service nova] Lock "7e2db5f6-1419-415e-b957-66cf2e3ec19d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 929.863799] env[61957]: DEBUG nova.compute.manager [req-61916fe8-1c6e-42bd-8a92-c94c5928b65b req-b7b99f8d-0ec4-43e3-8b91-7553f6316457 service nova] [instance: 7e2db5f6-1419-415e-b957-66cf2e3ec19d] No waiting events found dispatching network-vif-plugged-e3abccf7-4ab1-4e6e-b0f4-8680d99be3bd {{(pid=61957) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 929.863972] env[61957]: WARNING nova.compute.manager [req-61916fe8-1c6e-42bd-8a92-c94c5928b65b req-b7b99f8d-0ec4-43e3-8b91-7553f6316457 service nova] [instance: 7e2db5f6-1419-415e-b957-66cf2e3ec19d] Received unexpected event network-vif-plugged-e3abccf7-4ab1-4e6e-b0f4-8680d99be3bd for instance with vm_state building and task_state spawning. [ 929.874516] env[61957]: DEBUG oslo_vmware.api [None req-d61e8152-d9da-413c-b7fd-8eed0b01a69d tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': task-1277735, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064811} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.875150] env[61957]: DEBUG oslo_vmware.api [None req-d2e4cd3a-a43a-4a0c-a755-6e27eb102a3e tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1277734, 'name': ReconfigVM_Task, 'duration_secs': 1.169483} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.875961] env[61957]: DEBUG oslo_concurrency.lockutils [None req-7678a123-82e2-420d-9405-a713c30b30cb tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 929.878732] env[61957]: DEBUG nova.virt.hardware [None req-d28d4db3-4ba8-4b6b-b020-dcb0a978b0ab tempest-InstanceActionsNegativeTestJSON-259524954 tempest-InstanceActionsNegativeTestJSON-259524954-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T17:22:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T17:21:50Z,direct_url=,disk_format='vmdk',id=11c76a2c-f705-470a-ba9d-4657858bab38,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='81d3096d0f094373913fc3dc8f5c3c6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T17:21:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 929.878957] env[61957]: DEBUG nova.virt.hardware [None req-d28d4db3-4ba8-4b6b-b020-dcb0a978b0ab tempest-InstanceActionsNegativeTestJSON-259524954 tempest-InstanceActionsNegativeTestJSON-259524954-project-member] Flavor limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 929.879142] env[61957]: DEBUG nova.virt.hardware [None req-d28d4db3-4ba8-4b6b-b020-dcb0a978b0ab tempest-InstanceActionsNegativeTestJSON-259524954 tempest-InstanceActionsNegativeTestJSON-259524954-project-member] Image limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 929.879334] env[61957]: DEBUG nova.virt.hardware [None req-d28d4db3-4ba8-4b6b-b020-dcb0a978b0ab tempest-InstanceActionsNegativeTestJSON-259524954 tempest-InstanceActionsNegativeTestJSON-259524954-project-member] Flavor pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 929.879483] env[61957]: DEBUG nova.virt.hardware [None req-d28d4db3-4ba8-4b6b-b020-dcb0a978b0ab tempest-InstanceActionsNegativeTestJSON-259524954 tempest-InstanceActionsNegativeTestJSON-259524954-project-member] Image pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 929.879631] env[61957]: DEBUG nova.virt.hardware [None req-d28d4db3-4ba8-4b6b-b020-dcb0a978b0ab tempest-InstanceActionsNegativeTestJSON-259524954 tempest-InstanceActionsNegativeTestJSON-259524954-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 929.879835] env[61957]: DEBUG nova.virt.hardware [None req-d28d4db3-4ba8-4b6b-b020-dcb0a978b0ab tempest-InstanceActionsNegativeTestJSON-259524954 tempest-InstanceActionsNegativeTestJSON-259524954-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 929.879995] env[61957]: DEBUG nova.virt.hardware [None req-d28d4db3-4ba8-4b6b-b020-dcb0a978b0ab tempest-InstanceActionsNegativeTestJSON-259524954 tempest-InstanceActionsNegativeTestJSON-259524954-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 929.880352] env[61957]: DEBUG nova.virt.hardware [None req-d28d4db3-4ba8-4b6b-b020-dcb0a978b0ab tempest-InstanceActionsNegativeTestJSON-259524954 tempest-InstanceActionsNegativeTestJSON-259524954-project-member] Got 1 possible topologies {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 929.880352] env[61957]: DEBUG nova.virt.hardware [None req-d28d4db3-4ba8-4b6b-b020-dcb0a978b0ab tempest-InstanceActionsNegativeTestJSON-259524954 tempest-InstanceActionsNegativeTestJSON-259524954-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 929.880500] env[61957]: DEBUG nova.virt.hardware [None req-d28d4db3-4ba8-4b6b-b020-dcb0a978b0ab tempest-InstanceActionsNegativeTestJSON-259524954 tempest-InstanceActionsNegativeTestJSON-259524954-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 929.880788] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-d61e8152-d9da-413c-b7fd-8eed0b01a69d tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: b598b8bb-919e-4404-b264-7b76161b0f79] Extended root virtual disk {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 929.881069] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-d2e4cd3a-a43a-4a0c-a755-6e27eb102a3e tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 39b8b1df-efdd-4c22-9f31-85c85be4f0eb] Reconfigured VM instance instance-0000004c to attach disk [datastore1] volume-f5184167-7e0d-480b-b064-0361333ea716/volume-f5184167-7e0d-480b-b064-0361333ea716.vmdk or device None with type thin {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 929.885681] env[61957]: DEBUG oslo_vmware.api [None req-4cbf1ed6-4756-412e-9889-a64859683b60 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Waiting for the task: (returnval){ [ 929.885681] env[61957]: value = "task-1277736" [ 929.885681] env[61957]: _type = "Task" [ 929.885681] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.886328] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8900947-e9d5-42ae-968a-a3a17910594f {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.889153] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70d9f8d9-3fe6-4473-8e31-7e7be64a2eac {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.891999] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2783334a-ae16-4e42-9501-a9e1cc4b0c90 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.923448] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26bd17a7-cc51-4ddd-b068-9eb0e7edc477 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.936141] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-d61e8152-d9da-413c-b7fd-8eed0b01a69d tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: b598b8bb-919e-4404-b264-7b76161b0f79] Reconfiguring VM instance instance-0000004f to attach disk [datastore1] b598b8bb-919e-4404-b264-7b76161b0f79/b598b8bb-919e-4404-b264-7b76161b0f79.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 929.936405] env[61957]: DEBUG oslo_vmware.api [None req-4cbf1ed6-4756-412e-9889-a64859683b60 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': task-1277736, 'name': CloneVM_Task} progress is 16%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.936634] env[61957]: DEBUG oslo_vmware.api [None req-d2e4cd3a-a43a-4a0c-a755-6e27eb102a3e tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Waiting for the task: (returnval){ [ 929.936634] env[61957]: value = "task-1277737" [ 929.936634] env[61957]: _type = "Task" [ 929.936634] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.936832] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6b41db65-f15c-49d6-ad70-9e6d86b377cf {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.968208] env[61957]: DEBUG oslo_vmware.api [None req-d2e4cd3a-a43a-4a0c-a755-6e27eb102a3e tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1277737, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.969570] env[61957]: DEBUG oslo_vmware.api [None req-d61e8152-d9da-413c-b7fd-8eed0b01a69d tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Waiting for the task: (returnval){ [ 929.969570] env[61957]: value = "task-1277738" [ 929.969570] env[61957]: _type = "Task" [ 929.969570] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.979580] env[61957]: DEBUG oslo_vmware.api [None req-d61e8152-d9da-413c-b7fd-8eed0b01a69d tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': task-1277738, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.059420] env[61957]: DEBUG nova.network.neutron [None req-d28d4db3-4ba8-4b6b-b020-dcb0a978b0ab tempest-InstanceActionsNegativeTestJSON-259524954 tempest-InstanceActionsNegativeTestJSON-259524954-project-member] [instance: 7e2db5f6-1419-415e-b957-66cf2e3ec19d] Successfully updated port: e3abccf7-4ab1-4e6e-b0f4-8680d99be3bd {{(pid=61957) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 930.187442] env[61957]: DEBUG nova.network.neutron [None req-2a26d3c1-8d56-45d2-a36f-db3655d7eccb tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] [instance: 526ff179-62a6-4763-ab25-797617c4ed57] Updating instance_info_cache with network_info: [{"id": "25abeea6-dd92-49e0-b014-d35dd1c0f8ee", "address": "fa:16:3e:0a:44:84", "network": {"id": "8e60ab13-6dd3-4cda-9cdf-974ad8bebe1f", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1404238496-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.173", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ea955bd4d84548ff8e6a0c21542d6ee7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9f208df-1fb5-4403-9796-7fd19e4bfb85", "external-id": "cl2-zone-400", "segmentation_id": 400, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap25abeea6-dd", "ovs_interfaceid": "25abeea6-dd92-49e0-b014-d35dd1c0f8ee", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 930.353532] env[61957]: DEBUG oslo_concurrency.lockutils [None req-1036c4ee-30e7-4291-a697-7231cf748564 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.718s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 930.354203] env[61957]: DEBUG nova.compute.manager [None req-1036c4ee-30e7-4291-a697-7231cf748564 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] [instance: 9b0053fe-420e-4b92-86a8-ed44f6a6ec49] Start building networks asynchronously for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 930.357533] env[61957]: DEBUG oslo_concurrency.lockutils [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.638s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 930.359258] env[61957]: INFO nova.compute.claims [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: 50e09db7-9f8f-452b-8232-bd7473e9e63a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 930.411610] env[61957]: DEBUG oslo_vmware.api [None req-4cbf1ed6-4756-412e-9889-a64859683b60 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': task-1277736, 'name': CloneVM_Task} progress is 94%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.460985] env[61957]: DEBUG oslo_vmware.api [None req-d2e4cd3a-a43a-4a0c-a755-6e27eb102a3e tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1277737, 'name': ReconfigVM_Task, 'duration_secs': 0.147036} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.461544] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-d2e4cd3a-a43a-4a0c-a755-6e27eb102a3e tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 39b8b1df-efdd-4c22-9f31-85c85be4f0eb] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-274556', 'volume_id': 'f5184167-7e0d-480b-b064-0361333ea716', 'name': 'volume-f5184167-7e0d-480b-b064-0361333ea716', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '39b8b1df-efdd-4c22-9f31-85c85be4f0eb', 'attached_at': '', 'detached_at': '', 'volume_id': 'f5184167-7e0d-480b-b064-0361333ea716', 'serial': 'f5184167-7e0d-480b-b064-0361333ea716'} {{(pid=61957) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 930.479546] env[61957]: DEBUG oslo_vmware.api [None req-d61e8152-d9da-413c-b7fd-8eed0b01a69d tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': task-1277738, 'name': ReconfigVM_Task, 'duration_secs': 0.339821} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.479546] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-d61e8152-d9da-413c-b7fd-8eed0b01a69d tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: b598b8bb-919e-4404-b264-7b76161b0f79] Reconfigured VM instance instance-0000004f to attach disk [datastore1] b598b8bb-919e-4404-b264-7b76161b0f79/b598b8bb-919e-4404-b264-7b76161b0f79.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 930.480058] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f0b05b9a-c540-4d0b-aea4-b148e35c6124 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.487167] env[61957]: DEBUG oslo_vmware.api [None req-d61e8152-d9da-413c-b7fd-8eed0b01a69d tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Waiting for the task: (returnval){ [ 930.487167] env[61957]: value = "task-1277739" [ 930.487167] env[61957]: _type = "Task" [ 930.487167] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.495887] env[61957]: DEBUG oslo_vmware.api [None req-d61e8152-d9da-413c-b7fd-8eed0b01a69d tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': task-1277739, 'name': Rename_Task} progress is 5%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.562126] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d28d4db3-4ba8-4b6b-b020-dcb0a978b0ab tempest-InstanceActionsNegativeTestJSON-259524954 tempest-InstanceActionsNegativeTestJSON-259524954-project-member] Acquiring lock "refresh_cache-7e2db5f6-1419-415e-b957-66cf2e3ec19d" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 930.562276] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d28d4db3-4ba8-4b6b-b020-dcb0a978b0ab tempest-InstanceActionsNegativeTestJSON-259524954 tempest-InstanceActionsNegativeTestJSON-259524954-project-member] Acquired lock "refresh_cache-7e2db5f6-1419-415e-b957-66cf2e3ec19d" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 930.562369] env[61957]: DEBUG nova.network.neutron [None req-d28d4db3-4ba8-4b6b-b020-dcb0a978b0ab tempest-InstanceActionsNegativeTestJSON-259524954 tempest-InstanceActionsNegativeTestJSON-259524954-project-member] [instance: 7e2db5f6-1419-415e-b957-66cf2e3ec19d] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 930.690747] env[61957]: DEBUG oslo_concurrency.lockutils [None req-2a26d3c1-8d56-45d2-a36f-db3655d7eccb tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Releasing lock "refresh_cache-526ff179-62a6-4763-ab25-797617c4ed57" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 930.863870] env[61957]: DEBUG nova.compute.utils [None req-1036c4ee-30e7-4291-a697-7231cf748564 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Using /dev/sd instead of None {{(pid=61957) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 930.867513] env[61957]: DEBUG nova.compute.manager [None req-1036c4ee-30e7-4291-a697-7231cf748564 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] [instance: 9b0053fe-420e-4b92-86a8-ed44f6a6ec49] Allocating IP information in the background. {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 930.867686] env[61957]: DEBUG nova.network.neutron [None req-1036c4ee-30e7-4291-a697-7231cf748564 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] [instance: 9b0053fe-420e-4b92-86a8-ed44f6a6ec49] allocate_for_instance() {{(pid=61957) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 930.911756] env[61957]: DEBUG oslo_vmware.api [None req-4cbf1ed6-4756-412e-9889-a64859683b60 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': task-1277736, 'name': CloneVM_Task} progress is 95%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.915277] env[61957]: DEBUG nova.policy [None req-1036c4ee-30e7-4291-a697-7231cf748564 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6653838d4b8a42618da8f8dc35362cdc', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f7c20ab3c822423fa837a03f066e4239', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61957) authorize /opt/stack/nova/nova/policy.py:203}} [ 930.999122] env[61957]: DEBUG oslo_vmware.api [None req-d61e8152-d9da-413c-b7fd-8eed0b01a69d tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': task-1277739, 'name': Rename_Task, 'duration_secs': 0.137166} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.999122] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-d61e8152-d9da-413c-b7fd-8eed0b01a69d tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: b598b8bb-919e-4404-b264-7b76161b0f79] Powering on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 930.999122] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-fc290d52-5630-4408-89cf-3fe2f26497a9 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.005599] env[61957]: DEBUG oslo_vmware.api [None req-d61e8152-d9da-413c-b7fd-8eed0b01a69d tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Waiting for the task: (returnval){ [ 931.005599] env[61957]: value = "task-1277740" [ 931.005599] env[61957]: _type = "Task" [ 931.005599] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.011998] env[61957]: DEBUG oslo_vmware.api [None req-d61e8152-d9da-413c-b7fd-8eed0b01a69d tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': task-1277740, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.107276] env[61957]: DEBUG nova.network.neutron [None req-d28d4db3-4ba8-4b6b-b020-dcb0a978b0ab tempest-InstanceActionsNegativeTestJSON-259524954 tempest-InstanceActionsNegativeTestJSON-259524954-project-member] [instance: 7e2db5f6-1419-415e-b957-66cf2e3ec19d] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 931.194547] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-2a26d3c1-8d56-45d2-a36f-db3655d7eccb tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] [instance: 526ff179-62a6-4763-ab25-797617c4ed57] Powering on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 931.194866] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-be94b6b2-58fb-407d-bf44-c67fa823a7f6 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.203248] env[61957]: DEBUG oslo_vmware.api [None req-2a26d3c1-8d56-45d2-a36f-db3655d7eccb tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Waiting for the task: (returnval){ [ 931.203248] env[61957]: value = "task-1277741" [ 931.203248] env[61957]: _type = "Task" [ 931.203248] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.215394] env[61957]: DEBUG oslo_vmware.api [None req-2a26d3c1-8d56-45d2-a36f-db3655d7eccb tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Task: {'id': task-1277741, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.301955] env[61957]: DEBUG nova.network.neutron [None req-1036c4ee-30e7-4291-a697-7231cf748564 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] [instance: 9b0053fe-420e-4b92-86a8-ed44f6a6ec49] Successfully created port: 219230f0-0b14-4be9-81bb-67a765d8ced9 {{(pid=61957) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 931.346224] env[61957]: DEBUG nova.network.neutron [None req-d28d4db3-4ba8-4b6b-b020-dcb0a978b0ab tempest-InstanceActionsNegativeTestJSON-259524954 tempest-InstanceActionsNegativeTestJSON-259524954-project-member] [instance: 7e2db5f6-1419-415e-b957-66cf2e3ec19d] Updating instance_info_cache with network_info: [{"id": "e3abccf7-4ab1-4e6e-b0f4-8680d99be3bd", "address": "fa:16:3e:d9:83:4e", "network": {"id": "1b83a988-4a65-40f1-b9c5-b46cce5a6d2d", "bridge": "br-int", "label": "tempest-InstanceActionsNegativeTestJSON-280659726-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "04f963c71c6e468abb28f1cd30dc0787", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "20641d67-1612-4b9c-8924-7a77df9c8e6d", "external-id": "nsx-vlan-transportzone-884", "segmentation_id": 884, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape3abccf7-4a", "ovs_interfaceid": "e3abccf7-4ab1-4e6e-b0f4-8680d99be3bd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 931.368534] env[61957]: DEBUG nova.compute.manager [None req-1036c4ee-30e7-4291-a697-7231cf748564 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] [instance: 9b0053fe-420e-4b92-86a8-ed44f6a6ec49] Start building block device mappings for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 931.414951] env[61957]: DEBUG oslo_vmware.api [None req-4cbf1ed6-4756-412e-9889-a64859683b60 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': task-1277736, 'name': CloneVM_Task, 'duration_secs': 1.335732} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.415370] env[61957]: INFO nova.virt.vmwareapi.vmops [None req-4cbf1ed6-4756-412e-9889-a64859683b60 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: d4674fbc-47b1-42d1-aaba-e86d46c51e8f] Created linked-clone VM from snapshot [ 931.416252] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d561b56-1ecc-4a7a-ae01-1e0c3bc2213a {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.427816] env[61957]: DEBUG nova.virt.vmwareapi.images [None req-4cbf1ed6-4756-412e-9889-a64859683b60 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: d4674fbc-47b1-42d1-aaba-e86d46c51e8f] Uploading image fa797977-e3da-4f16-8445-e20c97e9e8ae {{(pid=61957) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 931.456043] env[61957]: DEBUG oslo_vmware.rw_handles [None req-4cbf1ed6-4756-412e-9889-a64859683b60 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 931.456043] env[61957]: value = "vm-274559" [ 931.456043] env[61957]: _type = "VirtualMachine" [ 931.456043] env[61957]: }. {{(pid=61957) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 931.456381] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-b0af3809-5e6e-4414-8757-b7a3b3a697bb {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.464695] env[61957]: DEBUG oslo_vmware.rw_handles [None req-4cbf1ed6-4756-412e-9889-a64859683b60 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Lease: (returnval){ [ 931.464695] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52724956-17af-b8de-fe74-7be599b7175d" [ 931.464695] env[61957]: _type = "HttpNfcLease" [ 931.464695] env[61957]: } obtained for exporting VM: (result){ [ 931.464695] env[61957]: value = "vm-274559" [ 931.464695] env[61957]: _type = "VirtualMachine" [ 931.464695] env[61957]: }. {{(pid=61957) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 931.464982] env[61957]: DEBUG oslo_vmware.api [None req-4cbf1ed6-4756-412e-9889-a64859683b60 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Waiting for the lease: (returnval){ [ 931.464982] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52724956-17af-b8de-fe74-7be599b7175d" [ 931.464982] env[61957]: _type = "HttpNfcLease" [ 931.464982] env[61957]: } to be ready. {{(pid=61957) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 931.472592] env[61957]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 931.472592] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52724956-17af-b8de-fe74-7be599b7175d" [ 931.472592] env[61957]: _type = "HttpNfcLease" [ 931.472592] env[61957]: } is initializing. {{(pid=61957) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 931.503410] env[61957]: DEBUG nova.objects.instance [None req-d2e4cd3a-a43a-4a0c-a755-6e27eb102a3e tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Lazy-loading 'flavor' on Instance uuid 39b8b1df-efdd-4c22-9f31-85c85be4f0eb {{(pid=61957) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 931.518070] env[61957]: DEBUG oslo_vmware.api [None req-d61e8152-d9da-413c-b7fd-8eed0b01a69d tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': task-1277740, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.719064] env[61957]: DEBUG oslo_vmware.api [None req-2a26d3c1-8d56-45d2-a36f-db3655d7eccb tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Task: {'id': task-1277741, 'name': PowerOnVM_Task, 'duration_secs': 0.429236} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.719369] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-2a26d3c1-8d56-45d2-a36f-db3655d7eccb tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] [instance: 526ff179-62a6-4763-ab25-797617c4ed57] Powered on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 931.720362] env[61957]: DEBUG nova.compute.manager [None req-2a26d3c1-8d56-45d2-a36f-db3655d7eccb tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] [instance: 526ff179-62a6-4763-ab25-797617c4ed57] Checking state {{(pid=61957) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 931.721159] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21591c36-30cc-4261-bc54-bf7ab1aaef8d {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.732160] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78e52905-f279-4856-9861-0b38281eb6c7 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.741603] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36c880b7-a37b-4dcc-92be-0ccf8fd566a0 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.774697] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9773fc35-d4e6-4bd4-b6e8-2eda1735fd9f {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.783736] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d8bde0c-58b7-4ef8-9683-306461946673 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.799131] env[61957]: DEBUG nova.compute.provider_tree [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 931.848506] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d28d4db3-4ba8-4b6b-b020-dcb0a978b0ab tempest-InstanceActionsNegativeTestJSON-259524954 tempest-InstanceActionsNegativeTestJSON-259524954-project-member] Releasing lock "refresh_cache-7e2db5f6-1419-415e-b957-66cf2e3ec19d" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 931.848837] env[61957]: DEBUG nova.compute.manager [None req-d28d4db3-4ba8-4b6b-b020-dcb0a978b0ab tempest-InstanceActionsNegativeTestJSON-259524954 tempest-InstanceActionsNegativeTestJSON-259524954-project-member] [instance: 7e2db5f6-1419-415e-b957-66cf2e3ec19d] Instance network_info: |[{"id": "e3abccf7-4ab1-4e6e-b0f4-8680d99be3bd", "address": "fa:16:3e:d9:83:4e", "network": {"id": "1b83a988-4a65-40f1-b9c5-b46cce5a6d2d", "bridge": "br-int", "label": "tempest-InstanceActionsNegativeTestJSON-280659726-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "04f963c71c6e468abb28f1cd30dc0787", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "20641d67-1612-4b9c-8924-7a77df9c8e6d", "external-id": "nsx-vlan-transportzone-884", "segmentation_id": 884, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape3abccf7-4a", "ovs_interfaceid": "e3abccf7-4ab1-4e6e-b0f4-8680d99be3bd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 931.850291] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-d28d4db3-4ba8-4b6b-b020-dcb0a978b0ab tempest-InstanceActionsNegativeTestJSON-259524954 tempest-InstanceActionsNegativeTestJSON-259524954-project-member] [instance: 7e2db5f6-1419-415e-b957-66cf2e3ec19d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d9:83:4e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '20641d67-1612-4b9c-8924-7a77df9c8e6d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e3abccf7-4ab1-4e6e-b0f4-8680d99be3bd', 'vif_model': 'vmxnet3'}] {{(pid=61957) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 931.857827] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-d28d4db3-4ba8-4b6b-b020-dcb0a978b0ab tempest-InstanceActionsNegativeTestJSON-259524954 tempest-InstanceActionsNegativeTestJSON-259524954-project-member] Creating folder: Project (04f963c71c6e468abb28f1cd30dc0787). Parent ref: group-v274445. {{(pid=61957) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 931.858148] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0d7b091d-f603-41a2-8260-8f1aa0b6213f {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.871028] env[61957]: INFO nova.virt.vmwareapi.vm_util [None req-d28d4db3-4ba8-4b6b-b020-dcb0a978b0ab tempest-InstanceActionsNegativeTestJSON-259524954 tempest-InstanceActionsNegativeTestJSON-259524954-project-member] Created folder: Project (04f963c71c6e468abb28f1cd30dc0787) in parent group-v274445. [ 931.871028] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-d28d4db3-4ba8-4b6b-b020-dcb0a978b0ab tempest-InstanceActionsNegativeTestJSON-259524954 tempest-InstanceActionsNegativeTestJSON-259524954-project-member] Creating folder: Instances. Parent ref: group-v274560. {{(pid=61957) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 931.871329] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-84851786-ef6b-42ea-a4ea-bbfd85b8403a {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.873295] env[61957]: DEBUG oslo_concurrency.lockutils [None req-5d7e1ad8-1e31-4b98-a89f-2f6c8c624f1c tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Acquiring lock "39b8b1df-efdd-4c22-9f31-85c85be4f0eb" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 931.882933] env[61957]: DEBUG nova.compute.manager [req-2e04e125-e057-4f6c-8ac2-30c1e5fc3730 req-5d82a577-07d7-4f5d-b1b0-47e0d21d4e9c service nova] [instance: 7e2db5f6-1419-415e-b957-66cf2e3ec19d] Received event network-changed-e3abccf7-4ab1-4e6e-b0f4-8680d99be3bd {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 931.883172] env[61957]: DEBUG nova.compute.manager [req-2e04e125-e057-4f6c-8ac2-30c1e5fc3730 req-5d82a577-07d7-4f5d-b1b0-47e0d21d4e9c service nova] [instance: 7e2db5f6-1419-415e-b957-66cf2e3ec19d] Refreshing instance network info cache due to event network-changed-e3abccf7-4ab1-4e6e-b0f4-8680d99be3bd. {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 931.883479] env[61957]: DEBUG oslo_concurrency.lockutils [req-2e04e125-e057-4f6c-8ac2-30c1e5fc3730 req-5d82a577-07d7-4f5d-b1b0-47e0d21d4e9c service nova] Acquiring lock "refresh_cache-7e2db5f6-1419-415e-b957-66cf2e3ec19d" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 931.883639] env[61957]: DEBUG oslo_concurrency.lockutils [req-2e04e125-e057-4f6c-8ac2-30c1e5fc3730 req-5d82a577-07d7-4f5d-b1b0-47e0d21d4e9c service nova] Acquired lock "refresh_cache-7e2db5f6-1419-415e-b957-66cf2e3ec19d" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 931.883860] env[61957]: DEBUG nova.network.neutron [req-2e04e125-e057-4f6c-8ac2-30c1e5fc3730 req-5d82a577-07d7-4f5d-b1b0-47e0d21d4e9c service nova] [instance: 7e2db5f6-1419-415e-b957-66cf2e3ec19d] Refreshing network info cache for port e3abccf7-4ab1-4e6e-b0f4-8680d99be3bd {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 931.889059] env[61957]: INFO nova.virt.vmwareapi.vm_util [None req-d28d4db3-4ba8-4b6b-b020-dcb0a978b0ab tempest-InstanceActionsNegativeTestJSON-259524954 tempest-InstanceActionsNegativeTestJSON-259524954-project-member] Created folder: Instances in parent group-v274560. [ 931.889580] env[61957]: DEBUG oslo.service.loopingcall [None req-d28d4db3-4ba8-4b6b-b020-dcb0a978b0ab tempest-InstanceActionsNegativeTestJSON-259524954 tempest-InstanceActionsNegativeTestJSON-259524954-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 931.889689] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7e2db5f6-1419-415e-b957-66cf2e3ec19d] Creating VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 931.889885] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ce81fc8c-bf6a-4c80-94b9-be14423779d8 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.910493] env[61957]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 931.910493] env[61957]: value = "task-1277745" [ 931.910493] env[61957]: _type = "Task" [ 931.910493] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.919722] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277745, 'name': CreateVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.973800] env[61957]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 931.973800] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52724956-17af-b8de-fe74-7be599b7175d" [ 931.973800] env[61957]: _type = "HttpNfcLease" [ 931.973800] env[61957]: } is ready. {{(pid=61957) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 931.974051] env[61957]: DEBUG oslo_vmware.rw_handles [None req-4cbf1ed6-4756-412e-9889-a64859683b60 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 931.974051] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52724956-17af-b8de-fe74-7be599b7175d" [ 931.974051] env[61957]: _type = "HttpNfcLease" [ 931.974051] env[61957]: }. {{(pid=61957) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 931.974803] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fb63bac-6793-4513-971c-1d34ffd79efc {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.983131] env[61957]: DEBUG oslo_vmware.rw_handles [None req-4cbf1ed6-4756-412e-9889-a64859683b60 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/521952db-7e22-536a-9cfd-d20a4fa400ec/disk-0.vmdk from lease info. {{(pid=61957) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 931.983333] env[61957]: DEBUG oslo_vmware.rw_handles [None req-4cbf1ed6-4756-412e-9889-a64859683b60 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/521952db-7e22-536a-9cfd-d20a4fa400ec/disk-0.vmdk for reading. {{(pid=61957) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 932.046656] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d2e4cd3a-a43a-4a0c-a755-6e27eb102a3e tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Lock "39b8b1df-efdd-4c22-9f31-85c85be4f0eb" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 8.659s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 932.047773] env[61957]: DEBUG oslo_concurrency.lockutils [None req-5d7e1ad8-1e31-4b98-a89f-2f6c8c624f1c tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Lock "39b8b1df-efdd-4c22-9f31-85c85be4f0eb" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.174s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 932.048744] env[61957]: DEBUG oslo_concurrency.lockutils [None req-5d7e1ad8-1e31-4b98-a89f-2f6c8c624f1c tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Acquiring lock "39b8b1df-efdd-4c22-9f31-85c85be4f0eb-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 932.048744] env[61957]: DEBUG oslo_concurrency.lockutils [None req-5d7e1ad8-1e31-4b98-a89f-2f6c8c624f1c tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Lock "39b8b1df-efdd-4c22-9f31-85c85be4f0eb-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 932.048744] env[61957]: DEBUG oslo_concurrency.lockutils [None req-5d7e1ad8-1e31-4b98-a89f-2f6c8c624f1c tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Lock "39b8b1df-efdd-4c22-9f31-85c85be4f0eb-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 932.050537] env[61957]: INFO nova.compute.manager [None req-5d7e1ad8-1e31-4b98-a89f-2f6c8c624f1c tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 39b8b1df-efdd-4c22-9f31-85c85be4f0eb] Terminating instance [ 932.052785] env[61957]: DEBUG nova.compute.manager [None req-5d7e1ad8-1e31-4b98-a89f-2f6c8c624f1c tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 39b8b1df-efdd-4c22-9f31-85c85be4f0eb] Start destroying the instance on the hypervisor. {{(pid=61957) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 932.053036] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-5d7e1ad8-1e31-4b98-a89f-2f6c8c624f1c tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 39b8b1df-efdd-4c22-9f31-85c85be4f0eb] Powering off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 932.053287] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e216088f-f1f0-46f7-b03a-ed5edc5589b3 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.058034] env[61957]: DEBUG oslo_vmware.api [None req-d61e8152-d9da-413c-b7fd-8eed0b01a69d tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': task-1277740, 'name': PowerOnVM_Task, 'duration_secs': 0.52136} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.058640] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-d61e8152-d9da-413c-b7fd-8eed0b01a69d tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: b598b8bb-919e-4404-b264-7b76161b0f79] Powered on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 932.058859] env[61957]: INFO nova.compute.manager [None req-d61e8152-d9da-413c-b7fd-8eed0b01a69d tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: b598b8bb-919e-4404-b264-7b76161b0f79] Took 9.29 seconds to spawn the instance on the hypervisor. [ 932.059054] env[61957]: DEBUG nova.compute.manager [None req-d61e8152-d9da-413c-b7fd-8eed0b01a69d tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: b598b8bb-919e-4404-b264-7b76161b0f79] Checking state {{(pid=61957) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 932.059872] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e65ced0d-ce7c-46fe-b6d5-e57a43199ef0 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.063915] env[61957]: DEBUG oslo_vmware.api [None req-5d7e1ad8-1e31-4b98-a89f-2f6c8c624f1c tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Waiting for the task: (returnval){ [ 932.063915] env[61957]: value = "task-1277746" [ 932.063915] env[61957]: _type = "Task" [ 932.063915] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.078251] env[61957]: DEBUG oslo_vmware.api [None req-5d7e1ad8-1e31-4b98-a89f-2f6c8c624f1c tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1277746, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.097359] env[61957]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-266cb1d5-0ea2-401c-83ff-050cad321212 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.302744] env[61957]: DEBUG nova.scheduler.client.report [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 932.382194] env[61957]: DEBUG nova.compute.manager [None req-1036c4ee-30e7-4291-a697-7231cf748564 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] [instance: 9b0053fe-420e-4b92-86a8-ed44f6a6ec49] Start spawning the instance on the hypervisor. {{(pid=61957) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 932.405443] env[61957]: DEBUG nova.virt.hardware [None req-1036c4ee-30e7-4291-a697-7231cf748564 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T17:22:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T17:21:50Z,direct_url=,disk_format='vmdk',id=11c76a2c-f705-470a-ba9d-4657858bab38,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='81d3096d0f094373913fc3dc8f5c3c6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T17:21:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 932.405914] env[61957]: DEBUG nova.virt.hardware [None req-1036c4ee-30e7-4291-a697-7231cf748564 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Flavor limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 932.406255] env[61957]: DEBUG nova.virt.hardware [None req-1036c4ee-30e7-4291-a697-7231cf748564 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Image limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 932.406548] env[61957]: DEBUG nova.virt.hardware [None req-1036c4ee-30e7-4291-a697-7231cf748564 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Flavor pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 932.406749] env[61957]: DEBUG nova.virt.hardware [None req-1036c4ee-30e7-4291-a697-7231cf748564 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Image pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 932.406927] env[61957]: DEBUG nova.virt.hardware [None req-1036c4ee-30e7-4291-a697-7231cf748564 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 932.407195] env[61957]: DEBUG nova.virt.hardware [None req-1036c4ee-30e7-4291-a697-7231cf748564 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 932.407982] env[61957]: DEBUG nova.virt.hardware [None req-1036c4ee-30e7-4291-a697-7231cf748564 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 932.407982] env[61957]: DEBUG nova.virt.hardware [None req-1036c4ee-30e7-4291-a697-7231cf748564 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Got 1 possible topologies {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 932.408250] env[61957]: DEBUG nova.virt.hardware [None req-1036c4ee-30e7-4291-a697-7231cf748564 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 932.408508] env[61957]: DEBUG nova.virt.hardware [None req-1036c4ee-30e7-4291-a697-7231cf748564 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 932.409513] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26dba135-1b04-4294-af64-48df80bdf050 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.425021] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e521b312-ccb4-4c21-be25-b24e2b82b17a {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.429087] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277745, 'name': CreateVM_Task, 'duration_secs': 0.424405} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.431779] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7e2db5f6-1419-415e-b957-66cf2e3ec19d] Created VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 932.432912] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d28d4db3-4ba8-4b6b-b020-dcb0a978b0ab tempest-InstanceActionsNegativeTestJSON-259524954 tempest-InstanceActionsNegativeTestJSON-259524954-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 932.433106] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d28d4db3-4ba8-4b6b-b020-dcb0a978b0ab tempest-InstanceActionsNegativeTestJSON-259524954 tempest-InstanceActionsNegativeTestJSON-259524954-project-member] Acquired lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 932.433438] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d28d4db3-4ba8-4b6b-b020-dcb0a978b0ab tempest-InstanceActionsNegativeTestJSON-259524954 tempest-InstanceActionsNegativeTestJSON-259524954-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 932.441694] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a7497d40-be4b-43b2-a4bd-1a184221ecb0 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.447440] env[61957]: DEBUG oslo_vmware.api [None req-d28d4db3-4ba8-4b6b-b020-dcb0a978b0ab tempest-InstanceActionsNegativeTestJSON-259524954 tempest-InstanceActionsNegativeTestJSON-259524954-project-member] Waiting for the task: (returnval){ [ 932.447440] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]529dbd70-8e27-e0bd-a472-e59678ef0e44" [ 932.447440] env[61957]: _type = "Task" [ 932.447440] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.458962] env[61957]: DEBUG oslo_vmware.api [None req-d28d4db3-4ba8-4b6b-b020-dcb0a978b0ab tempest-InstanceActionsNegativeTestJSON-259524954 tempest-InstanceActionsNegativeTestJSON-259524954-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]529dbd70-8e27-e0bd-a472-e59678ef0e44, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.579086] env[61957]: DEBUG oslo_vmware.api [None req-5d7e1ad8-1e31-4b98-a89f-2f6c8c624f1c tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1277746, 'name': PowerOffVM_Task, 'duration_secs': 0.309285} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.581483] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-5d7e1ad8-1e31-4b98-a89f-2f6c8c624f1c tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 39b8b1df-efdd-4c22-9f31-85c85be4f0eb] Powered off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 932.581563] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-5d7e1ad8-1e31-4b98-a89f-2f6c8c624f1c tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 39b8b1df-efdd-4c22-9f31-85c85be4f0eb] Volume detach. Driver type: vmdk {{(pid=61957) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 932.582891] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-5d7e1ad8-1e31-4b98-a89f-2f6c8c624f1c tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 39b8b1df-efdd-4c22-9f31-85c85be4f0eb] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-274556', 'volume_id': 'f5184167-7e0d-480b-b064-0361333ea716', 'name': 'volume-f5184167-7e0d-480b-b064-0361333ea716', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '39b8b1df-efdd-4c22-9f31-85c85be4f0eb', 'attached_at': '', 'detached_at': '', 'volume_id': 'f5184167-7e0d-480b-b064-0361333ea716', 'serial': 'f5184167-7e0d-480b-b064-0361333ea716'} {{(pid=61957) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 932.582891] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39607c6e-c52a-475d-bc27-80397f67645a {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.587118] env[61957]: INFO nova.compute.manager [None req-d61e8152-d9da-413c-b7fd-8eed0b01a69d tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: b598b8bb-919e-4404-b264-7b76161b0f79] Took 42.09 seconds to build instance. [ 932.614657] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-913382ea-9da4-4544-bb11-b593f7237ac1 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.623719] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-810647ff-025a-456b-ba13-4d99eab908f2 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.647429] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64baad2a-464b-40c3-87ab-b8e1262397c3 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.668557] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-5d7e1ad8-1e31-4b98-a89f-2f6c8c624f1c tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] The volume has not been displaced from its original location: [datastore1] volume-f5184167-7e0d-480b-b064-0361333ea716/volume-f5184167-7e0d-480b-b064-0361333ea716.vmdk. No consolidation needed. {{(pid=61957) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 932.674180] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-5d7e1ad8-1e31-4b98-a89f-2f6c8c624f1c tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 39b8b1df-efdd-4c22-9f31-85c85be4f0eb] Reconfiguring VM instance instance-0000004c to detach disk 2001 {{(pid=61957) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 932.675509] env[61957]: DEBUG nova.network.neutron [req-2e04e125-e057-4f6c-8ac2-30c1e5fc3730 req-5d82a577-07d7-4f5d-b1b0-47e0d21d4e9c service nova] [instance: 7e2db5f6-1419-415e-b957-66cf2e3ec19d] Updated VIF entry in instance network info cache for port e3abccf7-4ab1-4e6e-b0f4-8680d99be3bd. {{(pid=61957) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 932.675961] env[61957]: DEBUG nova.network.neutron [req-2e04e125-e057-4f6c-8ac2-30c1e5fc3730 req-5d82a577-07d7-4f5d-b1b0-47e0d21d4e9c service nova] [instance: 7e2db5f6-1419-415e-b957-66cf2e3ec19d] Updating instance_info_cache with network_info: [{"id": "e3abccf7-4ab1-4e6e-b0f4-8680d99be3bd", "address": "fa:16:3e:d9:83:4e", "network": {"id": "1b83a988-4a65-40f1-b9c5-b46cce5a6d2d", "bridge": "br-int", "label": "tempest-InstanceActionsNegativeTestJSON-280659726-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "04f963c71c6e468abb28f1cd30dc0787", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "20641d67-1612-4b9c-8924-7a77df9c8e6d", "external-id": "nsx-vlan-transportzone-884", "segmentation_id": 884, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape3abccf7-4a", "ovs_interfaceid": "e3abccf7-4ab1-4e6e-b0f4-8680d99be3bd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 932.677536] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-00ea2762-70ad-4667-981e-a989f7c79cdc {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.693182] env[61957]: DEBUG oslo_concurrency.lockutils [req-2e04e125-e057-4f6c-8ac2-30c1e5fc3730 req-5d82a577-07d7-4f5d-b1b0-47e0d21d4e9c service nova] Releasing lock "refresh_cache-7e2db5f6-1419-415e-b957-66cf2e3ec19d" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 932.699633] env[61957]: DEBUG oslo_vmware.api [None req-5d7e1ad8-1e31-4b98-a89f-2f6c8c624f1c tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Waiting for the task: (returnval){ [ 932.699633] env[61957]: value = "task-1277747" [ 932.699633] env[61957]: _type = "Task" [ 932.699633] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.709617] env[61957]: DEBUG oslo_vmware.api [None req-5d7e1ad8-1e31-4b98-a89f-2f6c8c624f1c tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1277747, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.808532] env[61957]: DEBUG oslo_concurrency.lockutils [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.451s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 932.809422] env[61957]: DEBUG nova.compute.manager [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: 50e09db7-9f8f-452b-8232-bd7473e9e63a] Start building networks asynchronously for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 932.812654] env[61957]: DEBUG oslo_concurrency.lockutils [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.280s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 932.815379] env[61957]: INFO nova.compute.claims [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: 2e32aae5-edd7-402f-98ad-75a93d26f7a1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 932.930931] env[61957]: DEBUG nova.network.neutron [None req-1036c4ee-30e7-4291-a697-7231cf748564 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] [instance: 9b0053fe-420e-4b92-86a8-ed44f6a6ec49] Successfully updated port: 219230f0-0b14-4be9-81bb-67a765d8ced9 {{(pid=61957) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 932.960902] env[61957]: DEBUG oslo_vmware.api [None req-d28d4db3-4ba8-4b6b-b020-dcb0a978b0ab tempest-InstanceActionsNegativeTestJSON-259524954 tempest-InstanceActionsNegativeTestJSON-259524954-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]529dbd70-8e27-e0bd-a472-e59678ef0e44, 'name': SearchDatastore_Task, 'duration_secs': 0.011176} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.962078] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d28d4db3-4ba8-4b6b-b020-dcb0a978b0ab tempest-InstanceActionsNegativeTestJSON-259524954 tempest-InstanceActionsNegativeTestJSON-259524954-project-member] Releasing lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 932.962078] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-d28d4db3-4ba8-4b6b-b020-dcb0a978b0ab tempest-InstanceActionsNegativeTestJSON-259524954 tempest-InstanceActionsNegativeTestJSON-259524954-project-member] [instance: 7e2db5f6-1419-415e-b957-66cf2e3ec19d] Processing image 11c76a2c-f705-470a-ba9d-4657858bab38 {{(pid=61957) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 932.963049] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d28d4db3-4ba8-4b6b-b020-dcb0a978b0ab tempest-InstanceActionsNegativeTestJSON-259524954 tempest-InstanceActionsNegativeTestJSON-259524954-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 932.963049] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d28d4db3-4ba8-4b6b-b020-dcb0a978b0ab tempest-InstanceActionsNegativeTestJSON-259524954 tempest-InstanceActionsNegativeTestJSON-259524954-project-member] Acquired lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 932.963049] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-d28d4db3-4ba8-4b6b-b020-dcb0a978b0ab tempest-InstanceActionsNegativeTestJSON-259524954 tempest-InstanceActionsNegativeTestJSON-259524954-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 932.963561] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-040378a1-255f-4495-bb05-c086c120ceff {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.972330] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-d28d4db3-4ba8-4b6b-b020-dcb0a978b0ab tempest-InstanceActionsNegativeTestJSON-259524954 tempest-InstanceActionsNegativeTestJSON-259524954-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 932.972597] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-d28d4db3-4ba8-4b6b-b020-dcb0a978b0ab tempest-InstanceActionsNegativeTestJSON-259524954 tempest-InstanceActionsNegativeTestJSON-259524954-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61957) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 932.973697] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ae79bc1f-a278-4c9b-8905-d612734963a9 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.980248] env[61957]: DEBUG oslo_vmware.api [None req-d28d4db3-4ba8-4b6b-b020-dcb0a978b0ab tempest-InstanceActionsNegativeTestJSON-259524954 tempest-InstanceActionsNegativeTestJSON-259524954-project-member] Waiting for the task: (returnval){ [ 932.980248] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52cb30e7-e73f-ef44-7466-8f5c6e00118c" [ 932.980248] env[61957]: _type = "Task" [ 932.980248] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.988402] env[61957]: DEBUG oslo_vmware.api [None req-d28d4db3-4ba8-4b6b-b020-dcb0a978b0ab tempest-InstanceActionsNegativeTestJSON-259524954 tempest-InstanceActionsNegativeTestJSON-259524954-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52cb30e7-e73f-ef44-7466-8f5c6e00118c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.092286] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d61e8152-d9da-413c-b7fd-8eed0b01a69d tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Lock "b598b8bb-919e-4404-b264-7b76161b0f79" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 43.604s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 933.213210] env[61957]: DEBUG oslo_vmware.api [None req-5d7e1ad8-1e31-4b98-a89f-2f6c8c624f1c tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1277747, 'name': ReconfigVM_Task, 'duration_secs': 0.23887} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.213783] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-5d7e1ad8-1e31-4b98-a89f-2f6c8c624f1c tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 39b8b1df-efdd-4c22-9f31-85c85be4f0eb] Reconfigured VM instance instance-0000004c to detach disk 2001 {{(pid=61957) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 933.219321] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-60ad25ff-8e41-4bae-b269-ab54f456dc6e {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.236812] env[61957]: DEBUG oslo_vmware.api [None req-5d7e1ad8-1e31-4b98-a89f-2f6c8c624f1c tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Waiting for the task: (returnval){ [ 933.236812] env[61957]: value = "task-1277748" [ 933.236812] env[61957]: _type = "Task" [ 933.236812] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.246213] env[61957]: DEBUG oslo_vmware.api [None req-5d7e1ad8-1e31-4b98-a89f-2f6c8c624f1c tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1277748, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.319962] env[61957]: DEBUG nova.compute.utils [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Using /dev/sd instead of None {{(pid=61957) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 933.324928] env[61957]: DEBUG nova.compute.manager [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: 50e09db7-9f8f-452b-8232-bd7473e9e63a] Allocating IP information in the background. {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 933.325398] env[61957]: DEBUG nova.network.neutron [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: 50e09db7-9f8f-452b-8232-bd7473e9e63a] allocate_for_instance() {{(pid=61957) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 933.381947] env[61957]: DEBUG nova.policy [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9b453c031e7a4294b14dd48e62e1fd30', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '55dd45b4291b456f83a446dfa87f550a', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61957) authorize /opt/stack/nova/nova/policy.py:203}} [ 933.436617] env[61957]: DEBUG oslo_concurrency.lockutils [None req-1036c4ee-30e7-4291-a697-7231cf748564 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Acquiring lock "refresh_cache-9b0053fe-420e-4b92-86a8-ed44f6a6ec49" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 933.436847] env[61957]: DEBUG oslo_concurrency.lockutils [None req-1036c4ee-30e7-4291-a697-7231cf748564 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Acquired lock "refresh_cache-9b0053fe-420e-4b92-86a8-ed44f6a6ec49" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 933.437264] env[61957]: DEBUG nova.network.neutron [None req-1036c4ee-30e7-4291-a697-7231cf748564 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] [instance: 9b0053fe-420e-4b92-86a8-ed44f6a6ec49] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 933.492372] env[61957]: DEBUG oslo_vmware.api [None req-d28d4db3-4ba8-4b6b-b020-dcb0a978b0ab tempest-InstanceActionsNegativeTestJSON-259524954 tempest-InstanceActionsNegativeTestJSON-259524954-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52cb30e7-e73f-ef44-7466-8f5c6e00118c, 'name': SearchDatastore_Task, 'duration_secs': 0.01084} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.492372] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-01f62efc-6039-4e08-a848-daff60da7f0e {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.497564] env[61957]: DEBUG oslo_vmware.api [None req-d28d4db3-4ba8-4b6b-b020-dcb0a978b0ab tempest-InstanceActionsNegativeTestJSON-259524954 tempest-InstanceActionsNegativeTestJSON-259524954-project-member] Waiting for the task: (returnval){ [ 933.497564] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]5247298d-d254-7be9-f308-b38175b3ff85" [ 933.497564] env[61957]: _type = "Task" [ 933.497564] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.505304] env[61957]: DEBUG oslo_vmware.api [None req-d28d4db3-4ba8-4b6b-b020-dcb0a978b0ab tempest-InstanceActionsNegativeTestJSON-259524954 tempest-InstanceActionsNegativeTestJSON-259524954-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]5247298d-d254-7be9-f308-b38175b3ff85, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.657550] env[61957]: DEBUG nova.network.neutron [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: 50e09db7-9f8f-452b-8232-bd7473e9e63a] Successfully created port: 224ed13c-938e-4dcf-b555-22fd97cd0210 {{(pid=61957) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 933.749858] env[61957]: DEBUG oslo_vmware.api [None req-5d7e1ad8-1e31-4b98-a89f-2f6c8c624f1c tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1277748, 'name': ReconfigVM_Task, 'duration_secs': 0.195383} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.750265] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-5d7e1ad8-1e31-4b98-a89f-2f6c8c624f1c tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 39b8b1df-efdd-4c22-9f31-85c85be4f0eb] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-274556', 'volume_id': 'f5184167-7e0d-480b-b064-0361333ea716', 'name': 'volume-f5184167-7e0d-480b-b064-0361333ea716', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '39b8b1df-efdd-4c22-9f31-85c85be4f0eb', 'attached_at': '', 'detached_at': '', 'volume_id': 'f5184167-7e0d-480b-b064-0361333ea716', 'serial': 'f5184167-7e0d-480b-b064-0361333ea716'} {{(pid=61957) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 933.750664] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-5d7e1ad8-1e31-4b98-a89f-2f6c8c624f1c tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 39b8b1df-efdd-4c22-9f31-85c85be4f0eb] Destroying instance {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 933.751713] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51916bf2-0a47-425e-95ce-26a3974c744f {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.759279] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-5d7e1ad8-1e31-4b98-a89f-2f6c8c624f1c tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 39b8b1df-efdd-4c22-9f31-85c85be4f0eb] Unregistering the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 933.759555] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-49fab8d8-c98f-4ec3-9557-70b99b5a1c22 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.764193] env[61957]: DEBUG oslo_concurrency.lockutils [None req-73ed12d4-fc9a-46df-a24a-8dbe4ec86160 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Acquiring lock "b598b8bb-919e-4404-b264-7b76161b0f79" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 933.765097] env[61957]: DEBUG oslo_concurrency.lockutils [None req-73ed12d4-fc9a-46df-a24a-8dbe4ec86160 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Lock "b598b8bb-919e-4404-b264-7b76161b0f79" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 933.765412] env[61957]: DEBUG oslo_concurrency.lockutils [None req-73ed12d4-fc9a-46df-a24a-8dbe4ec86160 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Acquiring lock "b598b8bb-919e-4404-b264-7b76161b0f79-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 933.765682] env[61957]: DEBUG oslo_concurrency.lockutils [None req-73ed12d4-fc9a-46df-a24a-8dbe4ec86160 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Lock "b598b8bb-919e-4404-b264-7b76161b0f79-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 933.765905] env[61957]: DEBUG oslo_concurrency.lockutils [None req-73ed12d4-fc9a-46df-a24a-8dbe4ec86160 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Lock "b598b8bb-919e-4404-b264-7b76161b0f79-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 933.768663] env[61957]: INFO nova.compute.manager [None req-73ed12d4-fc9a-46df-a24a-8dbe4ec86160 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: b598b8bb-919e-4404-b264-7b76161b0f79] Terminating instance [ 933.771072] env[61957]: DEBUG nova.compute.manager [None req-73ed12d4-fc9a-46df-a24a-8dbe4ec86160 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: b598b8bb-919e-4404-b264-7b76161b0f79] Start destroying the instance on the hypervisor. {{(pid=61957) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 933.771312] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-73ed12d4-fc9a-46df-a24a-8dbe4ec86160 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: b598b8bb-919e-4404-b264-7b76161b0f79] Destroying instance {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 933.772383] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9656ad2b-8788-4c95-b70a-7ae4f92e0e37 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.781549] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-73ed12d4-fc9a-46df-a24a-8dbe4ec86160 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: b598b8bb-919e-4404-b264-7b76161b0f79] Powering off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 933.781832] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9809bfc4-3a30-4486-9b1e-b50827f9cc53 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.789760] env[61957]: DEBUG oslo_vmware.api [None req-73ed12d4-fc9a-46df-a24a-8dbe4ec86160 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Waiting for the task: (returnval){ [ 933.789760] env[61957]: value = "task-1277750" [ 933.789760] env[61957]: _type = "Task" [ 933.789760] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.799070] env[61957]: DEBUG oslo_vmware.api [None req-73ed12d4-fc9a-46df-a24a-8dbe4ec86160 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': task-1277750, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.826388] env[61957]: DEBUG nova.compute.manager [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: 50e09db7-9f8f-452b-8232-bd7473e9e63a] Start building block device mappings for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 933.853163] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-5d7e1ad8-1e31-4b98-a89f-2f6c8c624f1c tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 39b8b1df-efdd-4c22-9f31-85c85be4f0eb] Unregistered the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 933.853499] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-5d7e1ad8-1e31-4b98-a89f-2f6c8c624f1c tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 39b8b1df-efdd-4c22-9f31-85c85be4f0eb] Deleting contents of the VM from datastore datastore1 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 933.853655] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-5d7e1ad8-1e31-4b98-a89f-2f6c8c624f1c tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Deleting the datastore file [datastore1] 39b8b1df-efdd-4c22-9f31-85c85be4f0eb {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 933.853950] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9fdced47-b126-4699-9d21-8ee5f9046e58 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.864594] env[61957]: DEBUG oslo_vmware.api [None req-5d7e1ad8-1e31-4b98-a89f-2f6c8c624f1c tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Waiting for the task: (returnval){ [ 933.864594] env[61957]: value = "task-1277751" [ 933.864594] env[61957]: _type = "Task" [ 933.864594] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.876228] env[61957]: DEBUG oslo_vmware.api [None req-5d7e1ad8-1e31-4b98-a89f-2f6c8c624f1c tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1277751, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.921230] env[61957]: DEBUG nova.compute.manager [req-58febe9d-b1b7-4127-8032-888f8c10d80c req-6d5ff5b6-36ef-4283-912b-779e11ade0a7 service nova] [instance: 9b0053fe-420e-4b92-86a8-ed44f6a6ec49] Received event network-vif-plugged-219230f0-0b14-4be9-81bb-67a765d8ced9 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 933.921544] env[61957]: DEBUG oslo_concurrency.lockutils [req-58febe9d-b1b7-4127-8032-888f8c10d80c req-6d5ff5b6-36ef-4283-912b-779e11ade0a7 service nova] Acquiring lock "9b0053fe-420e-4b92-86a8-ed44f6a6ec49-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 933.922172] env[61957]: DEBUG oslo_concurrency.lockutils [req-58febe9d-b1b7-4127-8032-888f8c10d80c req-6d5ff5b6-36ef-4283-912b-779e11ade0a7 service nova] Lock "9b0053fe-420e-4b92-86a8-ed44f6a6ec49-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 933.922390] env[61957]: DEBUG oslo_concurrency.lockutils [req-58febe9d-b1b7-4127-8032-888f8c10d80c req-6d5ff5b6-36ef-4283-912b-779e11ade0a7 service nova] Lock "9b0053fe-420e-4b92-86a8-ed44f6a6ec49-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 933.922596] env[61957]: DEBUG nova.compute.manager [req-58febe9d-b1b7-4127-8032-888f8c10d80c req-6d5ff5b6-36ef-4283-912b-779e11ade0a7 service nova] [instance: 9b0053fe-420e-4b92-86a8-ed44f6a6ec49] No waiting events found dispatching network-vif-plugged-219230f0-0b14-4be9-81bb-67a765d8ced9 {{(pid=61957) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 933.922762] env[61957]: WARNING nova.compute.manager [req-58febe9d-b1b7-4127-8032-888f8c10d80c req-6d5ff5b6-36ef-4283-912b-779e11ade0a7 service nova] [instance: 9b0053fe-420e-4b92-86a8-ed44f6a6ec49] Received unexpected event network-vif-plugged-219230f0-0b14-4be9-81bb-67a765d8ced9 for instance with vm_state building and task_state spawning. [ 933.923134] env[61957]: DEBUG nova.compute.manager [req-58febe9d-b1b7-4127-8032-888f8c10d80c req-6d5ff5b6-36ef-4283-912b-779e11ade0a7 service nova] [instance: 9b0053fe-420e-4b92-86a8-ed44f6a6ec49] Received event network-changed-219230f0-0b14-4be9-81bb-67a765d8ced9 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 933.923134] env[61957]: DEBUG nova.compute.manager [req-58febe9d-b1b7-4127-8032-888f8c10d80c req-6d5ff5b6-36ef-4283-912b-779e11ade0a7 service nova] [instance: 9b0053fe-420e-4b92-86a8-ed44f6a6ec49] Refreshing instance network info cache due to event network-changed-219230f0-0b14-4be9-81bb-67a765d8ced9. {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 933.923324] env[61957]: DEBUG oslo_concurrency.lockutils [req-58febe9d-b1b7-4127-8032-888f8c10d80c req-6d5ff5b6-36ef-4283-912b-779e11ade0a7 service nova] Acquiring lock "refresh_cache-9b0053fe-420e-4b92-86a8-ed44f6a6ec49" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 933.999209] env[61957]: DEBUG nova.network.neutron [None req-1036c4ee-30e7-4291-a697-7231cf748564 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] [instance: 9b0053fe-420e-4b92-86a8-ed44f6a6ec49] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 934.014639] env[61957]: DEBUG oslo_vmware.api [None req-d28d4db3-4ba8-4b6b-b020-dcb0a978b0ab tempest-InstanceActionsNegativeTestJSON-259524954 tempest-InstanceActionsNegativeTestJSON-259524954-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]5247298d-d254-7be9-f308-b38175b3ff85, 'name': SearchDatastore_Task, 'duration_secs': 0.011347} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.014983] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d28d4db3-4ba8-4b6b-b020-dcb0a978b0ab tempest-InstanceActionsNegativeTestJSON-259524954 tempest-InstanceActionsNegativeTestJSON-259524954-project-member] Releasing lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 934.015291] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-d28d4db3-4ba8-4b6b-b020-dcb0a978b0ab tempest-InstanceActionsNegativeTestJSON-259524954 tempest-InstanceActionsNegativeTestJSON-259524954-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore1] 7e2db5f6-1419-415e-b957-66cf2e3ec19d/7e2db5f6-1419-415e-b957-66cf2e3ec19d.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 934.015612] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7775f79d-0cf1-4393-97af-e60e73231260 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.023640] env[61957]: DEBUG oslo_vmware.api [None req-d28d4db3-4ba8-4b6b-b020-dcb0a978b0ab tempest-InstanceActionsNegativeTestJSON-259524954 tempest-InstanceActionsNegativeTestJSON-259524954-project-member] Waiting for the task: (returnval){ [ 934.023640] env[61957]: value = "task-1277752" [ 934.023640] env[61957]: _type = "Task" [ 934.023640] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.041616] env[61957]: DEBUG oslo_vmware.api [None req-d28d4db3-4ba8-4b6b-b020-dcb0a978b0ab tempest-InstanceActionsNegativeTestJSON-259524954 tempest-InstanceActionsNegativeTestJSON-259524954-project-member] Task: {'id': task-1277752, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.179974] env[61957]: DEBUG nova.network.neutron [None req-1036c4ee-30e7-4291-a697-7231cf748564 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] [instance: 9b0053fe-420e-4b92-86a8-ed44f6a6ec49] Updating instance_info_cache with network_info: [{"id": "219230f0-0b14-4be9-81bb-67a765d8ced9", "address": "fa:16:3e:f6:9e:ce", "network": {"id": "cf378e9c-eaae-4168-bd60-9818fcfaf405", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1243326371-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f7c20ab3c822423fa837a03f066e4239", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8cb478a6-872c-4a90-a8db-526b374e82ce", "external-id": "nsx-vlan-transportzone-835", "segmentation_id": 835, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap219230f0-0b", "ovs_interfaceid": "219230f0-0b14-4be9-81bb-67a765d8ced9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 934.234025] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8faf2f9f-a68e-4be4-8519-ce2d9354e51c {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.241916] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bf1c3eb-9d66-4074-9159-379332aba13e {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.274859] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26d2439c-a749-4fb3-bb77-c8e7e0f14847 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.283560] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-579794e0-db65-4de7-81e5-19622075b2d4 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.299596] env[61957]: DEBUG nova.compute.provider_tree [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 934.310497] env[61957]: DEBUG oslo_vmware.api [None req-73ed12d4-fc9a-46df-a24a-8dbe4ec86160 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': task-1277750, 'name': PowerOffVM_Task, 'duration_secs': 0.294251} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.311426] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-73ed12d4-fc9a-46df-a24a-8dbe4ec86160 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: b598b8bb-919e-4404-b264-7b76161b0f79] Powered off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 934.311594] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-73ed12d4-fc9a-46df-a24a-8dbe4ec86160 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: b598b8bb-919e-4404-b264-7b76161b0f79] Unregistering the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 934.312735] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-49da2411-f8e0-432b-acfa-afeb6757ebae {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.376611] env[61957]: DEBUG oslo_vmware.api [None req-5d7e1ad8-1e31-4b98-a89f-2f6c8c624f1c tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1277751, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.219602} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.377751] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-5d7e1ad8-1e31-4b98-a89f-2f6c8c624f1c tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Deleted the datastore file {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 934.377892] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-5d7e1ad8-1e31-4b98-a89f-2f6c8c624f1c tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 39b8b1df-efdd-4c22-9f31-85c85be4f0eb] Deleted contents of the VM from datastore datastore1 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 934.378030] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-5d7e1ad8-1e31-4b98-a89f-2f6c8c624f1c tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 39b8b1df-efdd-4c22-9f31-85c85be4f0eb] Instance destroyed {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 934.378211] env[61957]: INFO nova.compute.manager [None req-5d7e1ad8-1e31-4b98-a89f-2f6c8c624f1c tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 39b8b1df-efdd-4c22-9f31-85c85be4f0eb] Took 2.33 seconds to destroy the instance on the hypervisor. [ 934.378675] env[61957]: DEBUG oslo.service.loopingcall [None req-5d7e1ad8-1e31-4b98-a89f-2f6c8c624f1c tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 934.379544] env[61957]: DEBUG nova.compute.manager [-] [instance: 39b8b1df-efdd-4c22-9f31-85c85be4f0eb] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 934.379544] env[61957]: DEBUG nova.network.neutron [-] [instance: 39b8b1df-efdd-4c22-9f31-85c85be4f0eb] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 934.386986] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-73ed12d4-fc9a-46df-a24a-8dbe4ec86160 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: b598b8bb-919e-4404-b264-7b76161b0f79] Unregistered the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 934.387242] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-73ed12d4-fc9a-46df-a24a-8dbe4ec86160 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: b598b8bb-919e-4404-b264-7b76161b0f79] Deleting contents of the VM from datastore datastore1 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 934.387429] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-73ed12d4-fc9a-46df-a24a-8dbe4ec86160 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Deleting the datastore file [datastore1] b598b8bb-919e-4404-b264-7b76161b0f79 {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 934.387699] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1ecd2c66-a9be-4744-8b9b-5547d1439645 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.395333] env[61957]: DEBUG oslo_vmware.api [None req-73ed12d4-fc9a-46df-a24a-8dbe4ec86160 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Waiting for the task: (returnval){ [ 934.395333] env[61957]: value = "task-1277754" [ 934.395333] env[61957]: _type = "Task" [ 934.395333] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.404386] env[61957]: DEBUG oslo_vmware.api [None req-73ed12d4-fc9a-46df-a24a-8dbe4ec86160 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': task-1277754, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.538743] env[61957]: DEBUG oslo_vmware.api [None req-d28d4db3-4ba8-4b6b-b020-dcb0a978b0ab tempest-InstanceActionsNegativeTestJSON-259524954 tempest-InstanceActionsNegativeTestJSON-259524954-project-member] Task: {'id': task-1277752, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.687554] env[61957]: DEBUG oslo_concurrency.lockutils [None req-1036c4ee-30e7-4291-a697-7231cf748564 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Releasing lock "refresh_cache-9b0053fe-420e-4b92-86a8-ed44f6a6ec49" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 934.688046] env[61957]: DEBUG nova.compute.manager [None req-1036c4ee-30e7-4291-a697-7231cf748564 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] [instance: 9b0053fe-420e-4b92-86a8-ed44f6a6ec49] Instance network_info: |[{"id": "219230f0-0b14-4be9-81bb-67a765d8ced9", "address": "fa:16:3e:f6:9e:ce", "network": {"id": "cf378e9c-eaae-4168-bd60-9818fcfaf405", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1243326371-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f7c20ab3c822423fa837a03f066e4239", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8cb478a6-872c-4a90-a8db-526b374e82ce", "external-id": "nsx-vlan-transportzone-835", "segmentation_id": 835, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap219230f0-0b", "ovs_interfaceid": "219230f0-0b14-4be9-81bb-67a765d8ced9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 934.688539] env[61957]: DEBUG oslo_concurrency.lockutils [req-58febe9d-b1b7-4127-8032-888f8c10d80c req-6d5ff5b6-36ef-4283-912b-779e11ade0a7 service nova] Acquired lock "refresh_cache-9b0053fe-420e-4b92-86a8-ed44f6a6ec49" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 934.688848] env[61957]: DEBUG nova.network.neutron [req-58febe9d-b1b7-4127-8032-888f8c10d80c req-6d5ff5b6-36ef-4283-912b-779e11ade0a7 service nova] [instance: 9b0053fe-420e-4b92-86a8-ed44f6a6ec49] Refreshing network info cache for port 219230f0-0b14-4be9-81bb-67a765d8ced9 {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 934.690621] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-1036c4ee-30e7-4291-a697-7231cf748564 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] [instance: 9b0053fe-420e-4b92-86a8-ed44f6a6ec49] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f6:9e:ce', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8cb478a6-872c-4a90-a8db-526b374e82ce', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '219230f0-0b14-4be9-81bb-67a765d8ced9', 'vif_model': 'vmxnet3'}] {{(pid=61957) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 934.712364] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-1036c4ee-30e7-4291-a697-7231cf748564 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Creating folder: Project (f7c20ab3c822423fa837a03f066e4239). Parent ref: group-v274445. {{(pid=61957) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 934.714142] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a7c2b01d-8c23-4dc4-b1fe-e082bd2da5b1 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.727325] env[61957]: INFO nova.virt.vmwareapi.vm_util [None req-1036c4ee-30e7-4291-a697-7231cf748564 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Created folder: Project (f7c20ab3c822423fa837a03f066e4239) in parent group-v274445. [ 934.727578] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-1036c4ee-30e7-4291-a697-7231cf748564 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Creating folder: Instances. Parent ref: group-v274563. {{(pid=61957) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 934.727839] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a3e2c167-1e28-41e8-afdd-058643c7b193 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.740477] env[61957]: INFO nova.virt.vmwareapi.vm_util [None req-1036c4ee-30e7-4291-a697-7231cf748564 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Created folder: Instances in parent group-v274563. [ 934.740811] env[61957]: DEBUG oslo.service.loopingcall [None req-1036c4ee-30e7-4291-a697-7231cf748564 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 934.741125] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9b0053fe-420e-4b92-86a8-ed44f6a6ec49] Creating VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 934.741456] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a45bde23-9574-4f35-82c2-a919fdedf085 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.768207] env[61957]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 934.768207] env[61957]: value = "task-1277757" [ 934.768207] env[61957]: _type = "Task" [ 934.768207] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.779362] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277757, 'name': CreateVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.806612] env[61957]: DEBUG nova.scheduler.client.report [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 934.839812] env[61957]: DEBUG nova.compute.manager [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: 50e09db7-9f8f-452b-8232-bd7473e9e63a] Start spawning the instance on the hypervisor. {{(pid=61957) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 934.873585] env[61957]: DEBUG nova.virt.hardware [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T17:22:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T17:21:50Z,direct_url=,disk_format='vmdk',id=11c76a2c-f705-470a-ba9d-4657858bab38,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='81d3096d0f094373913fc3dc8f5c3c6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T17:21:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 934.874313] env[61957]: DEBUG nova.virt.hardware [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Flavor limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 934.874313] env[61957]: DEBUG nova.virt.hardware [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Image limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 934.874534] env[61957]: DEBUG nova.virt.hardware [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Flavor pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 934.874534] env[61957]: DEBUG nova.virt.hardware [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Image pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 934.874723] env[61957]: DEBUG nova.virt.hardware [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 934.874948] env[61957]: DEBUG nova.virt.hardware [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 934.875559] env[61957]: DEBUG nova.virt.hardware [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 934.875559] env[61957]: DEBUG nova.virt.hardware [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Got 1 possible topologies {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 934.875559] env[61957]: DEBUG nova.virt.hardware [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 934.876108] env[61957]: DEBUG nova.virt.hardware [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 934.876911] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9360a70-0b4e-4ef6-838b-15aa5e96d540 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.894426] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c822270-2818-4e36-8879-cefd2cf01092 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.925170] env[61957]: DEBUG oslo_vmware.api [None req-73ed12d4-fc9a-46df-a24a-8dbe4ec86160 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': task-1277754, 'name': DeleteDatastoreFile_Task} progress is 100%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.039025] env[61957]: DEBUG oslo_vmware.api [None req-d28d4db3-4ba8-4b6b-b020-dcb0a978b0ab tempest-InstanceActionsNegativeTestJSON-259524954 tempest-InstanceActionsNegativeTestJSON-259524954-project-member] Task: {'id': task-1277752, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.891434} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.039334] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-d28d4db3-4ba8-4b6b-b020-dcb0a978b0ab tempest-InstanceActionsNegativeTestJSON-259524954 tempest-InstanceActionsNegativeTestJSON-259524954-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore1] 7e2db5f6-1419-415e-b957-66cf2e3ec19d/7e2db5f6-1419-415e-b957-66cf2e3ec19d.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 935.039543] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-d28d4db3-4ba8-4b6b-b020-dcb0a978b0ab tempest-InstanceActionsNegativeTestJSON-259524954 tempest-InstanceActionsNegativeTestJSON-259524954-project-member] [instance: 7e2db5f6-1419-415e-b957-66cf2e3ec19d] Extending root virtual disk to 1048576 {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 935.039801] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-088bb2c8-090a-4796-9cee-c97b441783dc {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.047368] env[61957]: DEBUG oslo_vmware.api [None req-d28d4db3-4ba8-4b6b-b020-dcb0a978b0ab tempest-InstanceActionsNegativeTestJSON-259524954 tempest-InstanceActionsNegativeTestJSON-259524954-project-member] Waiting for the task: (returnval){ [ 935.047368] env[61957]: value = "task-1277758" [ 935.047368] env[61957]: _type = "Task" [ 935.047368] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.056761] env[61957]: DEBUG oslo_vmware.api [None req-d28d4db3-4ba8-4b6b-b020-dcb0a978b0ab tempest-InstanceActionsNegativeTestJSON-259524954 tempest-InstanceActionsNegativeTestJSON-259524954-project-member] Task: {'id': task-1277758, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.275619] env[61957]: DEBUG nova.network.neutron [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: 50e09db7-9f8f-452b-8232-bd7473e9e63a] Successfully updated port: 224ed13c-938e-4dcf-b555-22fd97cd0210 {{(pid=61957) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 935.281320] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277757, 'name': CreateVM_Task, 'duration_secs': 0.507684} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.281683] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9b0053fe-420e-4b92-86a8-ed44f6a6ec49] Created VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 935.282626] env[61957]: DEBUG oslo_concurrency.lockutils [None req-1036c4ee-30e7-4291-a697-7231cf748564 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 935.282821] env[61957]: DEBUG oslo_concurrency.lockutils [None req-1036c4ee-30e7-4291-a697-7231cf748564 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Acquired lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 935.283184] env[61957]: DEBUG oslo_concurrency.lockutils [None req-1036c4ee-30e7-4291-a697-7231cf748564 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 935.283453] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b5c26b5d-30ac-431b-bfaa-ffe9b987b877 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.290945] env[61957]: DEBUG oslo_vmware.api [None req-1036c4ee-30e7-4291-a697-7231cf748564 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Waiting for the task: (returnval){ [ 935.290945] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52fa3e44-7dc5-5db6-7a81-900ec2575b46" [ 935.290945] env[61957]: _type = "Task" [ 935.290945] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.301719] env[61957]: DEBUG oslo_vmware.api [None req-1036c4ee-30e7-4291-a697-7231cf748564 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52fa3e44-7dc5-5db6-7a81-900ec2575b46, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.312704] env[61957]: DEBUG oslo_concurrency.lockutils [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.500s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 935.313281] env[61957]: DEBUG nova.compute.manager [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: 2e32aae5-edd7-402f-98ad-75a93d26f7a1] Start building networks asynchronously for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 935.315920] env[61957]: DEBUG oslo_concurrency.lockutils [None req-35c8e096-fa6b-43c8-874d-91a51735ac8e tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 22.455s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 935.316125] env[61957]: DEBUG nova.objects.instance [None req-35c8e096-fa6b-43c8-874d-91a51735ac8e tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] [instance: a8ca1ad6-d636-4fa2-b0ac-53b020e392ce] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61957) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 935.403953] env[61957]: DEBUG nova.network.neutron [-] [instance: 39b8b1df-efdd-4c22-9f31-85c85be4f0eb] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 935.416661] env[61957]: DEBUG oslo_vmware.api [None req-73ed12d4-fc9a-46df-a24a-8dbe4ec86160 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': task-1277754, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.523623} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.417066] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-73ed12d4-fc9a-46df-a24a-8dbe4ec86160 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Deleted the datastore file {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 935.417925] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-73ed12d4-fc9a-46df-a24a-8dbe4ec86160 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: b598b8bb-919e-4404-b264-7b76161b0f79] Deleted contents of the VM from datastore datastore1 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 935.417925] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-73ed12d4-fc9a-46df-a24a-8dbe4ec86160 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: b598b8bb-919e-4404-b264-7b76161b0f79] Instance destroyed {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 935.418080] env[61957]: INFO nova.compute.manager [None req-73ed12d4-fc9a-46df-a24a-8dbe4ec86160 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: b598b8bb-919e-4404-b264-7b76161b0f79] Took 1.65 seconds to destroy the instance on the hypervisor. [ 935.418339] env[61957]: DEBUG oslo.service.loopingcall [None req-73ed12d4-fc9a-46df-a24a-8dbe4ec86160 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 935.419413] env[61957]: DEBUG nova.compute.manager [-] [instance: b598b8bb-919e-4404-b264-7b76161b0f79] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 935.419413] env[61957]: DEBUG nova.network.neutron [-] [instance: b598b8bb-919e-4404-b264-7b76161b0f79] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 935.501488] env[61957]: DEBUG nova.network.neutron [req-58febe9d-b1b7-4127-8032-888f8c10d80c req-6d5ff5b6-36ef-4283-912b-779e11ade0a7 service nova] [instance: 9b0053fe-420e-4b92-86a8-ed44f6a6ec49] Updated VIF entry in instance network info cache for port 219230f0-0b14-4be9-81bb-67a765d8ced9. {{(pid=61957) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 935.501488] env[61957]: DEBUG nova.network.neutron [req-58febe9d-b1b7-4127-8032-888f8c10d80c req-6d5ff5b6-36ef-4283-912b-779e11ade0a7 service nova] [instance: 9b0053fe-420e-4b92-86a8-ed44f6a6ec49] Updating instance_info_cache with network_info: [{"id": "219230f0-0b14-4be9-81bb-67a765d8ced9", "address": "fa:16:3e:f6:9e:ce", "network": {"id": "cf378e9c-eaae-4168-bd60-9818fcfaf405", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1243326371-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f7c20ab3c822423fa837a03f066e4239", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8cb478a6-872c-4a90-a8db-526b374e82ce", "external-id": "nsx-vlan-transportzone-835", "segmentation_id": 835, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap219230f0-0b", "ovs_interfaceid": "219230f0-0b14-4be9-81bb-67a765d8ced9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 935.559530] env[61957]: DEBUG oslo_vmware.api [None req-d28d4db3-4ba8-4b6b-b020-dcb0a978b0ab tempest-InstanceActionsNegativeTestJSON-259524954 tempest-InstanceActionsNegativeTestJSON-259524954-project-member] Task: {'id': task-1277758, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.111561} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.560249] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-d28d4db3-4ba8-4b6b-b020-dcb0a978b0ab tempest-InstanceActionsNegativeTestJSON-259524954 tempest-InstanceActionsNegativeTestJSON-259524954-project-member] [instance: 7e2db5f6-1419-415e-b957-66cf2e3ec19d] Extended root virtual disk {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 935.561206] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d72c5be-02b0-4f88-95ad-b866861690a4 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.586746] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-d28d4db3-4ba8-4b6b-b020-dcb0a978b0ab tempest-InstanceActionsNegativeTestJSON-259524954 tempest-InstanceActionsNegativeTestJSON-259524954-project-member] [instance: 7e2db5f6-1419-415e-b957-66cf2e3ec19d] Reconfiguring VM instance instance-00000050 to attach disk [datastore1] 7e2db5f6-1419-415e-b957-66cf2e3ec19d/7e2db5f6-1419-415e-b957-66cf2e3ec19d.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 935.587613] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-561a11ab-db26-4ed8-9012-ba21a57b9df3 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.612493] env[61957]: DEBUG oslo_vmware.api [None req-d28d4db3-4ba8-4b6b-b020-dcb0a978b0ab tempest-InstanceActionsNegativeTestJSON-259524954 tempest-InstanceActionsNegativeTestJSON-259524954-project-member] Waiting for the task: (returnval){ [ 935.612493] env[61957]: value = "task-1277759" [ 935.612493] env[61957]: _type = "Task" [ 935.612493] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.623176] env[61957]: DEBUG oslo_vmware.api [None req-d28d4db3-4ba8-4b6b-b020-dcb0a978b0ab tempest-InstanceActionsNegativeTestJSON-259524954 tempest-InstanceActionsNegativeTestJSON-259524954-project-member] Task: {'id': task-1277759, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.782290] env[61957]: DEBUG oslo_concurrency.lockutils [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Acquiring lock "refresh_cache-50e09db7-9f8f-452b-8232-bd7473e9e63a" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 935.782426] env[61957]: DEBUG oslo_concurrency.lockutils [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Acquired lock "refresh_cache-50e09db7-9f8f-452b-8232-bd7473e9e63a" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 935.782583] env[61957]: DEBUG nova.network.neutron [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: 50e09db7-9f8f-452b-8232-bd7473e9e63a] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 935.802275] env[61957]: DEBUG oslo_vmware.api [None req-1036c4ee-30e7-4291-a697-7231cf748564 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52fa3e44-7dc5-5db6-7a81-900ec2575b46, 'name': SearchDatastore_Task, 'duration_secs': 0.013143} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.802615] env[61957]: DEBUG oslo_concurrency.lockutils [None req-1036c4ee-30e7-4291-a697-7231cf748564 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Releasing lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 935.802846] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-1036c4ee-30e7-4291-a697-7231cf748564 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] [instance: 9b0053fe-420e-4b92-86a8-ed44f6a6ec49] Processing image 11c76a2c-f705-470a-ba9d-4657858bab38 {{(pid=61957) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 935.803124] env[61957]: DEBUG oslo_concurrency.lockutils [None req-1036c4ee-30e7-4291-a697-7231cf748564 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 935.803289] env[61957]: DEBUG oslo_concurrency.lockutils [None req-1036c4ee-30e7-4291-a697-7231cf748564 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Acquired lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 935.803495] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-1036c4ee-30e7-4291-a697-7231cf748564 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 935.803754] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7931c9ba-980b-4424-9dac-c01af6891eed {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.811681] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-1036c4ee-30e7-4291-a697-7231cf748564 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 935.811871] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-1036c4ee-30e7-4291-a697-7231cf748564 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61957) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 935.812577] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5ee2f27b-a199-4040-b5ab-ce2b68513e48 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.817692] env[61957]: DEBUG oslo_vmware.api [None req-1036c4ee-30e7-4291-a697-7231cf748564 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Waiting for the task: (returnval){ [ 935.817692] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52adfec1-e76b-1d19-c0e0-c49afd9f161e" [ 935.817692] env[61957]: _type = "Task" [ 935.817692] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.822755] env[61957]: DEBUG nova.compute.utils [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Using /dev/sd instead of None {{(pid=61957) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 935.827070] env[61957]: DEBUG nova.compute.manager [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: 2e32aae5-edd7-402f-98ad-75a93d26f7a1] Allocating IP information in the background. {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 935.827070] env[61957]: DEBUG nova.network.neutron [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: 2e32aae5-edd7-402f-98ad-75a93d26f7a1] allocate_for_instance() {{(pid=61957) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 935.835316] env[61957]: DEBUG oslo_vmware.api [None req-1036c4ee-30e7-4291-a697-7231cf748564 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52adfec1-e76b-1d19-c0e0-c49afd9f161e, 'name': SearchDatastore_Task, 'duration_secs': 0.008494} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.836919] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3a9d8cf2-6078-479d-a8e7-82b25c79d04e {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.843577] env[61957]: DEBUG oslo_vmware.api [None req-1036c4ee-30e7-4291-a697-7231cf748564 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Waiting for the task: (returnval){ [ 935.843577] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]527e71d3-9834-5810-3431-ba0f678dbde3" [ 935.843577] env[61957]: _type = "Task" [ 935.843577] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.851618] env[61957]: DEBUG oslo_vmware.api [None req-1036c4ee-30e7-4291-a697-7231cf748564 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]527e71d3-9834-5810-3431-ba0f678dbde3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.871344] env[61957]: DEBUG nova.policy [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9b453c031e7a4294b14dd48e62e1fd30', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '55dd45b4291b456f83a446dfa87f550a', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61957) authorize /opt/stack/nova/nova/policy.py:203}} [ 935.910704] env[61957]: INFO nova.compute.manager [-] [instance: 39b8b1df-efdd-4c22-9f31-85c85be4f0eb] Took 1.53 seconds to deallocate network for instance. [ 935.954317] env[61957]: DEBUG nova.compute.manager [req-11564a69-620e-4329-bf9a-ec3dc562ef1b req-7f3906a3-bb96-49b5-b7b1-d9bdc55881cf service nova] [instance: 39b8b1df-efdd-4c22-9f31-85c85be4f0eb] Received event network-vif-deleted-7e47ddcd-8c57-46be-8e68-62cec87e391b {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 935.954589] env[61957]: DEBUG nova.compute.manager [req-11564a69-620e-4329-bf9a-ec3dc562ef1b req-7f3906a3-bb96-49b5-b7b1-d9bdc55881cf service nova] [instance: 50e09db7-9f8f-452b-8232-bd7473e9e63a] Received event network-vif-plugged-224ed13c-938e-4dcf-b555-22fd97cd0210 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 935.954822] env[61957]: DEBUG oslo_concurrency.lockutils [req-11564a69-620e-4329-bf9a-ec3dc562ef1b req-7f3906a3-bb96-49b5-b7b1-d9bdc55881cf service nova] Acquiring lock "50e09db7-9f8f-452b-8232-bd7473e9e63a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 935.955605] env[61957]: DEBUG oslo_concurrency.lockutils [req-11564a69-620e-4329-bf9a-ec3dc562ef1b req-7f3906a3-bb96-49b5-b7b1-d9bdc55881cf service nova] Lock "50e09db7-9f8f-452b-8232-bd7473e9e63a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 935.955811] env[61957]: DEBUG oslo_concurrency.lockutils [req-11564a69-620e-4329-bf9a-ec3dc562ef1b req-7f3906a3-bb96-49b5-b7b1-d9bdc55881cf service nova] Lock "50e09db7-9f8f-452b-8232-bd7473e9e63a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 935.955994] env[61957]: DEBUG nova.compute.manager [req-11564a69-620e-4329-bf9a-ec3dc562ef1b req-7f3906a3-bb96-49b5-b7b1-d9bdc55881cf service nova] [instance: 50e09db7-9f8f-452b-8232-bd7473e9e63a] No waiting events found dispatching network-vif-plugged-224ed13c-938e-4dcf-b555-22fd97cd0210 {{(pid=61957) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 935.956207] env[61957]: WARNING nova.compute.manager [req-11564a69-620e-4329-bf9a-ec3dc562ef1b req-7f3906a3-bb96-49b5-b7b1-d9bdc55881cf service nova] [instance: 50e09db7-9f8f-452b-8232-bd7473e9e63a] Received unexpected event network-vif-plugged-224ed13c-938e-4dcf-b555-22fd97cd0210 for instance with vm_state building and task_state spawning. [ 935.956393] env[61957]: DEBUG nova.compute.manager [req-11564a69-620e-4329-bf9a-ec3dc562ef1b req-7f3906a3-bb96-49b5-b7b1-d9bdc55881cf service nova] [instance: 50e09db7-9f8f-452b-8232-bd7473e9e63a] Received event network-changed-224ed13c-938e-4dcf-b555-22fd97cd0210 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 935.956554] env[61957]: DEBUG nova.compute.manager [req-11564a69-620e-4329-bf9a-ec3dc562ef1b req-7f3906a3-bb96-49b5-b7b1-d9bdc55881cf service nova] [instance: 50e09db7-9f8f-452b-8232-bd7473e9e63a] Refreshing instance network info cache due to event network-changed-224ed13c-938e-4dcf-b555-22fd97cd0210. {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 935.956725] env[61957]: DEBUG oslo_concurrency.lockutils [req-11564a69-620e-4329-bf9a-ec3dc562ef1b req-7f3906a3-bb96-49b5-b7b1-d9bdc55881cf service nova] Acquiring lock "refresh_cache-50e09db7-9f8f-452b-8232-bd7473e9e63a" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 936.005171] env[61957]: DEBUG oslo_concurrency.lockutils [req-58febe9d-b1b7-4127-8032-888f8c10d80c req-6d5ff5b6-36ef-4283-912b-779e11ade0a7 service nova] Releasing lock "refresh_cache-9b0053fe-420e-4b92-86a8-ed44f6a6ec49" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 936.122446] env[61957]: DEBUG oslo_vmware.api [None req-d28d4db3-4ba8-4b6b-b020-dcb0a978b0ab tempest-InstanceActionsNegativeTestJSON-259524954 tempest-InstanceActionsNegativeTestJSON-259524954-project-member] Task: {'id': task-1277759, 'name': ReconfigVM_Task, 'duration_secs': 0.360835} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.122820] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-d28d4db3-4ba8-4b6b-b020-dcb0a978b0ab tempest-InstanceActionsNegativeTestJSON-259524954 tempest-InstanceActionsNegativeTestJSON-259524954-project-member] [instance: 7e2db5f6-1419-415e-b957-66cf2e3ec19d] Reconfigured VM instance instance-00000050 to attach disk [datastore1] 7e2db5f6-1419-415e-b957-66cf2e3ec19d/7e2db5f6-1419-415e-b957-66cf2e3ec19d.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 936.124304] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-cf395659-257d-4bfb-9ad3-952e46827321 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.130933] env[61957]: DEBUG oslo_vmware.api [None req-d28d4db3-4ba8-4b6b-b020-dcb0a978b0ab tempest-InstanceActionsNegativeTestJSON-259524954 tempest-InstanceActionsNegativeTestJSON-259524954-project-member] Waiting for the task: (returnval){ [ 936.130933] env[61957]: value = "task-1277760" [ 936.130933] env[61957]: _type = "Task" [ 936.130933] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.135385] env[61957]: DEBUG nova.network.neutron [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: 2e32aae5-edd7-402f-98ad-75a93d26f7a1] Successfully created port: a142b42a-efff-4834-a236-1d0ceb416178 {{(pid=61957) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 936.141036] env[61957]: DEBUG oslo_vmware.api [None req-d28d4db3-4ba8-4b6b-b020-dcb0a978b0ab tempest-InstanceActionsNegativeTestJSON-259524954 tempest-InstanceActionsNegativeTestJSON-259524954-project-member] Task: {'id': task-1277760, 'name': Rename_Task} progress is 5%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.230829] env[61957]: DEBUG nova.network.neutron [-] [instance: b598b8bb-919e-4404-b264-7b76161b0f79] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 936.327666] env[61957]: DEBUG nova.compute.manager [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: 2e32aae5-edd7-402f-98ad-75a93d26f7a1] Start building block device mappings for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 936.332144] env[61957]: DEBUG oslo_concurrency.lockutils [None req-35c8e096-fa6b-43c8-874d-91a51735ac8e tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.016s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 936.333240] env[61957]: DEBUG oslo_concurrency.lockutils [None req-9065790e-dc28-4275-a1c6-42f3b501d4d2 tempest-ServersV294TestFqdnHostnames-1426486100 tempest-ServersV294TestFqdnHostnames-1426486100-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 21.171s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 936.333471] env[61957]: DEBUG oslo_concurrency.lockutils [None req-9065790e-dc28-4275-a1c6-42f3b501d4d2 tempest-ServersV294TestFqdnHostnames-1426486100 tempest-ServersV294TestFqdnHostnames-1426486100-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 936.337905] env[61957]: DEBUG oslo_concurrency.lockutils [None req-94184bcc-7df6-4480-913f-5d000fd6dd4e tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 18.297s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 936.337905] env[61957]: DEBUG oslo_concurrency.lockutils [None req-94184bcc-7df6-4480-913f-5d000fd6dd4e tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 936.342022] env[61957]: DEBUG oslo_concurrency.lockutils [None req-63a396d5-1afb-44c8-a210-9949507fdc49 tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 17.461s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 936.342022] env[61957]: DEBUG oslo_concurrency.lockutils [None req-63a396d5-1afb-44c8-a210-9949507fdc49 tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 936.342022] env[61957]: DEBUG oslo_concurrency.lockutils [None req-23521104-b947-45c5-a62a-77ad796d9410 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 17.461s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 936.347945] env[61957]: DEBUG nova.network.neutron [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: 50e09db7-9f8f-452b-8232-bd7473e9e63a] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 936.357452] env[61957]: DEBUG oslo_vmware.api [None req-1036c4ee-30e7-4291-a697-7231cf748564 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]527e71d3-9834-5810-3431-ba0f678dbde3, 'name': SearchDatastore_Task, 'duration_secs': 0.009617} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.357776] env[61957]: DEBUG oslo_concurrency.lockutils [None req-1036c4ee-30e7-4291-a697-7231cf748564 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Releasing lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 936.358072] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-1036c4ee-30e7-4291-a697-7231cf748564 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore1] 9b0053fe-420e-4b92-86a8-ed44f6a6ec49/9b0053fe-420e-4b92-86a8-ed44f6a6ec49.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 936.358369] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0200b0d6-2e48-4949-86e0-a1ee603c0676 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.365717] env[61957]: DEBUG oslo_vmware.api [None req-1036c4ee-30e7-4291-a697-7231cf748564 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Waiting for the task: (returnval){ [ 936.365717] env[61957]: value = "task-1277761" [ 936.365717] env[61957]: _type = "Task" [ 936.365717] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.382333] env[61957]: DEBUG oslo_vmware.api [None req-1036c4ee-30e7-4291-a697-7231cf748564 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Task: {'id': task-1277761, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.387322] env[61957]: INFO nova.scheduler.client.report [None req-94184bcc-7df6-4480-913f-5d000fd6dd4e tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Deleted allocations for instance d5d10a31-0e78-4ed7-b944-9208138a4861 [ 936.392022] env[61957]: INFO nova.scheduler.client.report [None req-9065790e-dc28-4275-a1c6-42f3b501d4d2 tempest-ServersV294TestFqdnHostnames-1426486100 tempest-ServersV294TestFqdnHostnames-1426486100-project-member] Deleted allocations for instance 2e4a5344-600f-4b61-826e-c15f96b50af2 [ 936.415143] env[61957]: INFO nova.scheduler.client.report [None req-63a396d5-1afb-44c8-a210-9949507fdc49 tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Deleted allocations for instance a8ca1ad6-d636-4fa2-b0ac-53b020e392ce [ 936.477561] env[61957]: INFO nova.compute.manager [None req-5d7e1ad8-1e31-4b98-a89f-2f6c8c624f1c tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 39b8b1df-efdd-4c22-9f31-85c85be4f0eb] Took 0.57 seconds to detach 1 volumes for instance. [ 936.558129] env[61957]: DEBUG nova.network.neutron [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: 50e09db7-9f8f-452b-8232-bd7473e9e63a] Updating instance_info_cache with network_info: [{"id": "224ed13c-938e-4dcf-b555-22fd97cd0210", "address": "fa:16:3e:96:4d:07", "network": {"id": "2b5a7a9c-7143-4754-8d84-486c2ab13b21", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1508819166-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "55dd45b4291b456f83a446dfa87f550a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "572b7281-aad3-45fa-9cb2-fc1c70569948", "external-id": "nsx-vlan-transportzone-722", "segmentation_id": 722, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap224ed13c-93", "ovs_interfaceid": "224ed13c-938e-4dcf-b555-22fd97cd0210", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 936.642368] env[61957]: DEBUG oslo_vmware.api [None req-d28d4db3-4ba8-4b6b-b020-dcb0a978b0ab tempest-InstanceActionsNegativeTestJSON-259524954 tempest-InstanceActionsNegativeTestJSON-259524954-project-member] Task: {'id': task-1277760, 'name': Rename_Task, 'duration_secs': 0.179821} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.642682] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-d28d4db3-4ba8-4b6b-b020-dcb0a978b0ab tempest-InstanceActionsNegativeTestJSON-259524954 tempest-InstanceActionsNegativeTestJSON-259524954-project-member] [instance: 7e2db5f6-1419-415e-b957-66cf2e3ec19d] Powering on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 936.642950] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-174697bb-efcd-4637-962b-fef716172c16 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.652162] env[61957]: DEBUG oslo_vmware.api [None req-d28d4db3-4ba8-4b6b-b020-dcb0a978b0ab tempest-InstanceActionsNegativeTestJSON-259524954 tempest-InstanceActionsNegativeTestJSON-259524954-project-member] Waiting for the task: (returnval){ [ 936.652162] env[61957]: value = "task-1277762" [ 936.652162] env[61957]: _type = "Task" [ 936.652162] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.661152] env[61957]: DEBUG oslo_vmware.api [None req-d28d4db3-4ba8-4b6b-b020-dcb0a978b0ab tempest-InstanceActionsNegativeTestJSON-259524954 tempest-InstanceActionsNegativeTestJSON-259524954-project-member] Task: {'id': task-1277762, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.733612] env[61957]: INFO nova.compute.manager [-] [instance: b598b8bb-919e-4404-b264-7b76161b0f79] Took 1.31 seconds to deallocate network for instance. [ 936.839685] env[61957]: DEBUG oslo_service.periodic_task [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61957) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 936.839966] env[61957]: DEBUG oslo_service.periodic_task [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61957) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 936.847422] env[61957]: INFO nova.compute.claims [None req-23521104-b947-45c5-a62a-77ad796d9410 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: dbeeb200-70b9-4cb4-b5a4-182389d21918] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 936.877190] env[61957]: DEBUG oslo_vmware.api [None req-1036c4ee-30e7-4291-a697-7231cf748564 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Task: {'id': task-1277761, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.500447} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.877428] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-1036c4ee-30e7-4291-a697-7231cf748564 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore1] 9b0053fe-420e-4b92-86a8-ed44f6a6ec49/9b0053fe-420e-4b92-86a8-ed44f6a6ec49.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 936.877649] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-1036c4ee-30e7-4291-a697-7231cf748564 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] [instance: 9b0053fe-420e-4b92-86a8-ed44f6a6ec49] Extending root virtual disk to 1048576 {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 936.877932] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-92e314c9-df92-4c39-9b42-605bb95d7e41 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.884414] env[61957]: DEBUG oslo_vmware.api [None req-1036c4ee-30e7-4291-a697-7231cf748564 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Waiting for the task: (returnval){ [ 936.884414] env[61957]: value = "task-1277763" [ 936.884414] env[61957]: _type = "Task" [ 936.884414] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.897850] env[61957]: DEBUG oslo_vmware.api [None req-1036c4ee-30e7-4291-a697-7231cf748564 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Task: {'id': task-1277763, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.898408] env[61957]: DEBUG oslo_concurrency.lockutils [None req-9065790e-dc28-4275-a1c6-42f3b501d4d2 tempest-ServersV294TestFqdnHostnames-1426486100 tempest-ServersV294TestFqdnHostnames-1426486100-project-member] Lock "2e4a5344-600f-4b61-826e-c15f96b50af2" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 24.874s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 936.899968] env[61957]: DEBUG oslo_concurrency.lockutils [None req-94184bcc-7df6-4480-913f-5d000fd6dd4e tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Lock "d5d10a31-0e78-4ed7-b944-9208138a4861" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 21.919s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 936.931277] env[61957]: DEBUG oslo_concurrency.lockutils [None req-63a396d5-1afb-44c8-a210-9949507fdc49 tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Lock "a8ca1ad6-d636-4fa2-b0ac-53b020e392ce" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 22.042s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 936.985760] env[61957]: DEBUG oslo_concurrency.lockutils [None req-5d7e1ad8-1e31-4b98-a89f-2f6c8c624f1c tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 937.061661] env[61957]: DEBUG oslo_concurrency.lockutils [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Releasing lock "refresh_cache-50e09db7-9f8f-452b-8232-bd7473e9e63a" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 937.062197] env[61957]: DEBUG nova.compute.manager [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: 50e09db7-9f8f-452b-8232-bd7473e9e63a] Instance network_info: |[{"id": "224ed13c-938e-4dcf-b555-22fd97cd0210", "address": "fa:16:3e:96:4d:07", "network": {"id": "2b5a7a9c-7143-4754-8d84-486c2ab13b21", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1508819166-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "55dd45b4291b456f83a446dfa87f550a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "572b7281-aad3-45fa-9cb2-fc1c70569948", "external-id": "nsx-vlan-transportzone-722", "segmentation_id": 722, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap224ed13c-93", "ovs_interfaceid": "224ed13c-938e-4dcf-b555-22fd97cd0210", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 937.062591] env[61957]: DEBUG oslo_concurrency.lockutils [req-11564a69-620e-4329-bf9a-ec3dc562ef1b req-7f3906a3-bb96-49b5-b7b1-d9bdc55881cf service nova] Acquired lock "refresh_cache-50e09db7-9f8f-452b-8232-bd7473e9e63a" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 937.062889] env[61957]: DEBUG nova.network.neutron [req-11564a69-620e-4329-bf9a-ec3dc562ef1b req-7f3906a3-bb96-49b5-b7b1-d9bdc55881cf service nova] [instance: 50e09db7-9f8f-452b-8232-bd7473e9e63a] Refreshing network info cache for port 224ed13c-938e-4dcf-b555-22fd97cd0210 {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 937.064520] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: 50e09db7-9f8f-452b-8232-bd7473e9e63a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:96:4d:07', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '572b7281-aad3-45fa-9cb2-fc1c70569948', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '224ed13c-938e-4dcf-b555-22fd97cd0210', 'vif_model': 'vmxnet3'}] {{(pid=61957) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 937.073693] env[61957]: DEBUG oslo.service.loopingcall [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 937.074778] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 50e09db7-9f8f-452b-8232-bd7473e9e63a] Creating VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 937.075078] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-54203316-f058-4b3d-b708-cc922cfa2a8a {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.095842] env[61957]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 937.095842] env[61957]: value = "task-1277764" [ 937.095842] env[61957]: _type = "Task" [ 937.095842] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.104636] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277764, 'name': CreateVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.163715] env[61957]: DEBUG oslo_vmware.api [None req-d28d4db3-4ba8-4b6b-b020-dcb0a978b0ab tempest-InstanceActionsNegativeTestJSON-259524954 tempest-InstanceActionsNegativeTestJSON-259524954-project-member] Task: {'id': task-1277762, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.240642] env[61957]: DEBUG oslo_concurrency.lockutils [None req-73ed12d4-fc9a-46df-a24a-8dbe4ec86160 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 937.351027] env[61957]: DEBUG oslo_service.periodic_task [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61957) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 937.351027] env[61957]: DEBUG nova.compute.manager [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Starting heal instance info cache {{(pid=61957) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 937.357412] env[61957]: INFO nova.compute.resource_tracker [None req-23521104-b947-45c5-a62a-77ad796d9410 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: dbeeb200-70b9-4cb4-b5a4-182389d21918] Updating resource usage from migration f4269166-72d4-48ca-b806-e1e29520eb19 [ 937.385256] env[61957]: DEBUG nova.compute.manager [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: 2e32aae5-edd7-402f-98ad-75a93d26f7a1] Start spawning the instance on the hypervisor. {{(pid=61957) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 937.399024] env[61957]: DEBUG oslo_vmware.api [None req-1036c4ee-30e7-4291-a697-7231cf748564 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Task: {'id': task-1277763, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.120007} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.400012] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-1036c4ee-30e7-4291-a697-7231cf748564 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] [instance: 9b0053fe-420e-4b92-86a8-ed44f6a6ec49] Extended root virtual disk {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 937.400941] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0f4606c-fb9a-4ef3-86bd-b8c04a5d35fa {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.425547] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-1036c4ee-30e7-4291-a697-7231cf748564 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] [instance: 9b0053fe-420e-4b92-86a8-ed44f6a6ec49] Reconfiguring VM instance instance-00000051 to attach disk [datastore1] 9b0053fe-420e-4b92-86a8-ed44f6a6ec49/9b0053fe-420e-4b92-86a8-ed44f6a6ec49.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 937.432593] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8eef8fbc-faec-4143-af2f-a964eefe16b1 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.458981] env[61957]: DEBUG oslo_vmware.api [None req-1036c4ee-30e7-4291-a697-7231cf748564 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Waiting for the task: (returnval){ [ 937.458981] env[61957]: value = "task-1277765" [ 937.458981] env[61957]: _type = "Task" [ 937.458981] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.461460] env[61957]: DEBUG nova.virt.hardware [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T17:22:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T17:21:50Z,direct_url=,disk_format='vmdk',id=11c76a2c-f705-470a-ba9d-4657858bab38,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='81d3096d0f094373913fc3dc8f5c3c6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T17:21:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 937.461720] env[61957]: DEBUG nova.virt.hardware [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Flavor limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 937.461899] env[61957]: DEBUG nova.virt.hardware [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Image limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 937.463317] env[61957]: DEBUG nova.virt.hardware [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Flavor pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 937.463317] env[61957]: DEBUG nova.virt.hardware [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Image pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 937.463317] env[61957]: DEBUG nova.virt.hardware [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 937.463317] env[61957]: DEBUG nova.virt.hardware [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 937.463317] env[61957]: DEBUG nova.virt.hardware [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 937.463547] env[61957]: DEBUG nova.virt.hardware [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Got 1 possible topologies {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 937.463608] env[61957]: DEBUG nova.virt.hardware [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 937.463836] env[61957]: DEBUG nova.virt.hardware [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 937.465984] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d59c1aa0-a11a-4737-b413-b8e404e6016f {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.483044] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9411cbe0-8d0a-4c69-9b8c-72f73d4b9e8f {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.487480] env[61957]: DEBUG oslo_vmware.api [None req-1036c4ee-30e7-4291-a697-7231cf748564 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Task: {'id': task-1277765, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.610331] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277764, 'name': CreateVM_Task, 'duration_secs': 0.398961} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.610331] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 50e09db7-9f8f-452b-8232-bd7473e9e63a] Created VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 937.610883] env[61957]: DEBUG oslo_concurrency.lockutils [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 937.611243] env[61957]: DEBUG oslo_concurrency.lockutils [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Acquired lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 937.611691] env[61957]: DEBUG oslo_concurrency.lockutils [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 937.612096] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0670487a-ccca-4e6c-a0ea-3a13ce016718 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.617800] env[61957]: DEBUG oslo_vmware.api [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Waiting for the task: (returnval){ [ 937.617800] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]522671cb-652e-d789-dcb2-a46087974b41" [ 937.617800] env[61957]: _type = "Task" [ 937.617800] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.627303] env[61957]: DEBUG oslo_vmware.api [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]522671cb-652e-d789-dcb2-a46087974b41, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.663491] env[61957]: DEBUG oslo_vmware.api [None req-d28d4db3-4ba8-4b6b-b020-dcb0a978b0ab tempest-InstanceActionsNegativeTestJSON-259524954 tempest-InstanceActionsNegativeTestJSON-259524954-project-member] Task: {'id': task-1277762, 'name': PowerOnVM_Task, 'duration_secs': 0.70674} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.667304] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-d28d4db3-4ba8-4b6b-b020-dcb0a978b0ab tempest-InstanceActionsNegativeTestJSON-259524954 tempest-InstanceActionsNegativeTestJSON-259524954-project-member] [instance: 7e2db5f6-1419-415e-b957-66cf2e3ec19d] Powered on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 937.667551] env[61957]: INFO nova.compute.manager [None req-d28d4db3-4ba8-4b6b-b020-dcb0a978b0ab tempest-InstanceActionsNegativeTestJSON-259524954 tempest-InstanceActionsNegativeTestJSON-259524954-project-member] [instance: 7e2db5f6-1419-415e-b957-66cf2e3ec19d] Took 7.82 seconds to spawn the instance on the hypervisor. [ 937.667741] env[61957]: DEBUG nova.compute.manager [None req-d28d4db3-4ba8-4b6b-b020-dcb0a978b0ab tempest-InstanceActionsNegativeTestJSON-259524954 tempest-InstanceActionsNegativeTestJSON-259524954-project-member] [instance: 7e2db5f6-1419-415e-b957-66cf2e3ec19d] Checking state {{(pid=61957) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 937.668990] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70b5fc2c-0dd6-497c-b0cd-0d1d4814e9f5 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.753886] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-201b0b27-d68c-4d35-bf12-230a7ae311ca {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.765081] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efa9a610-cc38-41cb-b8fa-3c18d7ae4435 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.803924] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a289a31c-7103-4e33-892b-7c60fe436c6a {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.812099] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b87e7ce-4719-4110-8dfd-7e82cea7b9b9 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.828986] env[61957]: DEBUG nova.compute.provider_tree [None req-23521104-b947-45c5-a62a-77ad796d9410 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 937.953913] env[61957]: DEBUG oslo_concurrency.lockutils [None req-78652638-0a34-4302-a770-2b339b434047 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Acquiring lock "28ab7d23-487f-4ae1-8fe4-58db55b59918" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 937.954256] env[61957]: DEBUG oslo_concurrency.lockutils [None req-78652638-0a34-4302-a770-2b339b434047 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Lock "28ab7d23-487f-4ae1-8fe4-58db55b59918" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 937.954520] env[61957]: DEBUG oslo_concurrency.lockutils [None req-78652638-0a34-4302-a770-2b339b434047 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Acquiring lock "28ab7d23-487f-4ae1-8fe4-58db55b59918-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 937.955059] env[61957]: DEBUG oslo_concurrency.lockutils [None req-78652638-0a34-4302-a770-2b339b434047 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Lock "28ab7d23-487f-4ae1-8fe4-58db55b59918-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 937.955279] env[61957]: DEBUG oslo_concurrency.lockutils [None req-78652638-0a34-4302-a770-2b339b434047 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Lock "28ab7d23-487f-4ae1-8fe4-58db55b59918-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 937.957486] env[61957]: DEBUG oslo_concurrency.lockutils [None req-483cfd5f-2659-4d70-aee0-52ce6ba7fa9a tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Acquiring lock "0d0f2d34-de35-4e80-8d9f-12693add0786" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 937.957691] env[61957]: DEBUG oslo_concurrency.lockutils [None req-483cfd5f-2659-4d70-aee0-52ce6ba7fa9a tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Lock "0d0f2d34-de35-4e80-8d9f-12693add0786" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 937.957879] env[61957]: DEBUG oslo_concurrency.lockutils [None req-483cfd5f-2659-4d70-aee0-52ce6ba7fa9a tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Acquiring lock "0d0f2d34-de35-4e80-8d9f-12693add0786-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 937.958072] env[61957]: DEBUG oslo_concurrency.lockutils [None req-483cfd5f-2659-4d70-aee0-52ce6ba7fa9a tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Lock "0d0f2d34-de35-4e80-8d9f-12693add0786-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 937.958246] env[61957]: DEBUG oslo_concurrency.lockutils [None req-483cfd5f-2659-4d70-aee0-52ce6ba7fa9a tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Lock "0d0f2d34-de35-4e80-8d9f-12693add0786-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 937.960254] env[61957]: INFO nova.compute.manager [None req-78652638-0a34-4302-a770-2b339b434047 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] [instance: 28ab7d23-487f-4ae1-8fe4-58db55b59918] Terminating instance [ 937.962099] env[61957]: INFO nova.compute.manager [None req-483cfd5f-2659-4d70-aee0-52ce6ba7fa9a tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] [instance: 0d0f2d34-de35-4e80-8d9f-12693add0786] Terminating instance [ 937.968929] env[61957]: DEBUG nova.compute.manager [None req-78652638-0a34-4302-a770-2b339b434047 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] [instance: 28ab7d23-487f-4ae1-8fe4-58db55b59918] Start destroying the instance on the hypervisor. {{(pid=61957) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 937.969167] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-78652638-0a34-4302-a770-2b339b434047 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] [instance: 28ab7d23-487f-4ae1-8fe4-58db55b59918] Destroying instance {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 937.969848] env[61957]: DEBUG oslo_concurrency.lockutils [None req-483cfd5f-2659-4d70-aee0-52ce6ba7fa9a tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Acquiring lock "refresh_cache-0d0f2d34-de35-4e80-8d9f-12693add0786" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 937.970043] env[61957]: DEBUG oslo_concurrency.lockutils [None req-483cfd5f-2659-4d70-aee0-52ce6ba7fa9a tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Acquired lock "refresh_cache-0d0f2d34-de35-4e80-8d9f-12693add0786" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 937.970563] env[61957]: DEBUG nova.network.neutron [None req-483cfd5f-2659-4d70-aee0-52ce6ba7fa9a tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] [instance: 0d0f2d34-de35-4e80-8d9f-12693add0786] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 937.972954] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b03f5e1-508f-41a0-9aed-ee9325f3982d {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.988762] env[61957]: DEBUG oslo_vmware.api [None req-1036c4ee-30e7-4291-a697-7231cf748564 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Task: {'id': task-1277765, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.989591] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-78652638-0a34-4302-a770-2b339b434047 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] [instance: 28ab7d23-487f-4ae1-8fe4-58db55b59918] Powering off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 937.990021] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-363c9e2f-00c0-4f40-883f-23a78fb75b19 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.997292] env[61957]: DEBUG oslo_vmware.api [None req-78652638-0a34-4302-a770-2b339b434047 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Waiting for the task: (returnval){ [ 937.997292] env[61957]: value = "task-1277766" [ 937.997292] env[61957]: _type = "Task" [ 937.997292] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.007187] env[61957]: DEBUG oslo_vmware.api [None req-78652638-0a34-4302-a770-2b339b434047 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Task: {'id': task-1277766, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.044196] env[61957]: DEBUG nova.compute.manager [req-057a6efe-ee42-4bde-bdd2-e476bd7a4e83 req-3c5ab1eb-ec90-4349-8476-a0f4713ac85a service nova] [instance: 2e32aae5-edd7-402f-98ad-75a93d26f7a1] Received event network-vif-plugged-a142b42a-efff-4834-a236-1d0ceb416178 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 938.044196] env[61957]: DEBUG oslo_concurrency.lockutils [req-057a6efe-ee42-4bde-bdd2-e476bd7a4e83 req-3c5ab1eb-ec90-4349-8476-a0f4713ac85a service nova] Acquiring lock "2e32aae5-edd7-402f-98ad-75a93d26f7a1-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 938.044196] env[61957]: DEBUG oslo_concurrency.lockutils [req-057a6efe-ee42-4bde-bdd2-e476bd7a4e83 req-3c5ab1eb-ec90-4349-8476-a0f4713ac85a service nova] Lock "2e32aae5-edd7-402f-98ad-75a93d26f7a1-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 938.044196] env[61957]: DEBUG oslo_concurrency.lockutils [req-057a6efe-ee42-4bde-bdd2-e476bd7a4e83 req-3c5ab1eb-ec90-4349-8476-a0f4713ac85a service nova] Lock "2e32aae5-edd7-402f-98ad-75a93d26f7a1-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 938.044680] env[61957]: DEBUG nova.compute.manager [req-057a6efe-ee42-4bde-bdd2-e476bd7a4e83 req-3c5ab1eb-ec90-4349-8476-a0f4713ac85a service nova] [instance: 2e32aae5-edd7-402f-98ad-75a93d26f7a1] No waiting events found dispatching network-vif-plugged-a142b42a-efff-4834-a236-1d0ceb416178 {{(pid=61957) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 938.045030] env[61957]: WARNING nova.compute.manager [req-057a6efe-ee42-4bde-bdd2-e476bd7a4e83 req-3c5ab1eb-ec90-4349-8476-a0f4713ac85a service nova] [instance: 2e32aae5-edd7-402f-98ad-75a93d26f7a1] Received unexpected event network-vif-plugged-a142b42a-efff-4834-a236-1d0ceb416178 for instance with vm_state building and task_state spawning. [ 938.050091] env[61957]: DEBUG nova.network.neutron [req-11564a69-620e-4329-bf9a-ec3dc562ef1b req-7f3906a3-bb96-49b5-b7b1-d9bdc55881cf service nova] [instance: 50e09db7-9f8f-452b-8232-bd7473e9e63a] Updated VIF entry in instance network info cache for port 224ed13c-938e-4dcf-b555-22fd97cd0210. {{(pid=61957) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 938.050611] env[61957]: DEBUG nova.network.neutron [req-11564a69-620e-4329-bf9a-ec3dc562ef1b req-7f3906a3-bb96-49b5-b7b1-d9bdc55881cf service nova] [instance: 50e09db7-9f8f-452b-8232-bd7473e9e63a] Updating instance_info_cache with network_info: [{"id": "224ed13c-938e-4dcf-b555-22fd97cd0210", "address": "fa:16:3e:96:4d:07", "network": {"id": "2b5a7a9c-7143-4754-8d84-486c2ab13b21", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1508819166-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "55dd45b4291b456f83a446dfa87f550a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "572b7281-aad3-45fa-9cb2-fc1c70569948", "external-id": "nsx-vlan-transportzone-722", "segmentation_id": 722, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap224ed13c-93", "ovs_interfaceid": "224ed13c-938e-4dcf-b555-22fd97cd0210", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 938.136813] env[61957]: DEBUG oslo_vmware.api [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]522671cb-652e-d789-dcb2-a46087974b41, 'name': SearchDatastore_Task, 'duration_secs': 0.009592} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.137661] env[61957]: DEBUG oslo_concurrency.lockutils [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Releasing lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 938.138159] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: 50e09db7-9f8f-452b-8232-bd7473e9e63a] Processing image 11c76a2c-f705-470a-ba9d-4657858bab38 {{(pid=61957) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 938.138762] env[61957]: DEBUG oslo_concurrency.lockutils [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 938.139100] env[61957]: DEBUG oslo_concurrency.lockutils [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Acquired lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 938.139438] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 938.139917] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-65f9dc8c-d612-4683-8956-6ae22d893e34 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.151861] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 938.152497] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61957) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 938.153500] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-774f7266-831f-453d-91a1-98ac8c9b298f {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.169262] env[61957]: DEBUG oslo_vmware.api [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Waiting for the task: (returnval){ [ 938.169262] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52ffb3ec-e4ac-2a03-2726-1c02b8aef745" [ 938.169262] env[61957]: _type = "Task" [ 938.169262] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.187553] env[61957]: DEBUG oslo_vmware.api [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52ffb3ec-e4ac-2a03-2726-1c02b8aef745, 'name': SearchDatastore_Task, 'duration_secs': 0.015111} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.191252] env[61957]: INFO nova.compute.manager [None req-d28d4db3-4ba8-4b6b-b020-dcb0a978b0ab tempest-InstanceActionsNegativeTestJSON-259524954 tempest-InstanceActionsNegativeTestJSON-259524954-project-member] [instance: 7e2db5f6-1419-415e-b957-66cf2e3ec19d] Took 42.28 seconds to build instance. [ 938.192785] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-beae0511-e550-40f7-86e9-d2303c7fd6d5 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.200126] env[61957]: DEBUG oslo_vmware.api [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Waiting for the task: (returnval){ [ 938.200126] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52ed2ebe-2d09-3326-a1d5-35d0d48cad26" [ 938.200126] env[61957]: _type = "Task" [ 938.200126] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.211886] env[61957]: DEBUG oslo_vmware.api [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52ed2ebe-2d09-3326-a1d5-35d0d48cad26, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.332130] env[61957]: DEBUG nova.scheduler.client.report [None req-23521104-b947-45c5-a62a-77ad796d9410 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 938.474050] env[61957]: DEBUG oslo_vmware.api [None req-1036c4ee-30e7-4291-a697-7231cf748564 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Task: {'id': task-1277765, 'name': ReconfigVM_Task, 'duration_secs': 0.787965} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.474367] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-1036c4ee-30e7-4291-a697-7231cf748564 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] [instance: 9b0053fe-420e-4b92-86a8-ed44f6a6ec49] Reconfigured VM instance instance-00000051 to attach disk [datastore1] 9b0053fe-420e-4b92-86a8-ed44f6a6ec49/9b0053fe-420e-4b92-86a8-ed44f6a6ec49.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 938.475035] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-853c527d-f6e7-4395-ac7d-a0b58f21c5bf {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.481893] env[61957]: DEBUG oslo_vmware.api [None req-1036c4ee-30e7-4291-a697-7231cf748564 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Waiting for the task: (returnval){ [ 938.481893] env[61957]: value = "task-1277767" [ 938.481893] env[61957]: _type = "Task" [ 938.481893] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.489891] env[61957]: DEBUG oslo_vmware.api [None req-1036c4ee-30e7-4291-a697-7231cf748564 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Task: {'id': task-1277767, 'name': Rename_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.493723] env[61957]: DEBUG nova.network.neutron [None req-483cfd5f-2659-4d70-aee0-52ce6ba7fa9a tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] [instance: 0d0f2d34-de35-4e80-8d9f-12693add0786] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 938.506870] env[61957]: DEBUG oslo_vmware.api [None req-78652638-0a34-4302-a770-2b339b434047 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Task: {'id': task-1277766, 'name': PowerOffVM_Task, 'duration_secs': 0.20732} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.511744] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-78652638-0a34-4302-a770-2b339b434047 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] [instance: 28ab7d23-487f-4ae1-8fe4-58db55b59918] Powered off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 938.511984] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-78652638-0a34-4302-a770-2b339b434047 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] [instance: 28ab7d23-487f-4ae1-8fe4-58db55b59918] Unregistering the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 938.515391] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7cb12436-9ef3-49a9-a77e-63899adab477 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.554131] env[61957]: DEBUG oslo_concurrency.lockutils [req-11564a69-620e-4329-bf9a-ec3dc562ef1b req-7f3906a3-bb96-49b5-b7b1-d9bdc55881cf service nova] Releasing lock "refresh_cache-50e09db7-9f8f-452b-8232-bd7473e9e63a" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 938.554614] env[61957]: DEBUG nova.compute.manager [req-11564a69-620e-4329-bf9a-ec3dc562ef1b req-7f3906a3-bb96-49b5-b7b1-d9bdc55881cf service nova] [instance: b598b8bb-919e-4404-b264-7b76161b0f79] Received event network-vif-deleted-be3f3b45-e91d-4c38-b746-a10c838decf7 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 938.554867] env[61957]: INFO nova.compute.manager [req-11564a69-620e-4329-bf9a-ec3dc562ef1b req-7f3906a3-bb96-49b5-b7b1-d9bdc55881cf service nova] [instance: b598b8bb-919e-4404-b264-7b76161b0f79] Neutron deleted interface be3f3b45-e91d-4c38-b746-a10c838decf7; detaching it from the instance and deleting it from the info cache [ 938.555058] env[61957]: DEBUG nova.network.neutron [req-11564a69-620e-4329-bf9a-ec3dc562ef1b req-7f3906a3-bb96-49b5-b7b1-d9bdc55881cf service nova] [instance: b598b8bb-919e-4404-b264-7b76161b0f79] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 938.556837] env[61957]: DEBUG nova.network.neutron [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: 2e32aae5-edd7-402f-98ad-75a93d26f7a1] Successfully updated port: a142b42a-efff-4834-a236-1d0ceb416178 {{(pid=61957) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 938.573525] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-78652638-0a34-4302-a770-2b339b434047 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] [instance: 28ab7d23-487f-4ae1-8fe4-58db55b59918] Unregistered the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 938.573776] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-78652638-0a34-4302-a770-2b339b434047 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] [instance: 28ab7d23-487f-4ae1-8fe4-58db55b59918] Deleting contents of the VM from datastore datastore2 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 938.573969] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-78652638-0a34-4302-a770-2b339b434047 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Deleting the datastore file [datastore2] 28ab7d23-487f-4ae1-8fe4-58db55b59918 {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 938.574367] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-57565e21-50c5-478a-9fc5-06643d83c1fc {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.577793] env[61957]: DEBUG nova.network.neutron [None req-483cfd5f-2659-4d70-aee0-52ce6ba7fa9a tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] [instance: 0d0f2d34-de35-4e80-8d9f-12693add0786] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 938.584265] env[61957]: DEBUG oslo_vmware.api [None req-78652638-0a34-4302-a770-2b339b434047 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Waiting for the task: (returnval){ [ 938.584265] env[61957]: value = "task-1277769" [ 938.584265] env[61957]: _type = "Task" [ 938.584265] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.594449] env[61957]: DEBUG oslo_vmware.api [None req-78652638-0a34-4302-a770-2b339b434047 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Task: {'id': task-1277769, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.613900] env[61957]: DEBUG nova.compute.manager [req-9cb7126f-0512-48be-9451-549561edc6e2 req-f31a9876-7654-48e5-b03e-9a59474a165b service nova] [instance: 2e32aae5-edd7-402f-98ad-75a93d26f7a1] Received event network-changed-a142b42a-efff-4834-a236-1d0ceb416178 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 938.614191] env[61957]: DEBUG nova.compute.manager [req-9cb7126f-0512-48be-9451-549561edc6e2 req-f31a9876-7654-48e5-b03e-9a59474a165b service nova] [instance: 2e32aae5-edd7-402f-98ad-75a93d26f7a1] Refreshing instance network info cache due to event network-changed-a142b42a-efff-4834-a236-1d0ceb416178. {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 938.614481] env[61957]: DEBUG oslo_concurrency.lockutils [req-9cb7126f-0512-48be-9451-549561edc6e2 req-f31a9876-7654-48e5-b03e-9a59474a165b service nova] Acquiring lock "refresh_cache-2e32aae5-edd7-402f-98ad-75a93d26f7a1" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 938.614929] env[61957]: DEBUG oslo_concurrency.lockutils [req-9cb7126f-0512-48be-9451-549561edc6e2 req-f31a9876-7654-48e5-b03e-9a59474a165b service nova] Acquired lock "refresh_cache-2e32aae5-edd7-402f-98ad-75a93d26f7a1" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 938.615178] env[61957]: DEBUG nova.network.neutron [req-9cb7126f-0512-48be-9451-549561edc6e2 req-f31a9876-7654-48e5-b03e-9a59474a165b service nova] [instance: 2e32aae5-edd7-402f-98ad-75a93d26f7a1] Refreshing network info cache for port a142b42a-efff-4834-a236-1d0ceb416178 {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 938.697243] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d28d4db3-4ba8-4b6b-b020-dcb0a978b0ab tempest-InstanceActionsNegativeTestJSON-259524954 tempest-InstanceActionsNegativeTestJSON-259524954-project-member] Lock "7e2db5f6-1419-415e-b957-66cf2e3ec19d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 48.753s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 938.718032] env[61957]: DEBUG oslo_vmware.api [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52ed2ebe-2d09-3326-a1d5-35d0d48cad26, 'name': SearchDatastore_Task, 'duration_secs': 0.023242} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.718353] env[61957]: DEBUG oslo_concurrency.lockutils [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Releasing lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 938.718633] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore1] 50e09db7-9f8f-452b-8232-bd7473e9e63a/50e09db7-9f8f-452b-8232-bd7473e9e63a.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 938.718928] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2fdbd118-ca0b-4152-b741-d3f21fb046f5 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.727552] env[61957]: DEBUG oslo_vmware.api [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Waiting for the task: (returnval){ [ 938.727552] env[61957]: value = "task-1277770" [ 938.727552] env[61957]: _type = "Task" [ 938.727552] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.736749] env[61957]: DEBUG oslo_vmware.api [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Task: {'id': task-1277770, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.838084] env[61957]: DEBUG oslo_concurrency.lockutils [None req-23521104-b947-45c5-a62a-77ad796d9410 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.497s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 938.838359] env[61957]: INFO nova.compute.manager [None req-23521104-b947-45c5-a62a-77ad796d9410 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: dbeeb200-70b9-4cb4-b5a4-182389d21918] Migrating [ 938.838694] env[61957]: DEBUG oslo_concurrency.lockutils [None req-23521104-b947-45c5-a62a-77ad796d9410 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Acquiring lock "compute-rpcapi-router" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 938.838823] env[61957]: DEBUG oslo_concurrency.lockutils [None req-23521104-b947-45c5-a62a-77ad796d9410 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Acquired lock "compute-rpcapi-router" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 938.840392] env[61957]: DEBUG oslo_concurrency.lockutils [None req-3e0cd30b-989c-4521-a40d-70a59bfcf72a tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 16.552s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 938.840653] env[61957]: DEBUG nova.objects.instance [None req-3e0cd30b-989c-4521-a40d-70a59bfcf72a tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Lazy-loading 'resources' on Instance uuid c23141ee-0cbb-4d1b-8390-c3073fe354f1 {{(pid=61957) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 938.998295] env[61957]: DEBUG oslo_vmware.api [None req-1036c4ee-30e7-4291-a697-7231cf748564 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Task: {'id': task-1277767, 'name': Rename_Task, 'duration_secs': 0.21172} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.998295] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-1036c4ee-30e7-4291-a697-7231cf748564 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] [instance: 9b0053fe-420e-4b92-86a8-ed44f6a6ec49] Powering on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 938.998295] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-04f51db1-e100-445d-947e-0c14baa9e0aa {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.008321] env[61957]: DEBUG oslo_vmware.api [None req-1036c4ee-30e7-4291-a697-7231cf748564 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Waiting for the task: (returnval){ [ 939.008321] env[61957]: value = "task-1277771" [ 939.008321] env[61957]: _type = "Task" [ 939.008321] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.022061] env[61957]: DEBUG oslo_vmware.api [None req-1036c4ee-30e7-4291-a697-7231cf748564 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Task: {'id': task-1277771, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.025033] env[61957]: DEBUG oslo_concurrency.lockutils [None req-680ab45d-0a61-469d-82c9-fe13d79ebbb0 tempest-InstanceActionsNegativeTestJSON-259524954 tempest-InstanceActionsNegativeTestJSON-259524954-project-member] Acquiring lock "7e2db5f6-1419-415e-b957-66cf2e3ec19d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 939.025531] env[61957]: DEBUG oslo_concurrency.lockutils [None req-680ab45d-0a61-469d-82c9-fe13d79ebbb0 tempest-InstanceActionsNegativeTestJSON-259524954 tempest-InstanceActionsNegativeTestJSON-259524954-project-member] Lock "7e2db5f6-1419-415e-b957-66cf2e3ec19d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 939.025843] env[61957]: DEBUG oslo_concurrency.lockutils [None req-680ab45d-0a61-469d-82c9-fe13d79ebbb0 tempest-InstanceActionsNegativeTestJSON-259524954 tempest-InstanceActionsNegativeTestJSON-259524954-project-member] Acquiring lock "7e2db5f6-1419-415e-b957-66cf2e3ec19d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 939.026139] env[61957]: DEBUG oslo_concurrency.lockutils [None req-680ab45d-0a61-469d-82c9-fe13d79ebbb0 tempest-InstanceActionsNegativeTestJSON-259524954 tempest-InstanceActionsNegativeTestJSON-259524954-project-member] Lock "7e2db5f6-1419-415e-b957-66cf2e3ec19d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 939.026412] env[61957]: DEBUG oslo_concurrency.lockutils [None req-680ab45d-0a61-469d-82c9-fe13d79ebbb0 tempest-InstanceActionsNegativeTestJSON-259524954 tempest-InstanceActionsNegativeTestJSON-259524954-project-member] Lock "7e2db5f6-1419-415e-b957-66cf2e3ec19d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 939.029138] env[61957]: INFO nova.compute.manager [None req-680ab45d-0a61-469d-82c9-fe13d79ebbb0 tempest-InstanceActionsNegativeTestJSON-259524954 tempest-InstanceActionsNegativeTestJSON-259524954-project-member] [instance: 7e2db5f6-1419-415e-b957-66cf2e3ec19d] Terminating instance [ 939.031753] env[61957]: DEBUG nova.compute.manager [None req-680ab45d-0a61-469d-82c9-fe13d79ebbb0 tempest-InstanceActionsNegativeTestJSON-259524954 tempest-InstanceActionsNegativeTestJSON-259524954-project-member] [instance: 7e2db5f6-1419-415e-b957-66cf2e3ec19d] Start destroying the instance on the hypervisor. {{(pid=61957) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 939.032144] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-680ab45d-0a61-469d-82c9-fe13d79ebbb0 tempest-InstanceActionsNegativeTestJSON-259524954 tempest-InstanceActionsNegativeTestJSON-259524954-project-member] [instance: 7e2db5f6-1419-415e-b957-66cf2e3ec19d] Destroying instance {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 939.033363] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3899eed-8077-4d52-b0e2-4fb65433018c {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.044965] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-680ab45d-0a61-469d-82c9-fe13d79ebbb0 tempest-InstanceActionsNegativeTestJSON-259524954 tempest-InstanceActionsNegativeTestJSON-259524954-project-member] [instance: 7e2db5f6-1419-415e-b957-66cf2e3ec19d] Powering off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 939.045410] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7d7a15f3-c32e-4aaa-8f27-4254db9609ec {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.053283] env[61957]: DEBUG oslo_vmware.api [None req-680ab45d-0a61-469d-82c9-fe13d79ebbb0 tempest-InstanceActionsNegativeTestJSON-259524954 tempest-InstanceActionsNegativeTestJSON-259524954-project-member] Waiting for the task: (returnval){ [ 939.053283] env[61957]: value = "task-1277772" [ 939.053283] env[61957]: _type = "Task" [ 939.053283] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.060762] env[61957]: DEBUG oslo_concurrency.lockutils [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Acquiring lock "refresh_cache-2e32aae5-edd7-402f-98ad-75a93d26f7a1" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 939.060945] env[61957]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-21445a7a-5b2c-45aa-865c-e27764cdb92d {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.067805] env[61957]: DEBUG oslo_vmware.api [None req-680ab45d-0a61-469d-82c9-fe13d79ebbb0 tempest-InstanceActionsNegativeTestJSON-259524954 tempest-InstanceActionsNegativeTestJSON-259524954-project-member] Task: {'id': task-1277772, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.075257] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-324b14b7-be2c-4c0f-9963-429c6a43b0d5 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.087869] env[61957]: DEBUG oslo_concurrency.lockutils [None req-483cfd5f-2659-4d70-aee0-52ce6ba7fa9a tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Releasing lock "refresh_cache-0d0f2d34-de35-4e80-8d9f-12693add0786" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 939.090876] env[61957]: DEBUG nova.compute.manager [None req-483cfd5f-2659-4d70-aee0-52ce6ba7fa9a tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] [instance: 0d0f2d34-de35-4e80-8d9f-12693add0786] Start destroying the instance on the hypervisor. {{(pid=61957) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 939.091267] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-483cfd5f-2659-4d70-aee0-52ce6ba7fa9a tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] [instance: 0d0f2d34-de35-4e80-8d9f-12693add0786] Destroying instance {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 939.092533] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3790a6b3-399c-474c-ba53-cf9993f4b38b {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.113383] env[61957]: DEBUG oslo_vmware.api [None req-78652638-0a34-4302-a770-2b339b434047 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Task: {'id': task-1277769, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.146917} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.128494] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-78652638-0a34-4302-a770-2b339b434047 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Deleted the datastore file {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 939.128780] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-78652638-0a34-4302-a770-2b339b434047 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] [instance: 28ab7d23-487f-4ae1-8fe4-58db55b59918] Deleted contents of the VM from datastore datastore2 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 939.129601] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-78652638-0a34-4302-a770-2b339b434047 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] [instance: 28ab7d23-487f-4ae1-8fe4-58db55b59918] Instance destroyed {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 939.129601] env[61957]: INFO nova.compute.manager [None req-78652638-0a34-4302-a770-2b339b434047 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] [instance: 28ab7d23-487f-4ae1-8fe4-58db55b59918] Took 1.16 seconds to destroy the instance on the hypervisor. [ 939.129601] env[61957]: DEBUG oslo.service.loopingcall [None req-78652638-0a34-4302-a770-2b339b434047 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 939.129822] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-483cfd5f-2659-4d70-aee0-52ce6ba7fa9a tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] [instance: 0d0f2d34-de35-4e80-8d9f-12693add0786] Powering off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 939.132803] env[61957]: DEBUG nova.compute.manager [req-11564a69-620e-4329-bf9a-ec3dc562ef1b req-7f3906a3-bb96-49b5-b7b1-d9bdc55881cf service nova] [instance: b598b8bb-919e-4404-b264-7b76161b0f79] Detach interface failed, port_id=be3f3b45-e91d-4c38-b746-a10c838decf7, reason: Instance b598b8bb-919e-4404-b264-7b76161b0f79 could not be found. {{(pid=61957) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 939.133164] env[61957]: DEBUG nova.compute.manager [-] [instance: 28ab7d23-487f-4ae1-8fe4-58db55b59918] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 939.133270] env[61957]: DEBUG nova.network.neutron [-] [instance: 28ab7d23-487f-4ae1-8fe4-58db55b59918] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 939.135056] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4c1e7324-9dd1-45f3-957a-d5a9f80756c1 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.147454] env[61957]: DEBUG oslo_vmware.api [None req-483cfd5f-2659-4d70-aee0-52ce6ba7fa9a tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Waiting for the task: (returnval){ [ 939.147454] env[61957]: value = "task-1277773" [ 939.147454] env[61957]: _type = "Task" [ 939.147454] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.157747] env[61957]: DEBUG oslo_vmware.api [None req-483cfd5f-2659-4d70-aee0-52ce6ba7fa9a tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Task: {'id': task-1277773, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.202850] env[61957]: DEBUG nova.network.neutron [req-9cb7126f-0512-48be-9451-549561edc6e2 req-f31a9876-7654-48e5-b03e-9a59474a165b service nova] [instance: 2e32aae5-edd7-402f-98ad-75a93d26f7a1] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 939.238412] env[61957]: DEBUG oslo_vmware.api [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Task: {'id': task-1277770, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.342945] env[61957]: INFO nova.compute.rpcapi [None req-23521104-b947-45c5-a62a-77ad796d9410 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Automatically selected compute RPC version 6.3 from minimum service version 67 [ 939.343606] env[61957]: DEBUG oslo_concurrency.lockutils [None req-23521104-b947-45c5-a62a-77ad796d9410 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Releasing lock "compute-rpcapi-router" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 939.394250] env[61957]: DEBUG nova.network.neutron [req-9cb7126f-0512-48be-9451-549561edc6e2 req-f31a9876-7654-48e5-b03e-9a59474a165b service nova] [instance: 2e32aae5-edd7-402f-98ad-75a93d26f7a1] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 939.520563] env[61957]: DEBUG oslo_vmware.api [None req-1036c4ee-30e7-4291-a697-7231cf748564 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Task: {'id': task-1277771, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.567229] env[61957]: DEBUG oslo_vmware.api [None req-680ab45d-0a61-469d-82c9-fe13d79ebbb0 tempest-InstanceActionsNegativeTestJSON-259524954 tempest-InstanceActionsNegativeTestJSON-259524954-project-member] Task: {'id': task-1277772, 'name': PowerOffVM_Task, 'duration_secs': 0.269615} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.568176] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-680ab45d-0a61-469d-82c9-fe13d79ebbb0 tempest-InstanceActionsNegativeTestJSON-259524954 tempest-InstanceActionsNegativeTestJSON-259524954-project-member] [instance: 7e2db5f6-1419-415e-b957-66cf2e3ec19d] Powered off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 939.568176] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-680ab45d-0a61-469d-82c9-fe13d79ebbb0 tempest-InstanceActionsNegativeTestJSON-259524954 tempest-InstanceActionsNegativeTestJSON-259524954-project-member] [instance: 7e2db5f6-1419-415e-b957-66cf2e3ec19d] Unregistering the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 939.568176] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-eb58c3b1-939f-4eb1-8a5e-3a6aeabc6f8a {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.643297] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-680ab45d-0a61-469d-82c9-fe13d79ebbb0 tempest-InstanceActionsNegativeTestJSON-259524954 tempest-InstanceActionsNegativeTestJSON-259524954-project-member] [instance: 7e2db5f6-1419-415e-b957-66cf2e3ec19d] Unregistered the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 939.643527] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-680ab45d-0a61-469d-82c9-fe13d79ebbb0 tempest-InstanceActionsNegativeTestJSON-259524954 tempest-InstanceActionsNegativeTestJSON-259524954-project-member] [instance: 7e2db5f6-1419-415e-b957-66cf2e3ec19d] Deleting contents of the VM from datastore datastore1 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 939.644382] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-680ab45d-0a61-469d-82c9-fe13d79ebbb0 tempest-InstanceActionsNegativeTestJSON-259524954 tempest-InstanceActionsNegativeTestJSON-259524954-project-member] Deleting the datastore file [datastore1] 7e2db5f6-1419-415e-b957-66cf2e3ec19d {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 939.644382] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f2774513-d30f-4c24-8513-911621161f67 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.654023] env[61957]: DEBUG oslo_vmware.api [None req-680ab45d-0a61-469d-82c9-fe13d79ebbb0 tempest-InstanceActionsNegativeTestJSON-259524954 tempest-InstanceActionsNegativeTestJSON-259524954-project-member] Waiting for the task: (returnval){ [ 939.654023] env[61957]: value = "task-1277775" [ 939.654023] env[61957]: _type = "Task" [ 939.654023] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.661374] env[61957]: DEBUG oslo_vmware.api [None req-483cfd5f-2659-4d70-aee0-52ce6ba7fa9a tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Task: {'id': task-1277773, 'name': PowerOffVM_Task, 'duration_secs': 0.179376} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.662362] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-483cfd5f-2659-4d70-aee0-52ce6ba7fa9a tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] [instance: 0d0f2d34-de35-4e80-8d9f-12693add0786] Powered off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 939.662493] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-483cfd5f-2659-4d70-aee0-52ce6ba7fa9a tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] [instance: 0d0f2d34-de35-4e80-8d9f-12693add0786] Unregistering the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 939.662883] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-816b711b-8a22-4d48-840e-d31064c6590e {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.668731] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d54dfa7-fb1d-4dc2-b89f-20020296805b {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.671340] env[61957]: DEBUG oslo_vmware.api [None req-680ab45d-0a61-469d-82c9-fe13d79ebbb0 tempest-InstanceActionsNegativeTestJSON-259524954 tempest-InstanceActionsNegativeTestJSON-259524954-project-member] Task: {'id': task-1277775, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.676082] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc5b2622-c43c-464d-b72c-b8acbc5a22c5 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.708742] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aeee5c61-85f5-4812-a508-81cf84775159 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.711583] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-483cfd5f-2659-4d70-aee0-52ce6ba7fa9a tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] [instance: 0d0f2d34-de35-4e80-8d9f-12693add0786] Unregistered the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 939.711792] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-483cfd5f-2659-4d70-aee0-52ce6ba7fa9a tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] [instance: 0d0f2d34-de35-4e80-8d9f-12693add0786] Deleting contents of the VM from datastore datastore2 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 939.711965] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-483cfd5f-2659-4d70-aee0-52ce6ba7fa9a tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Deleting the datastore file [datastore2] 0d0f2d34-de35-4e80-8d9f-12693add0786 {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 939.712255] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d7843dfc-1718-4803-929d-da29c9cba173 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.720960] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37341bde-93c6-4003-a2ed-83d8cad03953 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.725214] env[61957]: DEBUG oslo_vmware.api [None req-483cfd5f-2659-4d70-aee0-52ce6ba7fa9a tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Waiting for the task: (returnval){ [ 939.725214] env[61957]: value = "task-1277777" [ 939.725214] env[61957]: _type = "Task" [ 939.725214] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.739610] env[61957]: DEBUG nova.compute.provider_tree [None req-3e0cd30b-989c-4521-a40d-70a59bfcf72a tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 939.749325] env[61957]: DEBUG oslo_vmware.api [None req-483cfd5f-2659-4d70-aee0-52ce6ba7fa9a tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Task: {'id': task-1277777, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.753353] env[61957]: DEBUG oslo_vmware.api [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Task: {'id': task-1277770, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.865244] env[61957]: DEBUG oslo_concurrency.lockutils [None req-23521104-b947-45c5-a62a-77ad796d9410 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Acquiring lock "refresh_cache-dbeeb200-70b9-4cb4-b5a4-182389d21918" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 939.865456] env[61957]: DEBUG oslo_concurrency.lockutils [None req-23521104-b947-45c5-a62a-77ad796d9410 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Acquired lock "refresh_cache-dbeeb200-70b9-4cb4-b5a4-182389d21918" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 939.865716] env[61957]: DEBUG nova.network.neutron [None req-23521104-b947-45c5-a62a-77ad796d9410 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: dbeeb200-70b9-4cb4-b5a4-182389d21918] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 939.900934] env[61957]: DEBUG oslo_concurrency.lockutils [req-9cb7126f-0512-48be-9451-549561edc6e2 req-f31a9876-7654-48e5-b03e-9a59474a165b service nova] Releasing lock "refresh_cache-2e32aae5-edd7-402f-98ad-75a93d26f7a1" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 939.901372] env[61957]: DEBUG oslo_concurrency.lockutils [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Acquired lock "refresh_cache-2e32aae5-edd7-402f-98ad-75a93d26f7a1" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 939.901554] env[61957]: DEBUG nova.network.neutron [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: 2e32aae5-edd7-402f-98ad-75a93d26f7a1] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 940.020171] env[61957]: DEBUG oslo_vmware.api [None req-1036c4ee-30e7-4291-a697-7231cf748564 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Task: {'id': task-1277771, 'name': PowerOnVM_Task, 'duration_secs': 0.770201} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.020468] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-1036c4ee-30e7-4291-a697-7231cf748564 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] [instance: 9b0053fe-420e-4b92-86a8-ed44f6a6ec49] Powered on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 940.020662] env[61957]: INFO nova.compute.manager [None req-1036c4ee-30e7-4291-a697-7231cf748564 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] [instance: 9b0053fe-420e-4b92-86a8-ed44f6a6ec49] Took 7.64 seconds to spawn the instance on the hypervisor. [ 940.020913] env[61957]: DEBUG nova.compute.manager [None req-1036c4ee-30e7-4291-a697-7231cf748564 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] [instance: 9b0053fe-420e-4b92-86a8-ed44f6a6ec49] Checking state {{(pid=61957) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 940.021724] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5d37a06-cf21-480a-bde6-c1328154cf19 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.097679] env[61957]: DEBUG nova.network.neutron [-] [instance: 28ab7d23-487f-4ae1-8fe4-58db55b59918] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 940.165496] env[61957]: DEBUG oslo_vmware.api [None req-680ab45d-0a61-469d-82c9-fe13d79ebbb0 tempest-InstanceActionsNegativeTestJSON-259524954 tempest-InstanceActionsNegativeTestJSON-259524954-project-member] Task: {'id': task-1277775, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.212472} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.165954] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-680ab45d-0a61-469d-82c9-fe13d79ebbb0 tempest-InstanceActionsNegativeTestJSON-259524954 tempest-InstanceActionsNegativeTestJSON-259524954-project-member] Deleted the datastore file {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 940.166303] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-680ab45d-0a61-469d-82c9-fe13d79ebbb0 tempest-InstanceActionsNegativeTestJSON-259524954 tempest-InstanceActionsNegativeTestJSON-259524954-project-member] [instance: 7e2db5f6-1419-415e-b957-66cf2e3ec19d] Deleted contents of the VM from datastore datastore1 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 940.167288] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-680ab45d-0a61-469d-82c9-fe13d79ebbb0 tempest-InstanceActionsNegativeTestJSON-259524954 tempest-InstanceActionsNegativeTestJSON-259524954-project-member] [instance: 7e2db5f6-1419-415e-b957-66cf2e3ec19d] Instance destroyed {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 940.167682] env[61957]: INFO nova.compute.manager [None req-680ab45d-0a61-469d-82c9-fe13d79ebbb0 tempest-InstanceActionsNegativeTestJSON-259524954 tempest-InstanceActionsNegativeTestJSON-259524954-project-member] [instance: 7e2db5f6-1419-415e-b957-66cf2e3ec19d] Took 1.14 seconds to destroy the instance on the hypervisor. [ 940.168097] env[61957]: DEBUG oslo.service.loopingcall [None req-680ab45d-0a61-469d-82c9-fe13d79ebbb0 tempest-InstanceActionsNegativeTestJSON-259524954 tempest-InstanceActionsNegativeTestJSON-259524954-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 940.168443] env[61957]: DEBUG nova.compute.manager [-] [instance: 7e2db5f6-1419-415e-b957-66cf2e3ec19d] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 940.168635] env[61957]: DEBUG nova.network.neutron [-] [instance: 7e2db5f6-1419-415e-b957-66cf2e3ec19d] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 940.236128] env[61957]: DEBUG oslo_vmware.api [None req-483cfd5f-2659-4d70-aee0-52ce6ba7fa9a tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Task: {'id': task-1277777, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.11014} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.236128] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-483cfd5f-2659-4d70-aee0-52ce6ba7fa9a tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Deleted the datastore file {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 940.236547] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-483cfd5f-2659-4d70-aee0-52ce6ba7fa9a tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] [instance: 0d0f2d34-de35-4e80-8d9f-12693add0786] Deleted contents of the VM from datastore datastore2 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 940.236547] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-483cfd5f-2659-4d70-aee0-52ce6ba7fa9a tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] [instance: 0d0f2d34-de35-4e80-8d9f-12693add0786] Instance destroyed {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 940.236639] env[61957]: INFO nova.compute.manager [None req-483cfd5f-2659-4d70-aee0-52ce6ba7fa9a tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] [instance: 0d0f2d34-de35-4e80-8d9f-12693add0786] Took 1.15 seconds to destroy the instance on the hypervisor. [ 940.236879] env[61957]: DEBUG oslo.service.loopingcall [None req-483cfd5f-2659-4d70-aee0-52ce6ba7fa9a tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 940.237091] env[61957]: DEBUG nova.compute.manager [-] [instance: 0d0f2d34-de35-4e80-8d9f-12693add0786] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 940.237181] env[61957]: DEBUG nova.network.neutron [-] [instance: 0d0f2d34-de35-4e80-8d9f-12693add0786] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 940.250296] env[61957]: DEBUG nova.scheduler.client.report [None req-3e0cd30b-989c-4521-a40d-70a59bfcf72a tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 940.253046] env[61957]: DEBUG oslo_vmware.api [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Task: {'id': task-1277770, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.498756} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.253208] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore1] 50e09db7-9f8f-452b-8232-bd7473e9e63a/50e09db7-9f8f-452b-8232-bd7473e9e63a.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 940.253419] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: 50e09db7-9f8f-452b-8232-bd7473e9e63a] Extending root virtual disk to 1048576 {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 940.253667] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5f3e4faf-0650-4fc0-80dd-22a90a615457 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.257783] env[61957]: DEBUG nova.network.neutron [-] [instance: 0d0f2d34-de35-4e80-8d9f-12693add0786] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 940.260461] env[61957]: DEBUG oslo_vmware.api [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Waiting for the task: (returnval){ [ 940.260461] env[61957]: value = "task-1277778" [ 940.260461] env[61957]: _type = "Task" [ 940.260461] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.269139] env[61957]: DEBUG oslo_vmware.api [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Task: {'id': task-1277778, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.464430] env[61957]: DEBUG nova.network.neutron [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: 2e32aae5-edd7-402f-98ad-75a93d26f7a1] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 940.541053] env[61957]: INFO nova.compute.manager [None req-1036c4ee-30e7-4291-a697-7231cf748564 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] [instance: 9b0053fe-420e-4b92-86a8-ed44f6a6ec49] Took 39.44 seconds to build instance. [ 940.600428] env[61957]: INFO nova.compute.manager [-] [instance: 28ab7d23-487f-4ae1-8fe4-58db55b59918] Took 1.47 seconds to deallocate network for instance. [ 940.643302] env[61957]: DEBUG nova.compute.manager [req-c114d245-d80a-4f79-8cff-e1ea11b7bf21 req-2c26732a-1a75-43d9-9280-0b645c2b72d6 service nova] [instance: 28ab7d23-487f-4ae1-8fe4-58db55b59918] Received event network-vif-deleted-1018a669-af2d-42cb-9558-0043b6d775b2 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 940.643698] env[61957]: DEBUG nova.compute.manager [req-c114d245-d80a-4f79-8cff-e1ea11b7bf21 req-2c26732a-1a75-43d9-9280-0b645c2b72d6 service nova] [instance: 7e2db5f6-1419-415e-b957-66cf2e3ec19d] Received event network-vif-deleted-e3abccf7-4ab1-4e6e-b0f4-8680d99be3bd {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 940.643824] env[61957]: INFO nova.compute.manager [req-c114d245-d80a-4f79-8cff-e1ea11b7bf21 req-2c26732a-1a75-43d9-9280-0b645c2b72d6 service nova] [instance: 7e2db5f6-1419-415e-b957-66cf2e3ec19d] Neutron deleted interface e3abccf7-4ab1-4e6e-b0f4-8680d99be3bd; detaching it from the instance and deleting it from the info cache [ 940.643957] env[61957]: DEBUG nova.network.neutron [req-c114d245-d80a-4f79-8cff-e1ea11b7bf21 req-2c26732a-1a75-43d9-9280-0b645c2b72d6 service nova] [instance: 7e2db5f6-1419-415e-b957-66cf2e3ec19d] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 940.735557] env[61957]: DEBUG nova.network.neutron [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: 2e32aae5-edd7-402f-98ad-75a93d26f7a1] Updating instance_info_cache with network_info: [{"id": "a142b42a-efff-4834-a236-1d0ceb416178", "address": "fa:16:3e:ef:4f:71", "network": {"id": "2b5a7a9c-7143-4754-8d84-486c2ab13b21", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1508819166-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "55dd45b4291b456f83a446dfa87f550a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "572b7281-aad3-45fa-9cb2-fc1c70569948", "external-id": "nsx-vlan-transportzone-722", "segmentation_id": 722, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa142b42a-ef", "ovs_interfaceid": "a142b42a-efff-4834-a236-1d0ceb416178", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 940.757385] env[61957]: DEBUG oslo_concurrency.lockutils [None req-3e0cd30b-989c-4521-a40d-70a59bfcf72a tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.917s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 940.760961] env[61957]: DEBUG nova.network.neutron [None req-23521104-b947-45c5-a62a-77ad796d9410 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: dbeeb200-70b9-4cb4-b5a4-182389d21918] Updating instance_info_cache with network_info: [{"id": "86553395-da58-4c44-b9f4-c67db304d3fe", "address": "fa:16:3e:a4:1b:84", "network": {"id": "6fbccc17-6674-4cd5-8f1b-aa73570baac7", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-394544811-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1dbdc699741a48af93c52068d88a5357", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "20e3f794-c7a3-4696-9488-ecf34c570ef9", "external-id": "nsx-vlan-transportzone-509", "segmentation_id": 509, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap86553395-da", "ovs_interfaceid": "86553395-da58-4c44-b9f4-c67db304d3fe", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 940.762351] env[61957]: DEBUG oslo_concurrency.lockutils [None req-9213a8e4-435d-4ff1-a82b-fdbd92c25d20 tempest-ServersNegativeTestMultiTenantJSON-2135399494 tempest-ServersNegativeTestMultiTenantJSON-2135399494-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 15.549s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 940.762593] env[61957]: DEBUG nova.objects.instance [None req-9213a8e4-435d-4ff1-a82b-fdbd92c25d20 tempest-ServersNegativeTestMultiTenantJSON-2135399494 tempest-ServersNegativeTestMultiTenantJSON-2135399494-project-member] Lazy-loading 'resources' on Instance uuid 19966b0a-53b7-48c5-849c-a9d00dc024f8 {{(pid=61957) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 940.763947] env[61957]: DEBUG nova.network.neutron [-] [instance: 0d0f2d34-de35-4e80-8d9f-12693add0786] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 940.777028] env[61957]: DEBUG oslo_vmware.api [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Task: {'id': task-1277778, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067897} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.777028] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: 50e09db7-9f8f-452b-8232-bd7473e9e63a] Extended root virtual disk {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 940.777028] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7df9099f-4118-46f7-8a53-96397357190b {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.800923] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: 50e09db7-9f8f-452b-8232-bd7473e9e63a] Reconfiguring VM instance instance-00000052 to attach disk [datastore1] 50e09db7-9f8f-452b-8232-bd7473e9e63a/50e09db7-9f8f-452b-8232-bd7473e9e63a.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 940.802656] env[61957]: INFO nova.scheduler.client.report [None req-3e0cd30b-989c-4521-a40d-70a59bfcf72a tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Deleted allocations for instance c23141ee-0cbb-4d1b-8390-c3073fe354f1 [ 940.805294] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0d275406-1329-4206-a260-dd6e76f52240 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.829205] env[61957]: DEBUG oslo_vmware.api [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Waiting for the task: (returnval){ [ 940.829205] env[61957]: value = "task-1277779" [ 940.829205] env[61957]: _type = "Task" [ 940.829205] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.839180] env[61957]: DEBUG oslo_vmware.api [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Task: {'id': task-1277779, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.885526] env[61957]: DEBUG oslo_vmware.rw_handles [None req-4cbf1ed6-4756-412e-9889-a64859683b60 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/521952db-7e22-536a-9cfd-d20a4fa400ec/disk-0.vmdk. {{(pid=61957) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 940.886596] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cb1bb02-0643-48c0-aa4b-6b6dd4e2e360 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.893625] env[61957]: DEBUG oslo_vmware.rw_handles [None req-4cbf1ed6-4756-412e-9889-a64859683b60 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/521952db-7e22-536a-9cfd-d20a4fa400ec/disk-0.vmdk is in state: ready. {{(pid=61957) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 940.893796] env[61957]: ERROR oslo_vmware.rw_handles [None req-4cbf1ed6-4756-412e-9889-a64859683b60 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/521952db-7e22-536a-9cfd-d20a4fa400ec/disk-0.vmdk due to incomplete transfer. [ 940.894273] env[61957]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-029917fa-e001-4d2f-a4c8-0b6a7bd037b2 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.900987] env[61957]: DEBUG oslo_vmware.rw_handles [None req-4cbf1ed6-4756-412e-9889-a64859683b60 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/521952db-7e22-536a-9cfd-d20a4fa400ec/disk-0.vmdk. {{(pid=61957) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 940.901217] env[61957]: DEBUG nova.virt.vmwareapi.images [None req-4cbf1ed6-4756-412e-9889-a64859683b60 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: d4674fbc-47b1-42d1-aaba-e86d46c51e8f] Uploaded image fa797977-e3da-4f16-8445-e20c97e9e8ae to the Glance image server {{(pid=61957) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 940.903346] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-4cbf1ed6-4756-412e-9889-a64859683b60 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: d4674fbc-47b1-42d1-aaba-e86d46c51e8f] Destroying the VM {{(pid=61957) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 940.903603] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-622a569b-9c48-4d96-ab71-cb5701f1f5ab {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.909332] env[61957]: DEBUG oslo_vmware.api [None req-4cbf1ed6-4756-412e-9889-a64859683b60 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Waiting for the task: (returnval){ [ 940.909332] env[61957]: value = "task-1277780" [ 940.909332] env[61957]: _type = "Task" [ 940.909332] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.917449] env[61957]: DEBUG oslo_vmware.api [None req-4cbf1ed6-4756-412e-9889-a64859683b60 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': task-1277780, 'name': Destroy_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.022812] env[61957]: DEBUG nova.network.neutron [-] [instance: 7e2db5f6-1419-415e-b957-66cf2e3ec19d] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 941.044314] env[61957]: DEBUG oslo_concurrency.lockutils [None req-1036c4ee-30e7-4291-a697-7231cf748564 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Lock "9b0053fe-420e-4b92-86a8-ed44f6a6ec49" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 48.337s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 941.107829] env[61957]: DEBUG oslo_concurrency.lockutils [None req-78652638-0a34-4302-a770-2b339b434047 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 941.149691] env[61957]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7fab97bd-5000-46de-8fca-5492a0065e94 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.158839] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b01731e0-0821-4ddd-9872-ef49440db1f3 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.189555] env[61957]: DEBUG nova.compute.manager [req-c114d245-d80a-4f79-8cff-e1ea11b7bf21 req-2c26732a-1a75-43d9-9280-0b645c2b72d6 service nova] [instance: 7e2db5f6-1419-415e-b957-66cf2e3ec19d] Detach interface failed, port_id=e3abccf7-4ab1-4e6e-b0f4-8680d99be3bd, reason: Instance 7e2db5f6-1419-415e-b957-66cf2e3ec19d could not be found. {{(pid=61957) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 941.237490] env[61957]: DEBUG oslo_concurrency.lockutils [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Releasing lock "refresh_cache-2e32aae5-edd7-402f-98ad-75a93d26f7a1" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 941.239029] env[61957]: DEBUG nova.compute.manager [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: 2e32aae5-edd7-402f-98ad-75a93d26f7a1] Instance network_info: |[{"id": "a142b42a-efff-4834-a236-1d0ceb416178", "address": "fa:16:3e:ef:4f:71", "network": {"id": "2b5a7a9c-7143-4754-8d84-486c2ab13b21", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1508819166-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "55dd45b4291b456f83a446dfa87f550a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "572b7281-aad3-45fa-9cb2-fc1c70569948", "external-id": "nsx-vlan-transportzone-722", "segmentation_id": 722, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa142b42a-ef", "ovs_interfaceid": "a142b42a-efff-4834-a236-1d0ceb416178", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 941.239029] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: 2e32aae5-edd7-402f-98ad-75a93d26f7a1] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ef:4f:71', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '572b7281-aad3-45fa-9cb2-fc1c70569948', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a142b42a-efff-4834-a236-1d0ceb416178', 'vif_model': 'vmxnet3'}] {{(pid=61957) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 941.246645] env[61957]: DEBUG oslo.service.loopingcall [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 941.246879] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2e32aae5-edd7-402f-98ad-75a93d26f7a1] Creating VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 941.247124] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9a8794e9-0470-4c61-b0b6-01b5a43c99e0 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.267896] env[61957]: DEBUG oslo_concurrency.lockutils [None req-23521104-b947-45c5-a62a-77ad796d9410 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Releasing lock "refresh_cache-dbeeb200-70b9-4cb4-b5a4-182389d21918" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 941.270637] env[61957]: INFO nova.compute.manager [-] [instance: 0d0f2d34-de35-4e80-8d9f-12693add0786] Took 1.03 seconds to deallocate network for instance. [ 941.270879] env[61957]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 941.270879] env[61957]: value = "task-1277781" [ 941.270879] env[61957]: _type = "Task" [ 941.270879] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.284623] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277781, 'name': CreateVM_Task} progress is 6%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.327190] env[61957]: DEBUG oslo_concurrency.lockutils [None req-3e0cd30b-989c-4521-a40d-70a59bfcf72a tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Lock "c23141ee-0cbb-4d1b-8390-c3073fe354f1" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 22.220s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 941.339223] env[61957]: DEBUG oslo_vmware.api [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Task: {'id': task-1277779, 'name': ReconfigVM_Task, 'duration_secs': 0.292788} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.339410] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: 50e09db7-9f8f-452b-8232-bd7473e9e63a] Reconfigured VM instance instance-00000052 to attach disk [datastore1] 50e09db7-9f8f-452b-8232-bd7473e9e63a/50e09db7-9f8f-452b-8232-bd7473e9e63a.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 941.342306] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-393c1343-6a5c-4ec7-b10b-3771e55a3689 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.351149] env[61957]: DEBUG oslo_vmware.api [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Waiting for the task: (returnval){ [ 941.351149] env[61957]: value = "task-1277782" [ 941.351149] env[61957]: _type = "Task" [ 941.351149] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.366266] env[61957]: DEBUG oslo_vmware.api [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Task: {'id': task-1277782, 'name': Rename_Task} progress is 6%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.375645] env[61957]: DEBUG nova.compute.manager [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Didn't find any instances for network info cache update. {{(pid=61957) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10012}} [ 941.375876] env[61957]: DEBUG oslo_service.periodic_task [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61957) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 941.376141] env[61957]: DEBUG oslo_service.periodic_task [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61957) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 941.376341] env[61957]: DEBUG oslo_service.periodic_task [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61957) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 941.376534] env[61957]: DEBUG oslo_service.periodic_task [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61957) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 941.376721] env[61957]: DEBUG oslo_service.periodic_task [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61957) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 941.376907] env[61957]: DEBUG oslo_service.periodic_task [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61957) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 941.377067] env[61957]: DEBUG nova.compute.manager [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61957) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 941.377219] env[61957]: DEBUG oslo_service.periodic_task [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61957) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 941.421667] env[61957]: DEBUG oslo_vmware.api [None req-4cbf1ed6-4756-412e-9889-a64859683b60 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': task-1277780, 'name': Destroy_Task} progress is 33%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.525736] env[61957]: INFO nova.compute.manager [-] [instance: 7e2db5f6-1419-415e-b957-66cf2e3ec19d] Took 1.36 seconds to deallocate network for instance. [ 941.561069] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3dc839a-f93f-497f-a9a3-f0c4b092d8e1 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.570267] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-581ab05e-26cb-4688-9b36-73ec9c1d1138 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.604950] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33373469-4c4f-44d4-a963-cf7abec373c7 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.611842] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d1aa370b-6bd9-49b7-9a71-30e9793a2856 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Acquiring lock "00ecc8c2-398e-4a29-b19f-a2013985b481" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 941.612128] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d1aa370b-6bd9-49b7-9a71-30e9793a2856 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Lock "00ecc8c2-398e-4a29-b19f-a2013985b481" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 941.619131] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e042d556-565a-4d7d-adf1-a1091a098c11 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.633943] env[61957]: DEBUG nova.compute.provider_tree [None req-9213a8e4-435d-4ff1-a82b-fdbd92c25d20 tempest-ServersNegativeTestMultiTenantJSON-2135399494 tempest-ServersNegativeTestMultiTenantJSON-2135399494-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 941.780237] env[61957]: DEBUG oslo_concurrency.lockutils [None req-483cfd5f-2659-4d70-aee0-52ce6ba7fa9a tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 941.786562] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277781, 'name': CreateVM_Task} progress is 25%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.861523] env[61957]: DEBUG oslo_vmware.api [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Task: {'id': task-1277782, 'name': Rename_Task} progress is 14%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.880253] env[61957]: DEBUG oslo_concurrency.lockutils [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 941.919741] env[61957]: DEBUG oslo_vmware.api [None req-4cbf1ed6-4756-412e-9889-a64859683b60 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': task-1277780, 'name': Destroy_Task, 'duration_secs': 0.670947} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.920052] env[61957]: INFO nova.virt.vmwareapi.vm_util [None req-4cbf1ed6-4756-412e-9889-a64859683b60 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: d4674fbc-47b1-42d1-aaba-e86d46c51e8f] Destroyed the VM [ 941.920381] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-4cbf1ed6-4756-412e-9889-a64859683b60 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: d4674fbc-47b1-42d1-aaba-e86d46c51e8f] Deleting Snapshot of the VM instance {{(pid=61957) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 941.920641] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-0e81b24c-03bb-4d20-837a-4b8f37a13e78 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.927061] env[61957]: DEBUG oslo_vmware.api [None req-4cbf1ed6-4756-412e-9889-a64859683b60 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Waiting for the task: (returnval){ [ 941.927061] env[61957]: value = "task-1277783" [ 941.927061] env[61957]: _type = "Task" [ 941.927061] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.934406] env[61957]: DEBUG oslo_vmware.api [None req-4cbf1ed6-4756-412e-9889-a64859683b60 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': task-1277783, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.035432] env[61957]: DEBUG oslo_concurrency.lockutils [None req-680ab45d-0a61-469d-82c9-fe13d79ebbb0 tempest-InstanceActionsNegativeTestJSON-259524954 tempest-InstanceActionsNegativeTestJSON-259524954-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 942.115331] env[61957]: DEBUG nova.compute.manager [None req-d1aa370b-6bd9-49b7-9a71-30e9793a2856 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] [instance: 00ecc8c2-398e-4a29-b19f-a2013985b481] Starting instance... {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 942.137427] env[61957]: DEBUG nova.scheduler.client.report [None req-9213a8e4-435d-4ff1-a82b-fdbd92c25d20 tempest-ServersNegativeTestMultiTenantJSON-2135399494 tempest-ServersNegativeTestMultiTenantJSON-2135399494-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 942.277634] env[61957]: DEBUG oslo_concurrency.lockutils [None req-9e60b82d-4fee-4f83-a3da-b48f03d27cc5 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Acquiring lock "d6c5c70d-86de-4dea-8b9e-76f321947a35" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 942.277634] env[61957]: DEBUG oslo_concurrency.lockutils [None req-9e60b82d-4fee-4f83-a3da-b48f03d27cc5 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Lock "d6c5c70d-86de-4dea-8b9e-76f321947a35" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 942.293529] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277781, 'name': CreateVM_Task} progress is 99%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.366147] env[61957]: DEBUG oslo_vmware.api [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Task: {'id': task-1277782, 'name': Rename_Task, 'duration_secs': 0.868209} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.370043] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: 50e09db7-9f8f-452b-8232-bd7473e9e63a] Powering on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 942.370043] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-da16f530-e705-415c-95b0-9ab3fb581c2e {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.376973] env[61957]: DEBUG oslo_vmware.api [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Waiting for the task: (returnval){ [ 942.376973] env[61957]: value = "task-1277784" [ 942.376973] env[61957]: _type = "Task" [ 942.376973] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.388366] env[61957]: DEBUG oslo_vmware.api [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Task: {'id': task-1277784, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.436796] env[61957]: DEBUG oslo_vmware.api [None req-4cbf1ed6-4756-412e-9889-a64859683b60 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': task-1277783, 'name': RemoveSnapshot_Task} progress is 80%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.640712] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d1aa370b-6bd9-49b7-9a71-30e9793a2856 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 942.647026] env[61957]: DEBUG oslo_concurrency.lockutils [None req-9213a8e4-435d-4ff1-a82b-fdbd92c25d20 tempest-ServersNegativeTestMultiTenantJSON-2135399494 tempest-ServersNegativeTestMultiTenantJSON-2135399494-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.882s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 942.647026] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d1519ec7-d973-47ee-af77-990ab0a419d4 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.369s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 942.647936] env[61957]: INFO nova.compute.claims [None req-d1519ec7-d973-47ee-af77-990ab0a419d4 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] [instance: 7793aa07-6aa2-459d-8a91-56c6b0412d3b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 942.674305] env[61957]: INFO nova.scheduler.client.report [None req-9213a8e4-435d-4ff1-a82b-fdbd92c25d20 tempest-ServersNegativeTestMultiTenantJSON-2135399494 tempest-ServersNegativeTestMultiTenantJSON-2135399494-project-member] Deleted allocations for instance 19966b0a-53b7-48c5-849c-a9d00dc024f8 [ 942.783286] env[61957]: DEBUG nova.compute.manager [None req-9e60b82d-4fee-4f83-a3da-b48f03d27cc5 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: d6c5c70d-86de-4dea-8b9e-76f321947a35] Starting instance... {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 942.792269] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277781, 'name': CreateVM_Task, 'duration_secs': 1.10365} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.792732] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2e32aae5-edd7-402f-98ad-75a93d26f7a1] Created VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 942.793281] env[61957]: DEBUG oslo_concurrency.lockutils [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 942.793449] env[61957]: DEBUG oslo_concurrency.lockutils [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Acquired lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 942.794026] env[61957]: DEBUG oslo_concurrency.lockutils [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 942.794812] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fc433247-30d3-4ba2-9590-fa724942c6cf {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.797781] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4301df2-52c9-4e8c-8ac0-d2a106239674 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.817805] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-23521104-b947-45c5-a62a-77ad796d9410 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: dbeeb200-70b9-4cb4-b5a4-182389d21918] Updating instance 'dbeeb200-70b9-4cb4-b5a4-182389d21918' progress to 0 {{(pid=61957) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 942.822832] env[61957]: DEBUG oslo_vmware.api [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Waiting for the task: (returnval){ [ 942.822832] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]529b7d7b-1b7a-1362-8912-13f9e9bf10a4" [ 942.822832] env[61957]: _type = "Task" [ 942.822832] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.834180] env[61957]: DEBUG oslo_vmware.api [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]529b7d7b-1b7a-1362-8912-13f9e9bf10a4, 'name': SearchDatastore_Task, 'duration_secs': 0.01155} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.834634] env[61957]: DEBUG oslo_concurrency.lockutils [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Releasing lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 942.834771] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: 2e32aae5-edd7-402f-98ad-75a93d26f7a1] Processing image 11c76a2c-f705-470a-ba9d-4657858bab38 {{(pid=61957) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 942.835015] env[61957]: DEBUG oslo_concurrency.lockutils [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 942.835179] env[61957]: DEBUG oslo_concurrency.lockutils [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Acquired lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 942.835373] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 942.835731] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ccd40f3c-43ed-4f6f-ad7d-a626f896578a {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.844709] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 942.844983] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61957) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 942.845829] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-557e4bbb-d6fa-4a57-b942-305b416bac95 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.852936] env[61957]: DEBUG oslo_vmware.api [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Waiting for the task: (returnval){ [ 942.852936] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]525735fc-4476-3925-c5eb-40226bdb2307" [ 942.852936] env[61957]: _type = "Task" [ 942.852936] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.862496] env[61957]: DEBUG oslo_vmware.api [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]525735fc-4476-3925-c5eb-40226bdb2307, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.887034] env[61957]: DEBUG oslo_vmware.api [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Task: {'id': task-1277784, 'name': PowerOnVM_Task, 'duration_secs': 0.486823} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.887034] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: 50e09db7-9f8f-452b-8232-bd7473e9e63a] Powered on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 942.887206] env[61957]: INFO nova.compute.manager [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: 50e09db7-9f8f-452b-8232-bd7473e9e63a] Took 8.05 seconds to spawn the instance on the hypervisor. [ 942.887330] env[61957]: DEBUG nova.compute.manager [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: 50e09db7-9f8f-452b-8232-bd7473e9e63a] Checking state {{(pid=61957) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 942.888009] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c02240d-41bd-4905-b838-34ae5f6eb0af {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.937572] env[61957]: DEBUG oslo_vmware.api [None req-4cbf1ed6-4756-412e-9889-a64859683b60 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': task-1277783, 'name': RemoveSnapshot_Task, 'duration_secs': 0.873842} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.937869] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-4cbf1ed6-4756-412e-9889-a64859683b60 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: d4674fbc-47b1-42d1-aaba-e86d46c51e8f] Deleted Snapshot of the VM instance {{(pid=61957) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 942.938114] env[61957]: INFO nova.compute.manager [None req-4cbf1ed6-4756-412e-9889-a64859683b60 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: d4674fbc-47b1-42d1-aaba-e86d46c51e8f] Took 15.84 seconds to snapshot the instance on the hypervisor. [ 943.187164] env[61957]: DEBUG oslo_concurrency.lockutils [None req-9213a8e4-435d-4ff1-a82b-fdbd92c25d20 tempest-ServersNegativeTestMultiTenantJSON-2135399494 tempest-ServersNegativeTestMultiTenantJSON-2135399494-project-member] Lock "19966b0a-53b7-48c5-849c-a9d00dc024f8" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 20.915s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 943.308015] env[61957]: DEBUG oslo_concurrency.lockutils [None req-9e60b82d-4fee-4f83-a3da-b48f03d27cc5 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 943.325523] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-23521104-b947-45c5-a62a-77ad796d9410 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: dbeeb200-70b9-4cb4-b5a4-182389d21918] Powering off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 943.325870] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a6293f7f-a98d-4c93-825d-1358e1a59dd4 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.333971] env[61957]: DEBUG oslo_vmware.api [None req-23521104-b947-45c5-a62a-77ad796d9410 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Waiting for the task: (returnval){ [ 943.333971] env[61957]: value = "task-1277785" [ 943.333971] env[61957]: _type = "Task" [ 943.333971] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.342931] env[61957]: DEBUG oslo_vmware.api [None req-23521104-b947-45c5-a62a-77ad796d9410 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': task-1277785, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.363275] env[61957]: DEBUG oslo_vmware.api [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]525735fc-4476-3925-c5eb-40226bdb2307, 'name': SearchDatastore_Task, 'duration_secs': 0.013953} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.364122] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4d682419-7491-4837-9fe6-56040da15e9a {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.369779] env[61957]: DEBUG oslo_vmware.api [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Waiting for the task: (returnval){ [ 943.369779] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52c470f0-6a5a-5199-5104-06f67191777c" [ 943.369779] env[61957]: _type = "Task" [ 943.369779] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.377597] env[61957]: DEBUG oslo_vmware.api [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52c470f0-6a5a-5199-5104-06f67191777c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.406673] env[61957]: INFO nova.compute.manager [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: 50e09db7-9f8f-452b-8232-bd7473e9e63a] Took 40.71 seconds to build instance. [ 943.755310] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ce0965c3-bc7d-4a77-9d32-52f4d059d837 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Acquiring lock "964c8c9d-0721-4ad3-a146-fce1fd1d6513" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 943.755310] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ce0965c3-bc7d-4a77-9d32-52f4d059d837 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Lock "964c8c9d-0721-4ad3-a146-fce1fd1d6513" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 943.845263] env[61957]: DEBUG oslo_vmware.api [None req-23521104-b947-45c5-a62a-77ad796d9410 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': task-1277785, 'name': PowerOffVM_Task, 'duration_secs': 0.260783} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.847680] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-23521104-b947-45c5-a62a-77ad796d9410 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: dbeeb200-70b9-4cb4-b5a4-182389d21918] Powered off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 943.847878] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-23521104-b947-45c5-a62a-77ad796d9410 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: dbeeb200-70b9-4cb4-b5a4-182389d21918] Updating instance 'dbeeb200-70b9-4cb4-b5a4-182389d21918' progress to 17 {{(pid=61957) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 943.886870] env[61957]: DEBUG oslo_vmware.api [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52c470f0-6a5a-5199-5104-06f67191777c, 'name': SearchDatastore_Task, 'duration_secs': 0.021562} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.888903] env[61957]: DEBUG oslo_concurrency.lockutils [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Releasing lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 943.889223] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore1] 2e32aae5-edd7-402f-98ad-75a93d26f7a1/2e32aae5-edd7-402f-98ad-75a93d26f7a1.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 943.889659] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6fabfef3-286e-4fa9-b4d7-fc5ba9d15ed1 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.896120] env[61957]: DEBUG oslo_vmware.api [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Waiting for the task: (returnval){ [ 943.896120] env[61957]: value = "task-1277786" [ 943.896120] env[61957]: _type = "Task" [ 943.896120] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.906354] env[61957]: DEBUG oslo_vmware.api [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Task: {'id': task-1277786, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.908856] env[61957]: DEBUG oslo_concurrency.lockutils [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Lock "50e09db7-9f8f-452b-8232-bd7473e9e63a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 43.185s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 943.980211] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5302d432-ff4a-4974-a889-00a1c73f8304 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.988569] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd9ddd6e-4ba8-4211-85e7-57cd69982123 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.019393] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7570ef0-b3a1-418f-91de-2874846dfb41 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.026605] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90b63429-1f48-4073-9207-6751bee0a08f {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.039469] env[61957]: DEBUG nova.compute.provider_tree [None req-d1519ec7-d973-47ee-af77-990ab0a419d4 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 944.260940] env[61957]: DEBUG nova.compute.manager [None req-ce0965c3-bc7d-4a77-9d32-52f4d059d837 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] [instance: 964c8c9d-0721-4ad3-a146-fce1fd1d6513] Starting instance... {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 944.355032] env[61957]: DEBUG nova.virt.hardware [None req-23521104-b947-45c5-a62a-77ad796d9410 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T17:22:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=11c76a2c-f705-470a-ba9d-4657858bab38,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 944.355032] env[61957]: DEBUG nova.virt.hardware [None req-23521104-b947-45c5-a62a-77ad796d9410 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Flavor limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 944.355032] env[61957]: DEBUG nova.virt.hardware [None req-23521104-b947-45c5-a62a-77ad796d9410 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Image limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 944.355032] env[61957]: DEBUG nova.virt.hardware [None req-23521104-b947-45c5-a62a-77ad796d9410 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Flavor pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 944.355564] env[61957]: DEBUG nova.virt.hardware [None req-23521104-b947-45c5-a62a-77ad796d9410 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Image pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 944.355875] env[61957]: DEBUG nova.virt.hardware [None req-23521104-b947-45c5-a62a-77ad796d9410 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 944.356228] env[61957]: DEBUG nova.virt.hardware [None req-23521104-b947-45c5-a62a-77ad796d9410 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 944.356506] env[61957]: DEBUG nova.virt.hardware [None req-23521104-b947-45c5-a62a-77ad796d9410 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 944.359018] env[61957]: DEBUG nova.virt.hardware [None req-23521104-b947-45c5-a62a-77ad796d9410 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Got 1 possible topologies {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 944.359018] env[61957]: DEBUG nova.virt.hardware [None req-23521104-b947-45c5-a62a-77ad796d9410 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 944.359018] env[61957]: DEBUG nova.virt.hardware [None req-23521104-b947-45c5-a62a-77ad796d9410 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 944.362666] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-182b742c-bf0e-4267-a1cc-407847001ab2 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.381653] env[61957]: DEBUG oslo_vmware.api [None req-23521104-b947-45c5-a62a-77ad796d9410 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Waiting for the task: (returnval){ [ 944.381653] env[61957]: value = "task-1277787" [ 944.381653] env[61957]: _type = "Task" [ 944.381653] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.394947] env[61957]: DEBUG oslo_vmware.api [None req-23521104-b947-45c5-a62a-77ad796d9410 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': task-1277787, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.410092] env[61957]: DEBUG oslo_vmware.api [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Task: {'id': task-1277786, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.543651] env[61957]: DEBUG nova.scheduler.client.report [None req-d1519ec7-d973-47ee-af77-990ab0a419d4 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 944.798208] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ce0965c3-bc7d-4a77-9d32-52f4d059d837 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 944.892884] env[61957]: DEBUG oslo_vmware.api [None req-23521104-b947-45c5-a62a-77ad796d9410 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': task-1277787, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.905743] env[61957]: DEBUG oslo_vmware.api [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Task: {'id': task-1277786, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.820271} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.906814] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore1] 2e32aae5-edd7-402f-98ad-75a93d26f7a1/2e32aae5-edd7-402f-98ad-75a93d26f7a1.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 944.906814] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: 2e32aae5-edd7-402f-98ad-75a93d26f7a1] Extending root virtual disk to 1048576 {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 944.906990] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-cad2e8b5-938d-4e73-9658-32c9a4b56075 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.914271] env[61957]: DEBUG oslo_vmware.api [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Waiting for the task: (returnval){ [ 944.914271] env[61957]: value = "task-1277788" [ 944.914271] env[61957]: _type = "Task" [ 944.914271] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.922057] env[61957]: DEBUG oslo_vmware.api [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Task: {'id': task-1277788, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.048562] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d1519ec7-d973-47ee-af77-990ab0a419d4 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.402s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 945.049183] env[61957]: DEBUG nova.compute.manager [None req-d1519ec7-d973-47ee-af77-990ab0a419d4 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] [instance: 7793aa07-6aa2-459d-8a91-56c6b0412d3b] Start building networks asynchronously for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 945.052070] env[61957]: DEBUG oslo_concurrency.lockutils [None req-7678a123-82e2-420d-9405-a713c30b30cb tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.176s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 945.053559] env[61957]: INFO nova.compute.claims [None req-7678a123-82e2-420d-9405-a713c30b30cb tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 945.396665] env[61957]: DEBUG oslo_vmware.api [None req-23521104-b947-45c5-a62a-77ad796d9410 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': task-1277787, 'name': ReconfigVM_Task, 'duration_secs': 0.978329} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.397049] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-23521104-b947-45c5-a62a-77ad796d9410 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: dbeeb200-70b9-4cb4-b5a4-182389d21918] Updating instance 'dbeeb200-70b9-4cb4-b5a4-182389d21918' progress to 33 {{(pid=61957) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 945.424765] env[61957]: DEBUG oslo_vmware.api [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Task: {'id': task-1277788, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.38264} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.425162] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: 2e32aae5-edd7-402f-98ad-75a93d26f7a1] Extended root virtual disk {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 945.426641] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-185e98f3-5e78-4d9a-a866-24bc50684645 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.449587] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: 2e32aae5-edd7-402f-98ad-75a93d26f7a1] Reconfiguring VM instance instance-00000053 to attach disk [datastore1] 2e32aae5-edd7-402f-98ad-75a93d26f7a1/2e32aae5-edd7-402f-98ad-75a93d26f7a1.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 945.449917] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-aaa3dd6d-f53c-40f2-9b74-f8c92f733f55 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.471433] env[61957]: DEBUG oslo_vmware.api [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Waiting for the task: (returnval){ [ 945.471433] env[61957]: value = "task-1277789" [ 945.471433] env[61957]: _type = "Task" [ 945.471433] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.479772] env[61957]: DEBUG oslo_vmware.api [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Task: {'id': task-1277789, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.561932] env[61957]: DEBUG nova.compute.utils [None req-d1519ec7-d973-47ee-af77-990ab0a419d4 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Using /dev/sd instead of None {{(pid=61957) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 945.566341] env[61957]: DEBUG nova.compute.manager [None req-d1519ec7-d973-47ee-af77-990ab0a419d4 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] [instance: 7793aa07-6aa2-459d-8a91-56c6b0412d3b] Allocating IP information in the background. {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 945.566579] env[61957]: DEBUG nova.network.neutron [None req-d1519ec7-d973-47ee-af77-990ab0a419d4 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] [instance: 7793aa07-6aa2-459d-8a91-56c6b0412d3b] allocate_for_instance() {{(pid=61957) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 945.615124] env[61957]: DEBUG nova.policy [None req-d1519ec7-d973-47ee-af77-990ab0a419d4 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd6d0d8fdb5f94a7f99c24ec9db21673c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7758bdd10505455b8cecb8129d70ebd5', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61957) authorize /opt/stack/nova/nova/policy.py:203}} [ 945.870433] env[61957]: DEBUG nova.network.neutron [None req-d1519ec7-d973-47ee-af77-990ab0a419d4 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] [instance: 7793aa07-6aa2-459d-8a91-56c6b0412d3b] Successfully created port: dfb387b2-224a-48bc-a581-17c9f58b213a {{(pid=61957) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 945.906082] env[61957]: DEBUG nova.virt.hardware [None req-23521104-b947-45c5-a62a-77ad796d9410 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T17:22:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=11c76a2c-f705-470a-ba9d-4657858bab38,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 945.906424] env[61957]: DEBUG nova.virt.hardware [None req-23521104-b947-45c5-a62a-77ad796d9410 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Flavor limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 945.906529] env[61957]: DEBUG nova.virt.hardware [None req-23521104-b947-45c5-a62a-77ad796d9410 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Image limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 945.906698] env[61957]: DEBUG nova.virt.hardware [None req-23521104-b947-45c5-a62a-77ad796d9410 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Flavor pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 945.906849] env[61957]: DEBUG nova.virt.hardware [None req-23521104-b947-45c5-a62a-77ad796d9410 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Image pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 945.907031] env[61957]: DEBUG nova.virt.hardware [None req-23521104-b947-45c5-a62a-77ad796d9410 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 945.907383] env[61957]: DEBUG nova.virt.hardware [None req-23521104-b947-45c5-a62a-77ad796d9410 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 945.907556] env[61957]: DEBUG nova.virt.hardware [None req-23521104-b947-45c5-a62a-77ad796d9410 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 945.907729] env[61957]: DEBUG nova.virt.hardware [None req-23521104-b947-45c5-a62a-77ad796d9410 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Got 1 possible topologies {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 945.907897] env[61957]: DEBUG nova.virt.hardware [None req-23521104-b947-45c5-a62a-77ad796d9410 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 945.908096] env[61957]: DEBUG nova.virt.hardware [None req-23521104-b947-45c5-a62a-77ad796d9410 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 945.914885] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-23521104-b947-45c5-a62a-77ad796d9410 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: dbeeb200-70b9-4cb4-b5a4-182389d21918] Reconfiguring VM instance instance-0000004b to detach disk 2000 {{(pid=61957) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 945.915233] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2e6d90ef-0f9d-44ca-9d6f-085811a65b22 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.938433] env[61957]: DEBUG oslo_vmware.api [None req-23521104-b947-45c5-a62a-77ad796d9410 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Waiting for the task: (returnval){ [ 945.938433] env[61957]: value = "task-1277790" [ 945.938433] env[61957]: _type = "Task" [ 945.938433] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.950286] env[61957]: DEBUG oslo_vmware.api [None req-23521104-b947-45c5-a62a-77ad796d9410 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': task-1277790, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.982664] env[61957]: DEBUG oslo_vmware.api [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Task: {'id': task-1277789, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.066853] env[61957]: DEBUG nova.compute.manager [None req-d1519ec7-d973-47ee-af77-990ab0a419d4 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] [instance: 7793aa07-6aa2-459d-8a91-56c6b0412d3b] Start building block device mappings for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 946.362702] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c83fad69-7244-4968-9b9d-2f5215d1328c {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.375938] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6059d58c-3063-421b-9076-8d6cd26c1821 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.408190] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3e968f7-5e89-49ba-ad60-5dbcd9e51a6c {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.415773] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58ea8e86-2efd-4dae-8303-5feb345f9b05 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.429018] env[61957]: DEBUG nova.compute.provider_tree [None req-7678a123-82e2-420d-9405-a713c30b30cb tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 946.449831] env[61957]: DEBUG oslo_vmware.api [None req-23521104-b947-45c5-a62a-77ad796d9410 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': task-1277790, 'name': ReconfigVM_Task, 'duration_secs': 0.182643} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.449831] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-23521104-b947-45c5-a62a-77ad796d9410 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: dbeeb200-70b9-4cb4-b5a4-182389d21918] Reconfigured VM instance instance-0000004b to detach disk 2000 {{(pid=61957) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 946.449831] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01453e4c-2ed6-491e-be2b-db8fcd3268c9 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.470476] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-23521104-b947-45c5-a62a-77ad796d9410 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: dbeeb200-70b9-4cb4-b5a4-182389d21918] Reconfiguring VM instance instance-0000004b to attach disk [datastore1] dbeeb200-70b9-4cb4-b5a4-182389d21918/dbeeb200-70b9-4cb4-b5a4-182389d21918.vmdk or device None with type thin {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 946.470844] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9f068555-97f2-41e1-ba9d-87b9468091f0 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.493958] env[61957]: DEBUG oslo_vmware.api [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Task: {'id': task-1277789, 'name': ReconfigVM_Task, 'duration_secs': 0.573604} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.495245] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: 2e32aae5-edd7-402f-98ad-75a93d26f7a1] Reconfigured VM instance instance-00000053 to attach disk [datastore1] 2e32aae5-edd7-402f-98ad-75a93d26f7a1/2e32aae5-edd7-402f-98ad-75a93d26f7a1.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 946.495938] env[61957]: DEBUG oslo_vmware.api [None req-23521104-b947-45c5-a62a-77ad796d9410 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Waiting for the task: (returnval){ [ 946.495938] env[61957]: value = "task-1277791" [ 946.495938] env[61957]: _type = "Task" [ 946.495938] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.496632] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-216bdfe2-337b-4e30-a095-cde39d8549a1 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.505755] env[61957]: DEBUG oslo_vmware.api [None req-23521104-b947-45c5-a62a-77ad796d9410 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': task-1277791, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.506839] env[61957]: DEBUG oslo_vmware.api [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Waiting for the task: (returnval){ [ 946.506839] env[61957]: value = "task-1277792" [ 946.506839] env[61957]: _type = "Task" [ 946.506839] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.514222] env[61957]: DEBUG oslo_vmware.api [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Task: {'id': task-1277792, 'name': Rename_Task} progress is 5%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.933248] env[61957]: DEBUG nova.scheduler.client.report [None req-7678a123-82e2-420d-9405-a713c30b30cb tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 947.007869] env[61957]: DEBUG oslo_vmware.api [None req-23521104-b947-45c5-a62a-77ad796d9410 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': task-1277791, 'name': ReconfigVM_Task, 'duration_secs': 0.290259} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.007869] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-23521104-b947-45c5-a62a-77ad796d9410 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: dbeeb200-70b9-4cb4-b5a4-182389d21918] Reconfigured VM instance instance-0000004b to attach disk [datastore1] dbeeb200-70b9-4cb4-b5a4-182389d21918/dbeeb200-70b9-4cb4-b5a4-182389d21918.vmdk or device None with type thin {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 947.007869] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-23521104-b947-45c5-a62a-77ad796d9410 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: dbeeb200-70b9-4cb4-b5a4-182389d21918] Updating instance 'dbeeb200-70b9-4cb4-b5a4-182389d21918' progress to 50 {{(pid=61957) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 947.020355] env[61957]: DEBUG oslo_vmware.api [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Task: {'id': task-1277792, 'name': Rename_Task, 'duration_secs': 0.146401} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.020864] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: 2e32aae5-edd7-402f-98ad-75a93d26f7a1] Powering on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 947.020864] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-42fdc533-9f86-434b-8158-d50f97d1e035 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.026895] env[61957]: DEBUG oslo_vmware.api [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Waiting for the task: (returnval){ [ 947.026895] env[61957]: value = "task-1277793" [ 947.026895] env[61957]: _type = "Task" [ 947.026895] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.034502] env[61957]: DEBUG oslo_vmware.api [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Task: {'id': task-1277793, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.079946] env[61957]: DEBUG nova.compute.manager [None req-d1519ec7-d973-47ee-af77-990ab0a419d4 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] [instance: 7793aa07-6aa2-459d-8a91-56c6b0412d3b] Start spawning the instance on the hypervisor. {{(pid=61957) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 947.107256] env[61957]: DEBUG nova.virt.hardware [None req-d1519ec7-d973-47ee-af77-990ab0a419d4 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T17:22:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T17:21:50Z,direct_url=,disk_format='vmdk',id=11c76a2c-f705-470a-ba9d-4657858bab38,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='81d3096d0f094373913fc3dc8f5c3c6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T17:21:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 947.107955] env[61957]: DEBUG nova.virt.hardware [None req-d1519ec7-d973-47ee-af77-990ab0a419d4 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Flavor limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 947.108200] env[61957]: DEBUG nova.virt.hardware [None req-d1519ec7-d973-47ee-af77-990ab0a419d4 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Image limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 947.108471] env[61957]: DEBUG nova.virt.hardware [None req-d1519ec7-d973-47ee-af77-990ab0a419d4 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Flavor pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 947.108788] env[61957]: DEBUG nova.virt.hardware [None req-d1519ec7-d973-47ee-af77-990ab0a419d4 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Image pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 947.108974] env[61957]: DEBUG nova.virt.hardware [None req-d1519ec7-d973-47ee-af77-990ab0a419d4 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 947.109294] env[61957]: DEBUG nova.virt.hardware [None req-d1519ec7-d973-47ee-af77-990ab0a419d4 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 947.109493] env[61957]: DEBUG nova.virt.hardware [None req-d1519ec7-d973-47ee-af77-990ab0a419d4 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 947.109673] env[61957]: DEBUG nova.virt.hardware [None req-d1519ec7-d973-47ee-af77-990ab0a419d4 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Got 1 possible topologies {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 947.109973] env[61957]: DEBUG nova.virt.hardware [None req-d1519ec7-d973-47ee-af77-990ab0a419d4 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 947.110178] env[61957]: DEBUG nova.virt.hardware [None req-d1519ec7-d973-47ee-af77-990ab0a419d4 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 947.111070] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fd5c254-37fa-4d88-a294-031f5a6dd314 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.119570] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-651db6cc-6d93-4bb7-bf46-436438de2291 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.376047] env[61957]: DEBUG nova.compute.manager [req-6beadcb5-cc2d-4f1c-b7bd-3f4aa7b3d18e req-e0fe82c6-894f-470d-bf96-58ae8a7534a7 service nova] [instance: 7793aa07-6aa2-459d-8a91-56c6b0412d3b] Received event network-vif-plugged-dfb387b2-224a-48bc-a581-17c9f58b213a {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 947.376312] env[61957]: DEBUG oslo_concurrency.lockutils [req-6beadcb5-cc2d-4f1c-b7bd-3f4aa7b3d18e req-e0fe82c6-894f-470d-bf96-58ae8a7534a7 service nova] Acquiring lock "7793aa07-6aa2-459d-8a91-56c6b0412d3b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 947.376536] env[61957]: DEBUG oslo_concurrency.lockutils [req-6beadcb5-cc2d-4f1c-b7bd-3f4aa7b3d18e req-e0fe82c6-894f-470d-bf96-58ae8a7534a7 service nova] Lock "7793aa07-6aa2-459d-8a91-56c6b0412d3b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 947.376717] env[61957]: DEBUG oslo_concurrency.lockutils [req-6beadcb5-cc2d-4f1c-b7bd-3f4aa7b3d18e req-e0fe82c6-894f-470d-bf96-58ae8a7534a7 service nova] Lock "7793aa07-6aa2-459d-8a91-56c6b0412d3b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 947.376882] env[61957]: DEBUG nova.compute.manager [req-6beadcb5-cc2d-4f1c-b7bd-3f4aa7b3d18e req-e0fe82c6-894f-470d-bf96-58ae8a7534a7 service nova] [instance: 7793aa07-6aa2-459d-8a91-56c6b0412d3b] No waiting events found dispatching network-vif-plugged-dfb387b2-224a-48bc-a581-17c9f58b213a {{(pid=61957) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 947.377273] env[61957]: WARNING nova.compute.manager [req-6beadcb5-cc2d-4f1c-b7bd-3f4aa7b3d18e req-e0fe82c6-894f-470d-bf96-58ae8a7534a7 service nova] [instance: 7793aa07-6aa2-459d-8a91-56c6b0412d3b] Received unexpected event network-vif-plugged-dfb387b2-224a-48bc-a581-17c9f58b213a for instance with vm_state building and task_state spawning. [ 947.437667] env[61957]: DEBUG oslo_concurrency.lockutils [None req-7678a123-82e2-420d-9405-a713c30b30cb tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.385s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 947.438235] env[61957]: DEBUG nova.compute.manager [None req-7678a123-82e2-420d-9405-a713c30b30cb tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2] Start building networks asynchronously for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 947.441210] env[61957]: DEBUG oslo_concurrency.lockutils [None req-5d7e1ad8-1e31-4b98-a89f-2f6c8c624f1c tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 10.456s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 947.441448] env[61957]: DEBUG nova.objects.instance [None req-5d7e1ad8-1e31-4b98-a89f-2f6c8c624f1c tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Lazy-loading 'resources' on Instance uuid 39b8b1df-efdd-4c22-9f31-85c85be4f0eb {{(pid=61957) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 947.493866] env[61957]: DEBUG nova.network.neutron [None req-d1519ec7-d973-47ee-af77-990ab0a419d4 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] [instance: 7793aa07-6aa2-459d-8a91-56c6b0412d3b] Successfully updated port: dfb387b2-224a-48bc-a581-17c9f58b213a {{(pid=61957) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 947.518427] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa1b2408-1372-435c-b3b0-a1d5fb163684 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.543576] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36622741-a0a6-4c80-a515-908b6f94879b {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.552056] env[61957]: DEBUG oslo_vmware.api [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Task: {'id': task-1277793, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.565012] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-23521104-b947-45c5-a62a-77ad796d9410 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: dbeeb200-70b9-4cb4-b5a4-182389d21918] Updating instance 'dbeeb200-70b9-4cb4-b5a4-182389d21918' progress to 67 {{(pid=61957) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 947.953068] env[61957]: DEBUG nova.compute.utils [None req-7678a123-82e2-420d-9405-a713c30b30cb tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Using /dev/sd instead of None {{(pid=61957) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 947.957218] env[61957]: DEBUG nova.compute.manager [None req-7678a123-82e2-420d-9405-a713c30b30cb tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2] Allocating IP information in the background. {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 947.957554] env[61957]: DEBUG nova.network.neutron [None req-7678a123-82e2-420d-9405-a713c30b30cb tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2] allocate_for_instance() {{(pid=61957) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 947.996378] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d1519ec7-d973-47ee-af77-990ab0a419d4 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Acquiring lock "refresh_cache-7793aa07-6aa2-459d-8a91-56c6b0412d3b" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 947.996488] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d1519ec7-d973-47ee-af77-990ab0a419d4 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Acquired lock "refresh_cache-7793aa07-6aa2-459d-8a91-56c6b0412d3b" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 947.996613] env[61957]: DEBUG nova.network.neutron [None req-d1519ec7-d973-47ee-af77-990ab0a419d4 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] [instance: 7793aa07-6aa2-459d-8a91-56c6b0412d3b] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 948.008156] env[61957]: DEBUG nova.policy [None req-7678a123-82e2-420d-9405-a713c30b30cb tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6fd1c19dc3a44212ada44445e0919106', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8713aa35bcb24b86ad0b58ca9fc991ba', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61957) authorize /opt/stack/nova/nova/policy.py:203}} [ 948.048556] env[61957]: DEBUG oslo_vmware.api [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Task: {'id': task-1277793, 'name': PowerOnVM_Task, 'duration_secs': 0.800819} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.051044] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: 2e32aae5-edd7-402f-98ad-75a93d26f7a1] Powered on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 948.051308] env[61957]: INFO nova.compute.manager [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: 2e32aae5-edd7-402f-98ad-75a93d26f7a1] Took 10.67 seconds to spawn the instance on the hypervisor. [ 948.051498] env[61957]: DEBUG nova.compute.manager [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: 2e32aae5-edd7-402f-98ad-75a93d26f7a1] Checking state {{(pid=61957) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 948.053844] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7d2d866-65fd-4457-9876-6fcb293976b1 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.130495] env[61957]: DEBUG nova.network.neutron [None req-23521104-b947-45c5-a62a-77ad796d9410 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: dbeeb200-70b9-4cb4-b5a4-182389d21918] Port 86553395-da58-4c44-b9f4-c67db304d3fe binding to destination host cpu-1 is already ACTIVE {{(pid=61957) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 948.210545] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e43ea98-58d4-4c94-8d7a-84a6c4aa7ab4 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.218216] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21b0b657-8d08-4723-ae13-03a60233ad1a {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.248572] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a086209-b526-491f-82f1-32abb2a7e741 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.256396] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a811793a-9746-482c-af4f-155e5e796637 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.269110] env[61957]: DEBUG nova.compute.provider_tree [None req-5d7e1ad8-1e31-4b98-a89f-2f6c8c624f1c tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 948.327556] env[61957]: DEBUG nova.network.neutron [None req-7678a123-82e2-420d-9405-a713c30b30cb tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2] Successfully created port: f62a9f12-0b55-4138-a40e-b431855487fe {{(pid=61957) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 948.458279] env[61957]: DEBUG nova.compute.manager [None req-7678a123-82e2-420d-9405-a713c30b30cb tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2] Start building block device mappings for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 948.539973] env[61957]: DEBUG nova.network.neutron [None req-d1519ec7-d973-47ee-af77-990ab0a419d4 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] [instance: 7793aa07-6aa2-459d-8a91-56c6b0412d3b] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 948.574124] env[61957]: INFO nova.compute.manager [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: 2e32aae5-edd7-402f-98ad-75a93d26f7a1] Took 44.06 seconds to build instance. [ 948.689426] env[61957]: DEBUG nova.network.neutron [None req-d1519ec7-d973-47ee-af77-990ab0a419d4 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] [instance: 7793aa07-6aa2-459d-8a91-56c6b0412d3b] Updating instance_info_cache with network_info: [{"id": "dfb387b2-224a-48bc-a581-17c9f58b213a", "address": "fa:16:3e:f4:50:7f", "network": {"id": "d55c72c6-107d-4965-b4e1-d18383813953", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-193304765-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "7758bdd10505455b8cecb8129d70ebd5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c24464bb-bb6b-43a2-bdcd-8086ad1a307f", "external-id": "nsx-vlan-transportzone-781", "segmentation_id": 781, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdfb387b2-22", "ovs_interfaceid": "dfb387b2-224a-48bc-a581-17c9f58b213a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 948.772854] env[61957]: DEBUG nova.scheduler.client.report [None req-5d7e1ad8-1e31-4b98-a89f-2f6c8c624f1c tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 949.076995] env[61957]: DEBUG oslo_concurrency.lockutils [None req-e054d924-0bf1-48a5-a76f-3d1ec240b80a tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Lock "2e32aae5-edd7-402f-98ad-75a93d26f7a1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 48.321s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 949.152846] env[61957]: DEBUG oslo_concurrency.lockutils [None req-23521104-b947-45c5-a62a-77ad796d9410 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Acquiring lock "dbeeb200-70b9-4cb4-b5a4-182389d21918-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 949.153055] env[61957]: DEBUG oslo_concurrency.lockutils [None req-23521104-b947-45c5-a62a-77ad796d9410 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Lock "dbeeb200-70b9-4cb4-b5a4-182389d21918-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 949.153314] env[61957]: DEBUG oslo_concurrency.lockutils [None req-23521104-b947-45c5-a62a-77ad796d9410 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Lock "dbeeb200-70b9-4cb4-b5a4-182389d21918-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 949.192265] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d1519ec7-d973-47ee-af77-990ab0a419d4 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Releasing lock "refresh_cache-7793aa07-6aa2-459d-8a91-56c6b0412d3b" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 949.192637] env[61957]: DEBUG nova.compute.manager [None req-d1519ec7-d973-47ee-af77-990ab0a419d4 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] [instance: 7793aa07-6aa2-459d-8a91-56c6b0412d3b] Instance network_info: |[{"id": "dfb387b2-224a-48bc-a581-17c9f58b213a", "address": "fa:16:3e:f4:50:7f", "network": {"id": "d55c72c6-107d-4965-b4e1-d18383813953", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-193304765-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "7758bdd10505455b8cecb8129d70ebd5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c24464bb-bb6b-43a2-bdcd-8086ad1a307f", "external-id": "nsx-vlan-transportzone-781", "segmentation_id": 781, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdfb387b2-22", "ovs_interfaceid": "dfb387b2-224a-48bc-a581-17c9f58b213a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 949.193042] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-d1519ec7-d973-47ee-af77-990ab0a419d4 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] [instance: 7793aa07-6aa2-459d-8a91-56c6b0412d3b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f4:50:7f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c24464bb-bb6b-43a2-bdcd-8086ad1a307f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'dfb387b2-224a-48bc-a581-17c9f58b213a', 'vif_model': 'vmxnet3'}] {{(pid=61957) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 949.200609] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-d1519ec7-d973-47ee-af77-990ab0a419d4 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Creating folder: Project (7758bdd10505455b8cecb8129d70ebd5). Parent ref: group-v274445. {{(pid=61957) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 949.201173] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e41f5889-4fee-4a04-bd1c-41359fec4538 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.211719] env[61957]: INFO nova.virt.vmwareapi.vm_util [None req-d1519ec7-d973-47ee-af77-990ab0a419d4 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Created folder: Project (7758bdd10505455b8cecb8129d70ebd5) in parent group-v274445. [ 949.212196] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-d1519ec7-d973-47ee-af77-990ab0a419d4 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Creating folder: Instances. Parent ref: group-v274568. {{(pid=61957) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 949.212196] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d4a18436-e1f4-4ad8-a430-aed8f5754baa {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.221192] env[61957]: INFO nova.virt.vmwareapi.vm_util [None req-d1519ec7-d973-47ee-af77-990ab0a419d4 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Created folder: Instances in parent group-v274568. [ 949.221411] env[61957]: DEBUG oslo.service.loopingcall [None req-d1519ec7-d973-47ee-af77-990ab0a419d4 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 949.221598] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7793aa07-6aa2-459d-8a91-56c6b0412d3b] Creating VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 949.221804] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f2d3b30a-3aa5-4787-a916-38e5de8c7b85 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.241482] env[61957]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 949.241482] env[61957]: value = "task-1277796" [ 949.241482] env[61957]: _type = "Task" [ 949.241482] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.249487] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277796, 'name': CreateVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.278835] env[61957]: DEBUG oslo_concurrency.lockutils [None req-5d7e1ad8-1e31-4b98-a89f-2f6c8c624f1c tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.838s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 949.281457] env[61957]: DEBUG oslo_concurrency.lockutils [None req-73ed12d4-fc9a-46df-a24a-8dbe4ec86160 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 12.041s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 949.281722] env[61957]: DEBUG nova.objects.instance [None req-73ed12d4-fc9a-46df-a24a-8dbe4ec86160 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Lazy-loading 'resources' on Instance uuid b598b8bb-919e-4404-b264-7b76161b0f79 {{(pid=61957) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 949.304573] env[61957]: INFO nova.scheduler.client.report [None req-5d7e1ad8-1e31-4b98-a89f-2f6c8c624f1c tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Deleted allocations for instance 39b8b1df-efdd-4c22-9f31-85c85be4f0eb [ 949.404207] env[61957]: DEBUG nova.compute.manager [req-733ec390-69e3-4f63-a93f-bf2ee4e34e02 req-e8b04c6e-f684-422b-8b7d-0abf98f2c884 service nova] [instance: 7793aa07-6aa2-459d-8a91-56c6b0412d3b] Received event network-changed-dfb387b2-224a-48bc-a581-17c9f58b213a {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 949.404207] env[61957]: DEBUG nova.compute.manager [req-733ec390-69e3-4f63-a93f-bf2ee4e34e02 req-e8b04c6e-f684-422b-8b7d-0abf98f2c884 service nova] [instance: 7793aa07-6aa2-459d-8a91-56c6b0412d3b] Refreshing instance network info cache due to event network-changed-dfb387b2-224a-48bc-a581-17c9f58b213a. {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 949.404207] env[61957]: DEBUG oslo_concurrency.lockutils [req-733ec390-69e3-4f63-a93f-bf2ee4e34e02 req-e8b04c6e-f684-422b-8b7d-0abf98f2c884 service nova] Acquiring lock "refresh_cache-7793aa07-6aa2-459d-8a91-56c6b0412d3b" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 949.404207] env[61957]: DEBUG oslo_concurrency.lockutils [req-733ec390-69e3-4f63-a93f-bf2ee4e34e02 req-e8b04c6e-f684-422b-8b7d-0abf98f2c884 service nova] Acquired lock "refresh_cache-7793aa07-6aa2-459d-8a91-56c6b0412d3b" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 949.404207] env[61957]: DEBUG nova.network.neutron [req-733ec390-69e3-4f63-a93f-bf2ee4e34e02 req-e8b04c6e-f684-422b-8b7d-0abf98f2c884 service nova] [instance: 7793aa07-6aa2-459d-8a91-56c6b0412d3b] Refreshing network info cache for port dfb387b2-224a-48bc-a581-17c9f58b213a {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 949.471742] env[61957]: DEBUG nova.compute.manager [None req-7678a123-82e2-420d-9405-a713c30b30cb tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2] Start spawning the instance on the hypervisor. {{(pid=61957) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 949.499382] env[61957]: DEBUG nova.virt.hardware [None req-7678a123-82e2-420d-9405-a713c30b30cb tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T17:22:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T17:21:50Z,direct_url=,disk_format='vmdk',id=11c76a2c-f705-470a-ba9d-4657858bab38,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='81d3096d0f094373913fc3dc8f5c3c6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T17:21:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 949.499826] env[61957]: DEBUG nova.virt.hardware [None req-7678a123-82e2-420d-9405-a713c30b30cb tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Flavor limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 949.500123] env[61957]: DEBUG nova.virt.hardware [None req-7678a123-82e2-420d-9405-a713c30b30cb tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Image limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 949.500476] env[61957]: DEBUG nova.virt.hardware [None req-7678a123-82e2-420d-9405-a713c30b30cb tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Flavor pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 949.500773] env[61957]: DEBUG nova.virt.hardware [None req-7678a123-82e2-420d-9405-a713c30b30cb tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Image pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 949.501056] env[61957]: DEBUG nova.virt.hardware [None req-7678a123-82e2-420d-9405-a713c30b30cb tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 949.501462] env[61957]: DEBUG nova.virt.hardware [None req-7678a123-82e2-420d-9405-a713c30b30cb tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 949.501735] env[61957]: DEBUG nova.virt.hardware [None req-7678a123-82e2-420d-9405-a713c30b30cb tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 949.502030] env[61957]: DEBUG nova.virt.hardware [None req-7678a123-82e2-420d-9405-a713c30b30cb tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Got 1 possible topologies {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 949.502319] env[61957]: DEBUG nova.virt.hardware [None req-7678a123-82e2-420d-9405-a713c30b30cb tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 949.502624] env[61957]: DEBUG nova.virt.hardware [None req-7678a123-82e2-420d-9405-a713c30b30cb tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 949.504093] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7086ab4e-fee3-4013-b1b1-6342283cea5a {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.518801] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81125f10-48a9-4954-9e79-87a00c1fa4c7 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.752934] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277796, 'name': CreateVM_Task, 'duration_secs': 0.467618} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.753191] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7793aa07-6aa2-459d-8a91-56c6b0412d3b] Created VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 949.753952] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d1519ec7-d973-47ee-af77-990ab0a419d4 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 949.754183] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d1519ec7-d973-47ee-af77-990ab0a419d4 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 949.754549] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d1519ec7-d973-47ee-af77-990ab0a419d4 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 949.754848] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fe5f8d37-3a19-446a-b57d-59b94ce8f3a0 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.759744] env[61957]: DEBUG oslo_vmware.api [None req-d1519ec7-d973-47ee-af77-990ab0a419d4 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Waiting for the task: (returnval){ [ 949.759744] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]5286c3a5-e842-66d3-cb3a-c72e0948cf02" [ 949.759744] env[61957]: _type = "Task" [ 949.759744] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.767387] env[61957]: DEBUG oslo_vmware.api [None req-d1519ec7-d973-47ee-af77-990ab0a419d4 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]5286c3a5-e842-66d3-cb3a-c72e0948cf02, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.812303] env[61957]: DEBUG oslo_concurrency.lockutils [None req-5d7e1ad8-1e31-4b98-a89f-2f6c8c624f1c tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Lock "39b8b1df-efdd-4c22-9f31-85c85be4f0eb" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 17.765s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 949.920902] env[61957]: DEBUG nova.network.neutron [None req-7678a123-82e2-420d-9405-a713c30b30cb tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2] Successfully updated port: f62a9f12-0b55-4138-a40e-b431855487fe {{(pid=61957) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 950.053588] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ba8a6c3-eea7-4307-bd4f-540824a11923 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.060914] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a85e9dae-4d17-4d6b-9e39-efcac5ada4bb {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.096271] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-280bc8a4-4e02-4994-b0b4-c261d82c506c {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.104539] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-426a02e1-9c8a-4574-8c43-a4b27dc3bc2a {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.109612] env[61957]: DEBUG nova.network.neutron [req-733ec390-69e3-4f63-a93f-bf2ee4e34e02 req-e8b04c6e-f684-422b-8b7d-0abf98f2c884 service nova] [instance: 7793aa07-6aa2-459d-8a91-56c6b0412d3b] Updated VIF entry in instance network info cache for port dfb387b2-224a-48bc-a581-17c9f58b213a. {{(pid=61957) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 950.109964] env[61957]: DEBUG nova.network.neutron [req-733ec390-69e3-4f63-a93f-bf2ee4e34e02 req-e8b04c6e-f684-422b-8b7d-0abf98f2c884 service nova] [instance: 7793aa07-6aa2-459d-8a91-56c6b0412d3b] Updating instance_info_cache with network_info: [{"id": "dfb387b2-224a-48bc-a581-17c9f58b213a", "address": "fa:16:3e:f4:50:7f", "network": {"id": "d55c72c6-107d-4965-b4e1-d18383813953", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-193304765-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "7758bdd10505455b8cecb8129d70ebd5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c24464bb-bb6b-43a2-bdcd-8086ad1a307f", "external-id": "nsx-vlan-transportzone-781", "segmentation_id": 781, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdfb387b2-22", "ovs_interfaceid": "dfb387b2-224a-48bc-a581-17c9f58b213a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 950.123181] env[61957]: DEBUG nova.compute.provider_tree [None req-73ed12d4-fc9a-46df-a24a-8dbe4ec86160 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 950.207681] env[61957]: DEBUG oslo_concurrency.lockutils [None req-23521104-b947-45c5-a62a-77ad796d9410 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Acquiring lock "refresh_cache-dbeeb200-70b9-4cb4-b5a4-182389d21918" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 950.208310] env[61957]: DEBUG oslo_concurrency.lockutils [None req-23521104-b947-45c5-a62a-77ad796d9410 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Acquired lock "refresh_cache-dbeeb200-70b9-4cb4-b5a4-182389d21918" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 950.208310] env[61957]: DEBUG nova.network.neutron [None req-23521104-b947-45c5-a62a-77ad796d9410 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: dbeeb200-70b9-4cb4-b5a4-182389d21918] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 950.274629] env[61957]: DEBUG oslo_vmware.api [None req-d1519ec7-d973-47ee-af77-990ab0a419d4 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]5286c3a5-e842-66d3-cb3a-c72e0948cf02, 'name': SearchDatastore_Task, 'duration_secs': 0.010284} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.275205] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d1519ec7-d973-47ee-af77-990ab0a419d4 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 950.275542] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-d1519ec7-d973-47ee-af77-990ab0a419d4 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] [instance: 7793aa07-6aa2-459d-8a91-56c6b0412d3b] Processing image 11c76a2c-f705-470a-ba9d-4657858bab38 {{(pid=61957) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 950.275875] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d1519ec7-d973-47ee-af77-990ab0a419d4 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 950.276118] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d1519ec7-d973-47ee-af77-990ab0a419d4 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 950.276392] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-d1519ec7-d973-47ee-af77-990ab0a419d4 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 950.276739] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-67a518e5-7b6c-40eb-8846-51e2bd3f6162 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.287048] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-d1519ec7-d973-47ee-af77-990ab0a419d4 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 950.287294] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-d1519ec7-d973-47ee-af77-990ab0a419d4 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61957) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 950.288122] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2acb8fa9-a88c-4880-9a6f-02b8f1132a68 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.294137] env[61957]: DEBUG oslo_vmware.api [None req-d1519ec7-d973-47ee-af77-990ab0a419d4 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Waiting for the task: (returnval){ [ 950.294137] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52f337b4-758a-912b-13d2-7a5a4c9561fd" [ 950.294137] env[61957]: _type = "Task" [ 950.294137] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.302474] env[61957]: DEBUG oslo_vmware.api [None req-d1519ec7-d973-47ee-af77-990ab0a419d4 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52f337b4-758a-912b-13d2-7a5a4c9561fd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.307031] env[61957]: DEBUG oslo_concurrency.lockutils [None req-5dd00032-4725-4b73-a1a7-7ef95ae8be93 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Acquiring lock "50e09db7-9f8f-452b-8232-bd7473e9e63a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 950.307312] env[61957]: DEBUG oslo_concurrency.lockutils [None req-5dd00032-4725-4b73-a1a7-7ef95ae8be93 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Lock "50e09db7-9f8f-452b-8232-bd7473e9e63a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 950.307555] env[61957]: DEBUG oslo_concurrency.lockutils [None req-5dd00032-4725-4b73-a1a7-7ef95ae8be93 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Acquiring lock "50e09db7-9f8f-452b-8232-bd7473e9e63a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 950.307813] env[61957]: DEBUG oslo_concurrency.lockutils [None req-5dd00032-4725-4b73-a1a7-7ef95ae8be93 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Lock "50e09db7-9f8f-452b-8232-bd7473e9e63a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 950.307948] env[61957]: DEBUG oslo_concurrency.lockutils [None req-5dd00032-4725-4b73-a1a7-7ef95ae8be93 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Lock "50e09db7-9f8f-452b-8232-bd7473e9e63a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 950.310279] env[61957]: INFO nova.compute.manager [None req-5dd00032-4725-4b73-a1a7-7ef95ae8be93 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: 50e09db7-9f8f-452b-8232-bd7473e9e63a] Terminating instance [ 950.312555] env[61957]: DEBUG nova.compute.manager [None req-5dd00032-4725-4b73-a1a7-7ef95ae8be93 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: 50e09db7-9f8f-452b-8232-bd7473e9e63a] Start destroying the instance on the hypervisor. {{(pid=61957) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 950.312787] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-5dd00032-4725-4b73-a1a7-7ef95ae8be93 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: 50e09db7-9f8f-452b-8232-bd7473e9e63a] Destroying instance {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 950.313651] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95e824fc-96cc-47cd-b8af-03a08f72f06c {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.321694] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-5dd00032-4725-4b73-a1a7-7ef95ae8be93 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: 50e09db7-9f8f-452b-8232-bd7473e9e63a] Powering off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 950.321960] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2aa3a04b-bd41-443a-a17f-1a33153c857d {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.328266] env[61957]: DEBUG oslo_vmware.api [None req-5dd00032-4725-4b73-a1a7-7ef95ae8be93 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Waiting for the task: (returnval){ [ 950.328266] env[61957]: value = "task-1277797" [ 950.328266] env[61957]: _type = "Task" [ 950.328266] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.335767] env[61957]: DEBUG oslo_vmware.api [None req-5dd00032-4725-4b73-a1a7-7ef95ae8be93 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Task: {'id': task-1277797, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.387452] env[61957]: DEBUG oslo_concurrency.lockutils [None req-7c0e7a82-e685-47c5-90d8-2de75a11cf75 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Acquiring lock "2e32aae5-edd7-402f-98ad-75a93d26f7a1" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 950.387614] env[61957]: DEBUG oslo_concurrency.lockutils [None req-7c0e7a82-e685-47c5-90d8-2de75a11cf75 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Lock "2e32aae5-edd7-402f-98ad-75a93d26f7a1" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 950.388392] env[61957]: DEBUG oslo_concurrency.lockutils [None req-7c0e7a82-e685-47c5-90d8-2de75a11cf75 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Acquiring lock "2e32aae5-edd7-402f-98ad-75a93d26f7a1-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 950.388392] env[61957]: DEBUG oslo_concurrency.lockutils [None req-7c0e7a82-e685-47c5-90d8-2de75a11cf75 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Lock "2e32aae5-edd7-402f-98ad-75a93d26f7a1-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 950.388561] env[61957]: DEBUG oslo_concurrency.lockutils [None req-7c0e7a82-e685-47c5-90d8-2de75a11cf75 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Lock "2e32aae5-edd7-402f-98ad-75a93d26f7a1-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 950.390846] env[61957]: INFO nova.compute.manager [None req-7c0e7a82-e685-47c5-90d8-2de75a11cf75 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: 2e32aae5-edd7-402f-98ad-75a93d26f7a1] Terminating instance [ 950.392782] env[61957]: DEBUG nova.compute.manager [None req-7c0e7a82-e685-47c5-90d8-2de75a11cf75 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: 2e32aae5-edd7-402f-98ad-75a93d26f7a1] Start destroying the instance on the hypervisor. {{(pid=61957) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 950.392985] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-7c0e7a82-e685-47c5-90d8-2de75a11cf75 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: 2e32aae5-edd7-402f-98ad-75a93d26f7a1] Destroying instance {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 950.393868] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c25ed70-d5b8-49fd-8751-6df2c6542fa6 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.402085] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-7c0e7a82-e685-47c5-90d8-2de75a11cf75 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: 2e32aae5-edd7-402f-98ad-75a93d26f7a1] Powering off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 950.402448] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-375c5899-46ca-42e6-bc91-f5446b1b40ce {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.409596] env[61957]: DEBUG oslo_vmware.api [None req-7c0e7a82-e685-47c5-90d8-2de75a11cf75 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Waiting for the task: (returnval){ [ 950.409596] env[61957]: value = "task-1277798" [ 950.409596] env[61957]: _type = "Task" [ 950.409596] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.417675] env[61957]: DEBUG oslo_vmware.api [None req-7c0e7a82-e685-47c5-90d8-2de75a11cf75 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Task: {'id': task-1277798, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.422376] env[61957]: DEBUG oslo_concurrency.lockutils [None req-7678a123-82e2-420d-9405-a713c30b30cb tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Acquiring lock "refresh_cache-08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 950.422535] env[61957]: DEBUG oslo_concurrency.lockutils [None req-7678a123-82e2-420d-9405-a713c30b30cb tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Acquired lock "refresh_cache-08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 950.422697] env[61957]: DEBUG nova.network.neutron [None req-7678a123-82e2-420d-9405-a713c30b30cb tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 950.612883] env[61957]: DEBUG oslo_concurrency.lockutils [req-733ec390-69e3-4f63-a93f-bf2ee4e34e02 req-e8b04c6e-f684-422b-8b7d-0abf98f2c884 service nova] Releasing lock "refresh_cache-7793aa07-6aa2-459d-8a91-56c6b0412d3b" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 950.629096] env[61957]: DEBUG nova.scheduler.client.report [None req-73ed12d4-fc9a-46df-a24a-8dbe4ec86160 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 950.804878] env[61957]: DEBUG oslo_vmware.api [None req-d1519ec7-d973-47ee-af77-990ab0a419d4 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52f337b4-758a-912b-13d2-7a5a4c9561fd, 'name': SearchDatastore_Task, 'duration_secs': 0.009437} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.805673] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ea23b1c8-3925-47d4-9fb7-c8ad71c342b8 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.811060] env[61957]: DEBUG oslo_vmware.api [None req-d1519ec7-d973-47ee-af77-990ab0a419d4 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Waiting for the task: (returnval){ [ 950.811060] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]524bc961-ec78-df72-b3b1-548d527234e9" [ 950.811060] env[61957]: _type = "Task" [ 950.811060] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.818754] env[61957]: DEBUG oslo_vmware.api [None req-d1519ec7-d973-47ee-af77-990ab0a419d4 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]524bc961-ec78-df72-b3b1-548d527234e9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.837432] env[61957]: DEBUG oslo_vmware.api [None req-5dd00032-4725-4b73-a1a7-7ef95ae8be93 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Task: {'id': task-1277797, 'name': PowerOffVM_Task, 'duration_secs': 0.263601} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.837704] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-5dd00032-4725-4b73-a1a7-7ef95ae8be93 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: 50e09db7-9f8f-452b-8232-bd7473e9e63a] Powered off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 950.837875] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-5dd00032-4725-4b73-a1a7-7ef95ae8be93 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: 50e09db7-9f8f-452b-8232-bd7473e9e63a] Unregistering the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 950.838155] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ef709888-e53e-44bc-9813-239ba5303136 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.922951] env[61957]: DEBUG oslo_vmware.api [None req-7c0e7a82-e685-47c5-90d8-2de75a11cf75 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Task: {'id': task-1277798, 'name': PowerOffVM_Task, 'duration_secs': 0.187259} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.923315] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-7c0e7a82-e685-47c5-90d8-2de75a11cf75 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: 2e32aae5-edd7-402f-98ad-75a93d26f7a1] Powered off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 950.923507] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-7c0e7a82-e685-47c5-90d8-2de75a11cf75 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: 2e32aae5-edd7-402f-98ad-75a93d26f7a1] Unregistering the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 950.923734] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f6b9e36f-8ac8-4cc6-a158-9ab0554ee1f0 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.926163] env[61957]: DEBUG nova.network.neutron [None req-23521104-b947-45c5-a62a-77ad796d9410 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: dbeeb200-70b9-4cb4-b5a4-182389d21918] Updating instance_info_cache with network_info: [{"id": "86553395-da58-4c44-b9f4-c67db304d3fe", "address": "fa:16:3e:a4:1b:84", "network": {"id": "6fbccc17-6674-4cd5-8f1b-aa73570baac7", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-394544811-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1dbdc699741a48af93c52068d88a5357", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "20e3f794-c7a3-4696-9488-ecf34c570ef9", "external-id": "nsx-vlan-transportzone-509", "segmentation_id": 509, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap86553395-da", "ovs_interfaceid": "86553395-da58-4c44-b9f4-c67db304d3fe", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 950.956193] env[61957]: DEBUG oslo_concurrency.lockutils [None req-3141e8c0-2e34-4c8b-a8ac-0f12d351422c tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Acquiring lock "736eabcf-70f8-4e1e-9830-270d9971fc9a" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 950.956584] env[61957]: DEBUG oslo_concurrency.lockutils [None req-3141e8c0-2e34-4c8b-a8ac-0f12d351422c tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Lock "736eabcf-70f8-4e1e-9830-270d9971fc9a" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.001s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 950.956831] env[61957]: INFO nova.compute.manager [None req-3141e8c0-2e34-4c8b-a8ac-0f12d351422c tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] [instance: 736eabcf-70f8-4e1e-9830-270d9971fc9a] Shelving [ 950.958896] env[61957]: DEBUG nova.network.neutron [None req-7678a123-82e2-420d-9405-a713c30b30cb tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 951.101042] env[61957]: DEBUG nova.network.neutron [None req-7678a123-82e2-420d-9405-a713c30b30cb tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2] Updating instance_info_cache with network_info: [{"id": "f62a9f12-0b55-4138-a40e-b431855487fe", "address": "fa:16:3e:af:48:d3", "network": {"id": "364cabcb-046d-4f91-b1ee-5dd3adfae6d7", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1432396967-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8713aa35bcb24b86ad0b58ca9fc991ba", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b7a73c01-1bb9-4612-a1a7-16d71b732e81", "external-id": "nsx-vlan-transportzone-711", "segmentation_id": 711, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf62a9f12-0b", "ovs_interfaceid": "f62a9f12-0b55-4138-a40e-b431855487fe", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 951.132439] env[61957]: DEBUG oslo_concurrency.lockutils [None req-73ed12d4-fc9a-46df-a24a-8dbe4ec86160 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.851s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 951.137331] env[61957]: DEBUG oslo_concurrency.lockutils [None req-78652638-0a34-4302-a770-2b339b434047 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 10.029s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 951.137483] env[61957]: DEBUG nova.objects.instance [None req-78652638-0a34-4302-a770-2b339b434047 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Lazy-loading 'resources' on Instance uuid 28ab7d23-487f-4ae1-8fe4-58db55b59918 {{(pid=61957) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 951.158892] env[61957]: INFO nova.scheduler.client.report [None req-73ed12d4-fc9a-46df-a24a-8dbe4ec86160 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Deleted allocations for instance b598b8bb-919e-4404-b264-7b76161b0f79 [ 951.324244] env[61957]: DEBUG oslo_vmware.api [None req-d1519ec7-d973-47ee-af77-990ab0a419d4 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]524bc961-ec78-df72-b3b1-548d527234e9, 'name': SearchDatastore_Task, 'duration_secs': 0.013767} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.324244] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d1519ec7-d973-47ee-af77-990ab0a419d4 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 951.324244] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-d1519ec7-d973-47ee-af77-990ab0a419d4 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore2] 7793aa07-6aa2-459d-8a91-56c6b0412d3b/7793aa07-6aa2-459d-8a91-56c6b0412d3b.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 951.324244] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1ccabd09-154c-4c16-af0b-8f22d11e90d3 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.330058] env[61957]: DEBUG oslo_vmware.api [None req-d1519ec7-d973-47ee-af77-990ab0a419d4 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Waiting for the task: (returnval){ [ 951.330058] env[61957]: value = "task-1277801" [ 951.330058] env[61957]: _type = "Task" [ 951.330058] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.337054] env[61957]: DEBUG oslo_vmware.api [None req-d1519ec7-d973-47ee-af77-990ab0a419d4 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Task: {'id': task-1277801, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.430430] env[61957]: DEBUG oslo_concurrency.lockutils [None req-23521104-b947-45c5-a62a-77ad796d9410 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Releasing lock "refresh_cache-dbeeb200-70b9-4cb4-b5a4-182389d21918" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 951.467532] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-3141e8c0-2e34-4c8b-a8ac-0f12d351422c tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] [instance: 736eabcf-70f8-4e1e-9830-270d9971fc9a] Powering off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 951.467846] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5464b0f0-7ccf-4357-bb38-364eef8a7c1e {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.475283] env[61957]: DEBUG oslo_vmware.api [None req-3141e8c0-2e34-4c8b-a8ac-0f12d351422c tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Waiting for the task: (returnval){ [ 951.475283] env[61957]: value = "task-1277802" [ 951.475283] env[61957]: _type = "Task" [ 951.475283] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.477626] env[61957]: DEBUG nova.compute.manager [req-1beb0f1d-56bf-4918-8e6d-180c001b49b1 req-8d6e48c3-6549-45e0-ae82-0859528c7da0 service nova] [instance: 08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2] Received event network-vif-plugged-f62a9f12-0b55-4138-a40e-b431855487fe {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 951.477795] env[61957]: DEBUG oslo_concurrency.lockutils [req-1beb0f1d-56bf-4918-8e6d-180c001b49b1 req-8d6e48c3-6549-45e0-ae82-0859528c7da0 service nova] Acquiring lock "08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 951.478048] env[61957]: DEBUG oslo_concurrency.lockutils [req-1beb0f1d-56bf-4918-8e6d-180c001b49b1 req-8d6e48c3-6549-45e0-ae82-0859528c7da0 service nova] Lock "08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 951.478239] env[61957]: DEBUG oslo_concurrency.lockutils [req-1beb0f1d-56bf-4918-8e6d-180c001b49b1 req-8d6e48c3-6549-45e0-ae82-0859528c7da0 service nova] Lock "08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 951.478422] env[61957]: DEBUG nova.compute.manager [req-1beb0f1d-56bf-4918-8e6d-180c001b49b1 req-8d6e48c3-6549-45e0-ae82-0859528c7da0 service nova] [instance: 08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2] No waiting events found dispatching network-vif-plugged-f62a9f12-0b55-4138-a40e-b431855487fe {{(pid=61957) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 951.478624] env[61957]: WARNING nova.compute.manager [req-1beb0f1d-56bf-4918-8e6d-180c001b49b1 req-8d6e48c3-6549-45e0-ae82-0859528c7da0 service nova] [instance: 08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2] Received unexpected event network-vif-plugged-f62a9f12-0b55-4138-a40e-b431855487fe for instance with vm_state building and task_state spawning. [ 951.478849] env[61957]: DEBUG nova.compute.manager [req-1beb0f1d-56bf-4918-8e6d-180c001b49b1 req-8d6e48c3-6549-45e0-ae82-0859528c7da0 service nova] [instance: 08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2] Received event network-changed-f62a9f12-0b55-4138-a40e-b431855487fe {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 951.479051] env[61957]: DEBUG nova.compute.manager [req-1beb0f1d-56bf-4918-8e6d-180c001b49b1 req-8d6e48c3-6549-45e0-ae82-0859528c7da0 service nova] [instance: 08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2] Refreshing instance network info cache due to event network-changed-f62a9f12-0b55-4138-a40e-b431855487fe. {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 951.479340] env[61957]: DEBUG oslo_concurrency.lockutils [req-1beb0f1d-56bf-4918-8e6d-180c001b49b1 req-8d6e48c3-6549-45e0-ae82-0859528c7da0 service nova] Acquiring lock "refresh_cache-08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 951.489525] env[61957]: DEBUG oslo_vmware.api [None req-3141e8c0-2e34-4c8b-a8ac-0f12d351422c tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Task: {'id': task-1277802, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.603820] env[61957]: DEBUG oslo_concurrency.lockutils [None req-7678a123-82e2-420d-9405-a713c30b30cb tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Releasing lock "refresh_cache-08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 951.604074] env[61957]: DEBUG nova.compute.manager [None req-7678a123-82e2-420d-9405-a713c30b30cb tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2] Instance network_info: |[{"id": "f62a9f12-0b55-4138-a40e-b431855487fe", "address": "fa:16:3e:af:48:d3", "network": {"id": "364cabcb-046d-4f91-b1ee-5dd3adfae6d7", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1432396967-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8713aa35bcb24b86ad0b58ca9fc991ba", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b7a73c01-1bb9-4612-a1a7-16d71b732e81", "external-id": "nsx-vlan-transportzone-711", "segmentation_id": 711, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf62a9f12-0b", "ovs_interfaceid": "f62a9f12-0b55-4138-a40e-b431855487fe", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 951.604955] env[61957]: DEBUG oslo_concurrency.lockutils [req-1beb0f1d-56bf-4918-8e6d-180c001b49b1 req-8d6e48c3-6549-45e0-ae82-0859528c7da0 service nova] Acquired lock "refresh_cache-08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 951.604955] env[61957]: DEBUG nova.network.neutron [req-1beb0f1d-56bf-4918-8e6d-180c001b49b1 req-8d6e48c3-6549-45e0-ae82-0859528c7da0 service nova] [instance: 08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2] Refreshing network info cache for port f62a9f12-0b55-4138-a40e-b431855487fe {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 951.606379] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-7678a123-82e2-420d-9405-a713c30b30cb tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:af:48:d3', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'b7a73c01-1bb9-4612-a1a7-16d71b732e81', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f62a9f12-0b55-4138-a40e-b431855487fe', 'vif_model': 'vmxnet3'}] {{(pid=61957) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 951.616972] env[61957]: DEBUG oslo.service.loopingcall [None req-7678a123-82e2-420d-9405-a713c30b30cb tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 951.621329] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2] Creating VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 951.622860] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-04d9b289-78cd-4ebc-85d0-ff9622d593f2 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.655606] env[61957]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 951.655606] env[61957]: value = "task-1277803" [ 951.655606] env[61957]: _type = "Task" [ 951.655606] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.670697] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277803, 'name': CreateVM_Task} progress is 5%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.674547] env[61957]: DEBUG oslo_concurrency.lockutils [None req-73ed12d4-fc9a-46df-a24a-8dbe4ec86160 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Lock "b598b8bb-919e-4404-b264-7b76161b0f79" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 17.909s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 951.703637] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-5dd00032-4725-4b73-a1a7-7ef95ae8be93 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: 50e09db7-9f8f-452b-8232-bd7473e9e63a] Unregistered the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 951.703882] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-5dd00032-4725-4b73-a1a7-7ef95ae8be93 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: 50e09db7-9f8f-452b-8232-bd7473e9e63a] Deleting contents of the VM from datastore datastore1 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 951.704098] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-5dd00032-4725-4b73-a1a7-7ef95ae8be93 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Deleting the datastore file [datastore1] 50e09db7-9f8f-452b-8232-bd7473e9e63a {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 951.707026] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4d6b4700-4f6f-4d7b-86d5-8976b1c9dd6b {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.713967] env[61957]: DEBUG oslo_vmware.api [None req-5dd00032-4725-4b73-a1a7-7ef95ae8be93 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Waiting for the task: (returnval){ [ 951.713967] env[61957]: value = "task-1277804" [ 951.713967] env[61957]: _type = "Task" [ 951.713967] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.720271] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-7c0e7a82-e685-47c5-90d8-2de75a11cf75 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: 2e32aae5-edd7-402f-98ad-75a93d26f7a1] Unregistered the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 951.721095] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-7c0e7a82-e685-47c5-90d8-2de75a11cf75 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: 2e32aae5-edd7-402f-98ad-75a93d26f7a1] Deleting contents of the VM from datastore datastore1 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 951.721095] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-7c0e7a82-e685-47c5-90d8-2de75a11cf75 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Deleting the datastore file [datastore1] 2e32aae5-edd7-402f-98ad-75a93d26f7a1 {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 951.723976] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6b744bf8-0893-437b-b6a3-2ed0875ef8d7 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.726188] env[61957]: DEBUG oslo_vmware.api [None req-5dd00032-4725-4b73-a1a7-7ef95ae8be93 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Task: {'id': task-1277804, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.735211] env[61957]: DEBUG oslo_vmware.api [None req-7c0e7a82-e685-47c5-90d8-2de75a11cf75 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Waiting for the task: (returnval){ [ 951.735211] env[61957]: value = "task-1277805" [ 951.735211] env[61957]: _type = "Task" [ 951.735211] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.744856] env[61957]: DEBUG oslo_vmware.api [None req-7c0e7a82-e685-47c5-90d8-2de75a11cf75 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Task: {'id': task-1277805, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.840744] env[61957]: DEBUG oslo_vmware.api [None req-d1519ec7-d973-47ee-af77-990ab0a419d4 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Task: {'id': task-1277801, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.884456] env[61957]: DEBUG nova.network.neutron [req-1beb0f1d-56bf-4918-8e6d-180c001b49b1 req-8d6e48c3-6549-45e0-ae82-0859528c7da0 service nova] [instance: 08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2] Updated VIF entry in instance network info cache for port f62a9f12-0b55-4138-a40e-b431855487fe. {{(pid=61957) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 951.884904] env[61957]: DEBUG nova.network.neutron [req-1beb0f1d-56bf-4918-8e6d-180c001b49b1 req-8d6e48c3-6549-45e0-ae82-0859528c7da0 service nova] [instance: 08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2] Updating instance_info_cache with network_info: [{"id": "f62a9f12-0b55-4138-a40e-b431855487fe", "address": "fa:16:3e:af:48:d3", "network": {"id": "364cabcb-046d-4f91-b1ee-5dd3adfae6d7", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1432396967-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8713aa35bcb24b86ad0b58ca9fc991ba", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b7a73c01-1bb9-4612-a1a7-16d71b732e81", "external-id": "nsx-vlan-transportzone-711", "segmentation_id": 711, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf62a9f12-0b", "ovs_interfaceid": "f62a9f12-0b55-4138-a40e-b431855487fe", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 951.954195] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38fb730d-61c7-402c-8e8e-0d6a179d8ce0 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.958613] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60ee9c5a-d22e-4771-a59b-512c801786f0 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.978081] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97aeea38-d6c8-4225-bea7-3bb064db77fb {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.985194] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d491339-833b-4c38-b2f4-cf4d6d46b937 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.166400] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9f6fa5a-5006-4964-b190-deebcda404ae {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.166400] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-23521104-b947-45c5-a62a-77ad796d9410 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: dbeeb200-70b9-4cb4-b5a4-182389d21918] Updating instance 'dbeeb200-70b9-4cb4-b5a4-182389d21918' progress to 83 {{(pid=61957) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 952.166400] env[61957]: DEBUG oslo_vmware.api [None req-3141e8c0-2e34-4c8b-a8ac-0f12d351422c tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Task: {'id': task-1277802, 'name': PowerOffVM_Task, 'duration_secs': 0.189096} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.166400] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-3141e8c0-2e34-4c8b-a8ac-0f12d351422c tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] [instance: 736eabcf-70f8-4e1e-9830-270d9971fc9a] Powered off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 952.166400] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21651022-b48d-436a-9792-6656dcd479dc {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.166400] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48230e67-00ea-4932-9818-6b0f3b0f0e52 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.166400] env[61957]: DEBUG nova.compute.provider_tree [None req-78652638-0a34-4302-a770-2b339b434047 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 952.166400] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88d76c36-6afc-48ab-b2d7-3528fca7efa5 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.168022] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277803, 'name': CreateVM_Task, 'duration_secs': 0.481389} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.168022] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2] Created VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 952.168564] env[61957]: DEBUG oslo_concurrency.lockutils [None req-7678a123-82e2-420d-9405-a713c30b30cb tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 952.168783] env[61957]: DEBUG oslo_concurrency.lockutils [None req-7678a123-82e2-420d-9405-a713c30b30cb tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 952.169480] env[61957]: DEBUG oslo_concurrency.lockutils [None req-7678a123-82e2-420d-9405-a713c30b30cb tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 952.169480] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-04a66850-b678-484f-8b01-aa26bc127a17 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.174658] env[61957]: DEBUG oslo_vmware.api [None req-7678a123-82e2-420d-9405-a713c30b30cb tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Waiting for the task: (returnval){ [ 952.174658] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]523ca31a-8f13-5d57-26c0-2ab6f1d48eaa" [ 952.174658] env[61957]: _type = "Task" [ 952.174658] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.182593] env[61957]: DEBUG oslo_vmware.api [None req-7678a123-82e2-420d-9405-a713c30b30cb tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]523ca31a-8f13-5d57-26c0-2ab6f1d48eaa, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.223314] env[61957]: DEBUG oslo_vmware.api [None req-5dd00032-4725-4b73-a1a7-7ef95ae8be93 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Task: {'id': task-1277804, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.275924} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.223468] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-5dd00032-4725-4b73-a1a7-7ef95ae8be93 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Deleted the datastore file {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 952.223672] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-5dd00032-4725-4b73-a1a7-7ef95ae8be93 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: 50e09db7-9f8f-452b-8232-bd7473e9e63a] Deleted contents of the VM from datastore datastore1 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 952.223925] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-5dd00032-4725-4b73-a1a7-7ef95ae8be93 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: 50e09db7-9f8f-452b-8232-bd7473e9e63a] Instance destroyed {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 952.224136] env[61957]: INFO nova.compute.manager [None req-5dd00032-4725-4b73-a1a7-7ef95ae8be93 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: 50e09db7-9f8f-452b-8232-bd7473e9e63a] Took 1.91 seconds to destroy the instance on the hypervisor. [ 952.224383] env[61957]: DEBUG oslo.service.loopingcall [None req-5dd00032-4725-4b73-a1a7-7ef95ae8be93 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 952.225984] env[61957]: DEBUG nova.compute.manager [-] [instance: 50e09db7-9f8f-452b-8232-bd7473e9e63a] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 952.226107] env[61957]: DEBUG nova.network.neutron [-] [instance: 50e09db7-9f8f-452b-8232-bd7473e9e63a] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 952.230107] env[61957]: DEBUG oslo_concurrency.lockutils [None req-cedd08d9-20a0-42aa-9f1f-33859d216b7f tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Acquiring lock "98929ed3-d420-45cd-9cde-9738b2ac8251" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 952.230335] env[61957]: DEBUG oslo_concurrency.lockutils [None req-cedd08d9-20a0-42aa-9f1f-33859d216b7f tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Lock "98929ed3-d420-45cd-9cde-9738b2ac8251" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 952.247690] env[61957]: DEBUG oslo_vmware.api [None req-7c0e7a82-e685-47c5-90d8-2de75a11cf75 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Task: {'id': task-1277805, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.266821} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.247974] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-7c0e7a82-e685-47c5-90d8-2de75a11cf75 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Deleted the datastore file {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 952.248183] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-7c0e7a82-e685-47c5-90d8-2de75a11cf75 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: 2e32aae5-edd7-402f-98ad-75a93d26f7a1] Deleted contents of the VM from datastore datastore1 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 952.248367] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-7c0e7a82-e685-47c5-90d8-2de75a11cf75 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: 2e32aae5-edd7-402f-98ad-75a93d26f7a1] Instance destroyed {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 952.248539] env[61957]: INFO nova.compute.manager [None req-7c0e7a82-e685-47c5-90d8-2de75a11cf75 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] [instance: 2e32aae5-edd7-402f-98ad-75a93d26f7a1] Took 1.86 seconds to destroy the instance on the hypervisor. [ 952.248791] env[61957]: DEBUG oslo.service.loopingcall [None req-7c0e7a82-e685-47c5-90d8-2de75a11cf75 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 952.248977] env[61957]: DEBUG nova.compute.manager [-] [instance: 2e32aae5-edd7-402f-98ad-75a93d26f7a1] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 952.249087] env[61957]: DEBUG nova.network.neutron [-] [instance: 2e32aae5-edd7-402f-98ad-75a93d26f7a1] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 952.341250] env[61957]: DEBUG oslo_vmware.api [None req-d1519ec7-d973-47ee-af77-990ab0a419d4 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Task: {'id': task-1277801, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.555369} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.341619] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-d1519ec7-d973-47ee-af77-990ab0a419d4 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore2] 7793aa07-6aa2-459d-8a91-56c6b0412d3b/7793aa07-6aa2-459d-8a91-56c6b0412d3b.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 952.341883] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-d1519ec7-d973-47ee-af77-990ab0a419d4 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] [instance: 7793aa07-6aa2-459d-8a91-56c6b0412d3b] Extending root virtual disk to 1048576 {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 952.342156] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d6b6ede7-e2c1-4d8c-bcc2-ac438a6c9748 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.348218] env[61957]: DEBUG oslo_vmware.api [None req-d1519ec7-d973-47ee-af77-990ab0a419d4 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Waiting for the task: (returnval){ [ 952.348218] env[61957]: value = "task-1277806" [ 952.348218] env[61957]: _type = "Task" [ 952.348218] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.358112] env[61957]: DEBUG oslo_vmware.api [None req-d1519ec7-d973-47ee-af77-990ab0a419d4 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Task: {'id': task-1277806, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.388036] env[61957]: DEBUG oslo_concurrency.lockutils [req-1beb0f1d-56bf-4918-8e6d-180c001b49b1 req-8d6e48c3-6549-45e0-ae82-0859528c7da0 service nova] Releasing lock "refresh_cache-08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 952.535645] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-23521104-b947-45c5-a62a-77ad796d9410 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: dbeeb200-70b9-4cb4-b5a4-182389d21918] Powering on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 952.536013] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7edcaa88-318c-462d-bb6a-56e2d1749603 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.544745] env[61957]: DEBUG oslo_vmware.api [None req-23521104-b947-45c5-a62a-77ad796d9410 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Waiting for the task: (returnval){ [ 952.544745] env[61957]: value = "task-1277807" [ 952.544745] env[61957]: _type = "Task" [ 952.544745] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.559881] env[61957]: DEBUG oslo_vmware.api [None req-23521104-b947-45c5-a62a-77ad796d9410 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': task-1277807, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.579219] env[61957]: DEBUG nova.scheduler.client.report [None req-78652638-0a34-4302-a770-2b339b434047 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 952.587478] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-3141e8c0-2e34-4c8b-a8ac-0f12d351422c tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] [instance: 736eabcf-70f8-4e1e-9830-270d9971fc9a] Creating Snapshot of the VM instance {{(pid=61957) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 952.588085] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-efc81350-baf6-443a-982c-0ca73f1b5c3b {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.596082] env[61957]: DEBUG oslo_vmware.api [None req-3141e8c0-2e34-4c8b-a8ac-0f12d351422c tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Waiting for the task: (returnval){ [ 952.596082] env[61957]: value = "task-1277808" [ 952.596082] env[61957]: _type = "Task" [ 952.596082] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.606228] env[61957]: DEBUG oslo_vmware.api [None req-3141e8c0-2e34-4c8b-a8ac-0f12d351422c tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Task: {'id': task-1277808, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.686542] env[61957]: DEBUG oslo_vmware.api [None req-7678a123-82e2-420d-9405-a713c30b30cb tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]523ca31a-8f13-5d57-26c0-2ab6f1d48eaa, 'name': SearchDatastore_Task, 'duration_secs': 0.0099} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.686896] env[61957]: DEBUG oslo_concurrency.lockutils [None req-7678a123-82e2-420d-9405-a713c30b30cb tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 952.687149] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-7678a123-82e2-420d-9405-a713c30b30cb tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2] Processing image 11c76a2c-f705-470a-ba9d-4657858bab38 {{(pid=61957) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 952.687390] env[61957]: DEBUG oslo_concurrency.lockutils [None req-7678a123-82e2-420d-9405-a713c30b30cb tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 952.687553] env[61957]: DEBUG oslo_concurrency.lockutils [None req-7678a123-82e2-420d-9405-a713c30b30cb tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 952.687731] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-7678a123-82e2-420d-9405-a713c30b30cb tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 952.687998] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ed76d5ba-a31b-4f92-b8c7-cd8e89d04467 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.696900] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-7678a123-82e2-420d-9405-a713c30b30cb tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 952.696900] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-7678a123-82e2-420d-9405-a713c30b30cb tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61957) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 952.697579] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-114c25d1-1a97-4dd6-8385-52b037c7df35 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.704744] env[61957]: DEBUG oslo_vmware.api [None req-7678a123-82e2-420d-9405-a713c30b30cb tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Waiting for the task: (returnval){ [ 952.704744] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]5208c22f-368c-5958-c2a0-3745a0e36cd8" [ 952.704744] env[61957]: _type = "Task" [ 952.704744] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.712646] env[61957]: DEBUG oslo_vmware.api [None req-7678a123-82e2-420d-9405-a713c30b30cb tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]5208c22f-368c-5958-c2a0-3745a0e36cd8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.732728] env[61957]: DEBUG nova.compute.manager [None req-cedd08d9-20a0-42aa-9f1f-33859d216b7f tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 98929ed3-d420-45cd-9cde-9738b2ac8251] Starting instance... {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 952.858621] env[61957]: DEBUG oslo_vmware.api [None req-d1519ec7-d973-47ee-af77-990ab0a419d4 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Task: {'id': task-1277806, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06872} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.858899] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-d1519ec7-d973-47ee-af77-990ab0a419d4 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] [instance: 7793aa07-6aa2-459d-8a91-56c6b0412d3b] Extended root virtual disk {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 952.859820] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d1edc5b-4f1e-434d-b80a-17406b6d60af {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.882964] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-d1519ec7-d973-47ee-af77-990ab0a419d4 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] [instance: 7793aa07-6aa2-459d-8a91-56c6b0412d3b] Reconfiguring VM instance instance-00000054 to attach disk [datastore2] 7793aa07-6aa2-459d-8a91-56c6b0412d3b/7793aa07-6aa2-459d-8a91-56c6b0412d3b.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 952.883339] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-351d3d0d-f844-416c-b6e1-ec9dac077fc3 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.903814] env[61957]: DEBUG oslo_vmware.api [None req-d1519ec7-d973-47ee-af77-990ab0a419d4 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Waiting for the task: (returnval){ [ 952.903814] env[61957]: value = "task-1277809" [ 952.903814] env[61957]: _type = "Task" [ 952.903814] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.913931] env[61957]: DEBUG oslo_vmware.api [None req-d1519ec7-d973-47ee-af77-990ab0a419d4 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Task: {'id': task-1277809, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.055252] env[61957]: DEBUG oslo_vmware.api [None req-23521104-b947-45c5-a62a-77ad796d9410 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': task-1277807, 'name': PowerOnVM_Task} progress is 90%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.085368] env[61957]: DEBUG oslo_concurrency.lockutils [None req-78652638-0a34-4302-a770-2b339b434047 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.948s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 953.088672] env[61957]: DEBUG oslo_concurrency.lockutils [None req-483cfd5f-2659-4d70-aee0-52ce6ba7fa9a tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 11.308s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 953.088945] env[61957]: DEBUG nova.objects.instance [None req-483cfd5f-2659-4d70-aee0-52ce6ba7fa9a tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Lazy-loading 'resources' on Instance uuid 0d0f2d34-de35-4e80-8d9f-12693add0786 {{(pid=61957) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 953.103302] env[61957]: INFO nova.scheduler.client.report [None req-78652638-0a34-4302-a770-2b339b434047 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Deleted allocations for instance 28ab7d23-487f-4ae1-8fe4-58db55b59918 [ 953.107550] env[61957]: DEBUG oslo_vmware.api [None req-3141e8c0-2e34-4c8b-a8ac-0f12d351422c tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Task: {'id': task-1277808, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.162498] env[61957]: DEBUG nova.network.neutron [-] [instance: 50e09db7-9f8f-452b-8232-bd7473e9e63a] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 953.166217] env[61957]: DEBUG nova.network.neutron [-] [instance: 2e32aae5-edd7-402f-98ad-75a93d26f7a1] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 953.216658] env[61957]: DEBUG oslo_vmware.api [None req-7678a123-82e2-420d-9405-a713c30b30cb tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]5208c22f-368c-5958-c2a0-3745a0e36cd8, 'name': SearchDatastore_Task, 'duration_secs': 0.008869} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.217530] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3c7dff2a-4849-4347-8fc9-3393ea0ed33a {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.222689] env[61957]: DEBUG oslo_vmware.api [None req-7678a123-82e2-420d-9405-a713c30b30cb tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Waiting for the task: (returnval){ [ 953.222689] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52314083-1ebb-8048-db47-28007ba55e54" [ 953.222689] env[61957]: _type = "Task" [ 953.222689] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.230325] env[61957]: DEBUG oslo_vmware.api [None req-7678a123-82e2-420d-9405-a713c30b30cb tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52314083-1ebb-8048-db47-28007ba55e54, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.251309] env[61957]: DEBUG oslo_concurrency.lockutils [None req-cedd08d9-20a0-42aa-9f1f-33859d216b7f tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 953.384376] env[61957]: DEBUG oslo_concurrency.lockutils [None req-5145e111-afa1-43c9-8275-d20aceeb3d2b tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Acquiring lock "6f45c7c8-de07-4d21-bccc-e4afe3c8ad3d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 953.384616] env[61957]: DEBUG oslo_concurrency.lockutils [None req-5145e111-afa1-43c9-8275-d20aceeb3d2b tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Lock "6f45c7c8-de07-4d21-bccc-e4afe3c8ad3d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 953.414538] env[61957]: DEBUG oslo_vmware.api [None req-d1519ec7-d973-47ee-af77-990ab0a419d4 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Task: {'id': task-1277809, 'name': ReconfigVM_Task, 'duration_secs': 0.264216} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.414826] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-d1519ec7-d973-47ee-af77-990ab0a419d4 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] [instance: 7793aa07-6aa2-459d-8a91-56c6b0412d3b] Reconfigured VM instance instance-00000054 to attach disk [datastore2] 7793aa07-6aa2-459d-8a91-56c6b0412d3b/7793aa07-6aa2-459d-8a91-56c6b0412d3b.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 953.415473] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-bb87b97d-9ebf-46d7-a3f8-314c98727f20 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.422540] env[61957]: DEBUG oslo_vmware.api [None req-d1519ec7-d973-47ee-af77-990ab0a419d4 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Waiting for the task: (returnval){ [ 953.422540] env[61957]: value = "task-1277810" [ 953.422540] env[61957]: _type = "Task" [ 953.422540] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.429973] env[61957]: DEBUG oslo_vmware.api [None req-d1519ec7-d973-47ee-af77-990ab0a419d4 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Task: {'id': task-1277810, 'name': Rename_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.509591] env[61957]: DEBUG nova.compute.manager [req-fd611ac9-57da-4948-8730-417d29e866ee req-627b5162-7e8b-4e2e-96aa-7d867f3ec376 service nova] [instance: 50e09db7-9f8f-452b-8232-bd7473e9e63a] Received event network-vif-deleted-224ed13c-938e-4dcf-b555-22fd97cd0210 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 953.509787] env[61957]: DEBUG nova.compute.manager [req-fd611ac9-57da-4948-8730-417d29e866ee req-627b5162-7e8b-4e2e-96aa-7d867f3ec376 service nova] [instance: 2e32aae5-edd7-402f-98ad-75a93d26f7a1] Received event network-vif-deleted-a142b42a-efff-4834-a236-1d0ceb416178 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 953.554956] env[61957]: DEBUG oslo_vmware.api [None req-23521104-b947-45c5-a62a-77ad796d9410 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': task-1277807, 'name': PowerOnVM_Task, 'duration_secs': 0.701474} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.555947] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-23521104-b947-45c5-a62a-77ad796d9410 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: dbeeb200-70b9-4cb4-b5a4-182389d21918] Powered on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 953.558507] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-23521104-b947-45c5-a62a-77ad796d9410 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: dbeeb200-70b9-4cb4-b5a4-182389d21918] Updating instance 'dbeeb200-70b9-4cb4-b5a4-182389d21918' progress to 100 {{(pid=61957) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 953.606680] env[61957]: DEBUG oslo_vmware.api [None req-3141e8c0-2e34-4c8b-a8ac-0f12d351422c tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Task: {'id': task-1277808, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.615803] env[61957]: DEBUG oslo_concurrency.lockutils [None req-78652638-0a34-4302-a770-2b339b434047 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Lock "28ab7d23-487f-4ae1-8fe4-58db55b59918" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 15.662s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 953.664463] env[61957]: INFO nova.compute.manager [-] [instance: 50e09db7-9f8f-452b-8232-bd7473e9e63a] Took 1.44 seconds to deallocate network for instance. [ 953.668317] env[61957]: INFO nova.compute.manager [-] [instance: 2e32aae5-edd7-402f-98ad-75a93d26f7a1] Took 1.42 seconds to deallocate network for instance. [ 953.734271] env[61957]: DEBUG oslo_vmware.api [None req-7678a123-82e2-420d-9405-a713c30b30cb tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52314083-1ebb-8048-db47-28007ba55e54, 'name': SearchDatastore_Task, 'duration_secs': 0.009613} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.736665] env[61957]: DEBUG oslo_concurrency.lockutils [None req-7678a123-82e2-420d-9405-a713c30b30cb tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 953.736872] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-7678a123-82e2-420d-9405-a713c30b30cb tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore2] 08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2/08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 953.737322] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-fe568809-4c6e-4b05-8748-4a1009d0e3f7 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.744787] env[61957]: DEBUG oslo_vmware.api [None req-7678a123-82e2-420d-9405-a713c30b30cb tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Waiting for the task: (returnval){ [ 953.744787] env[61957]: value = "task-1277811" [ 953.744787] env[61957]: _type = "Task" [ 953.744787] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.754897] env[61957]: DEBUG oslo_vmware.api [None req-7678a123-82e2-420d-9405-a713c30b30cb tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': task-1277811, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.840905] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddb3d58f-9003-4c2b-8eef-501cb9e08b4b {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.849467] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae467bc1-8b2c-4c86-a356-4068bf8b556c {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.880307] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3d0eaba-61d0-43b1-a949-a7a5a175cb3a {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.887757] env[61957]: DEBUG nova.compute.manager [None req-5145e111-afa1-43c9-8275-d20aceeb3d2b tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 6f45c7c8-de07-4d21-bccc-e4afe3c8ad3d] Starting instance... {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 953.891647] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fc0db04-663e-4641-9419-786ab2aba185 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.908949] env[61957]: DEBUG nova.compute.provider_tree [None req-483cfd5f-2659-4d70-aee0-52ce6ba7fa9a tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 953.933700] env[61957]: DEBUG oslo_vmware.api [None req-d1519ec7-d973-47ee-af77-990ab0a419d4 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Task: {'id': task-1277810, 'name': Rename_Task, 'duration_secs': 0.131477} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.933999] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-d1519ec7-d973-47ee-af77-990ab0a419d4 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] [instance: 7793aa07-6aa2-459d-8a91-56c6b0412d3b] Powering on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 953.935384] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-bae01795-c9bc-4230-a15c-c58b0a31a778 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.942733] env[61957]: DEBUG oslo_vmware.api [None req-d1519ec7-d973-47ee-af77-990ab0a419d4 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Waiting for the task: (returnval){ [ 953.942733] env[61957]: value = "task-1277812" [ 953.942733] env[61957]: _type = "Task" [ 953.942733] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.952405] env[61957]: DEBUG oslo_vmware.api [None req-d1519ec7-d973-47ee-af77-990ab0a419d4 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Task: {'id': task-1277812, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.113268] env[61957]: DEBUG oslo_vmware.api [None req-3141e8c0-2e34-4c8b-a8ac-0f12d351422c tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Task: {'id': task-1277808, 'name': CreateSnapshot_Task, 'duration_secs': 1.219367} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.113983] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-3141e8c0-2e34-4c8b-a8ac-0f12d351422c tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] [instance: 736eabcf-70f8-4e1e-9830-270d9971fc9a] Created Snapshot of the VM instance {{(pid=61957) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 954.114816] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e276ae09-7c36-4497-8fbd-8be175fb4dd5 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.173392] env[61957]: DEBUG oslo_concurrency.lockutils [None req-5dd00032-4725-4b73-a1a7-7ef95ae8be93 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 954.178898] env[61957]: DEBUG oslo_concurrency.lockutils [None req-7c0e7a82-e685-47c5-90d8-2de75a11cf75 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 954.255053] env[61957]: DEBUG oslo_vmware.api [None req-7678a123-82e2-420d-9405-a713c30b30cb tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': task-1277811, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.457406} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.255492] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-7678a123-82e2-420d-9405-a713c30b30cb tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore2] 08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2/08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 954.255827] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-7678a123-82e2-420d-9405-a713c30b30cb tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2] Extending root virtual disk to 1048576 {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 954.256162] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a331f4ec-884d-4fc4-a8b5-fc68edbf7697 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.265482] env[61957]: DEBUG oslo_vmware.api [None req-7678a123-82e2-420d-9405-a713c30b30cb tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Waiting for the task: (returnval){ [ 954.265482] env[61957]: value = "task-1277813" [ 954.265482] env[61957]: _type = "Task" [ 954.265482] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.274632] env[61957]: DEBUG oslo_vmware.api [None req-7678a123-82e2-420d-9405-a713c30b30cb tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': task-1277813, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.413743] env[61957]: DEBUG nova.scheduler.client.report [None req-483cfd5f-2659-4d70-aee0-52ce6ba7fa9a tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 954.419406] env[61957]: DEBUG oslo_concurrency.lockutils [None req-5145e111-afa1-43c9-8275-d20aceeb3d2b tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 954.455187] env[61957]: DEBUG oslo_vmware.api [None req-d1519ec7-d973-47ee-af77-990ab0a419d4 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Task: {'id': task-1277812, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.636101] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-3141e8c0-2e34-4c8b-a8ac-0f12d351422c tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] [instance: 736eabcf-70f8-4e1e-9830-270d9971fc9a] Creating linked-clone VM from snapshot {{(pid=61957) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 954.636514] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-f9c6e4b0-4b29-43c2-b93f-f5b664e7b87c {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.645310] env[61957]: DEBUG oslo_vmware.api [None req-3141e8c0-2e34-4c8b-a8ac-0f12d351422c tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Waiting for the task: (returnval){ [ 954.645310] env[61957]: value = "task-1277814" [ 954.645310] env[61957]: _type = "Task" [ 954.645310] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.653647] env[61957]: DEBUG oslo_vmware.api [None req-3141e8c0-2e34-4c8b-a8ac-0f12d351422c tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Task: {'id': task-1277814, 'name': CloneVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.763835] env[61957]: DEBUG oslo_concurrency.lockutils [None req-7a8537d3-1c8b-4082-adcd-02831c20c149 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Acquiring lock "e40fa112-4648-428f-a403-b3e3b8319ea8" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 954.764200] env[61957]: DEBUG oslo_concurrency.lockutils [None req-7a8537d3-1c8b-4082-adcd-02831c20c149 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Lock "e40fa112-4648-428f-a403-b3e3b8319ea8" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 954.764526] env[61957]: DEBUG oslo_concurrency.lockutils [None req-7a8537d3-1c8b-4082-adcd-02831c20c149 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Acquiring lock "e40fa112-4648-428f-a403-b3e3b8319ea8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 954.764744] env[61957]: DEBUG oslo_concurrency.lockutils [None req-7a8537d3-1c8b-4082-adcd-02831c20c149 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Lock "e40fa112-4648-428f-a403-b3e3b8319ea8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 954.764962] env[61957]: DEBUG oslo_concurrency.lockutils [None req-7a8537d3-1c8b-4082-adcd-02831c20c149 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Lock "e40fa112-4648-428f-a403-b3e3b8319ea8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 954.771151] env[61957]: INFO nova.compute.manager [None req-7a8537d3-1c8b-4082-adcd-02831c20c149 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] [instance: e40fa112-4648-428f-a403-b3e3b8319ea8] Terminating instance [ 954.773205] env[61957]: DEBUG nova.compute.manager [None req-7a8537d3-1c8b-4082-adcd-02831c20c149 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] [instance: e40fa112-4648-428f-a403-b3e3b8319ea8] Start destroying the instance on the hypervisor. {{(pid=61957) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 954.773408] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-7a8537d3-1c8b-4082-adcd-02831c20c149 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] [instance: e40fa112-4648-428f-a403-b3e3b8319ea8] Destroying instance {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 954.774342] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-939b65a2-585d-4fd2-83e3-ce6d5b94c24c {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.782268] env[61957]: DEBUG oslo_vmware.api [None req-7678a123-82e2-420d-9405-a713c30b30cb tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': task-1277813, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.0619} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.784506] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-7678a123-82e2-420d-9405-a713c30b30cb tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2] Extended root virtual disk {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 954.785188] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-7a8537d3-1c8b-4082-adcd-02831c20c149 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] [instance: e40fa112-4648-428f-a403-b3e3b8319ea8] Powering off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 954.785924] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89562606-1631-4757-a4cf-4774c7f6ff8d {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.788425] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5048cdd1-6808-401b-9f10-e9b2cbfdf0db {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.809688] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-7678a123-82e2-420d-9405-a713c30b30cb tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2] Reconfiguring VM instance instance-00000055 to attach disk [datastore2] 08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2/08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 954.811316] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-08607ec4-19fd-4626-bfc1-0d085ed05d64 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.826355] env[61957]: DEBUG oslo_vmware.api [None req-7a8537d3-1c8b-4082-adcd-02831c20c149 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Waiting for the task: (returnval){ [ 954.826355] env[61957]: value = "task-1277815" [ 954.826355] env[61957]: _type = "Task" [ 954.826355] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.834133] env[61957]: DEBUG oslo_vmware.api [None req-7678a123-82e2-420d-9405-a713c30b30cb tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Waiting for the task: (returnval){ [ 954.834133] env[61957]: value = "task-1277816" [ 954.834133] env[61957]: _type = "Task" [ 954.834133] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.837384] env[61957]: DEBUG oslo_vmware.api [None req-7a8537d3-1c8b-4082-adcd-02831c20c149 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Task: {'id': task-1277815, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.846203] env[61957]: DEBUG oslo_vmware.api [None req-7678a123-82e2-420d-9405-a713c30b30cb tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': task-1277816, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.921238] env[61957]: DEBUG oslo_concurrency.lockutils [None req-483cfd5f-2659-4d70-aee0-52ce6ba7fa9a tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.833s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 954.923937] env[61957]: DEBUG oslo_concurrency.lockutils [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 13.043s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 954.923937] env[61957]: DEBUG oslo_concurrency.lockutils [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 954.924120] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61957) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 954.924333] env[61957]: DEBUG oslo_concurrency.lockutils [None req-680ab45d-0a61-469d-82c9-fe13d79ebbb0 tempest-InstanceActionsNegativeTestJSON-259524954 tempest-InstanceActionsNegativeTestJSON-259524954-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 12.889s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 954.924532] env[61957]: DEBUG nova.objects.instance [None req-680ab45d-0a61-469d-82c9-fe13d79ebbb0 tempest-InstanceActionsNegativeTestJSON-259524954 tempest-InstanceActionsNegativeTestJSON-259524954-project-member] Lazy-loading 'resources' on Instance uuid 7e2db5f6-1419-415e-b957-66cf2e3ec19d {{(pid=61957) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 954.926302] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5bfbbf4-4f99-4072-af14-54646f64eb40 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.935614] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48593191-6839-4ba5-a458-f55d147cf25f {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.941038] env[61957]: INFO nova.scheduler.client.report [None req-483cfd5f-2659-4d70-aee0-52ce6ba7fa9a tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Deleted allocations for instance 0d0f2d34-de35-4e80-8d9f-12693add0786 [ 954.960979] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8970bd1-4d1f-49e5-818b-4bc0f53d8a57 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.968444] env[61957]: DEBUG oslo_vmware.api [None req-d1519ec7-d973-47ee-af77-990ab0a419d4 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Task: {'id': task-1277812, 'name': PowerOnVM_Task, 'duration_secs': 0.5205} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.969317] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-d1519ec7-d973-47ee-af77-990ab0a419d4 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] [instance: 7793aa07-6aa2-459d-8a91-56c6b0412d3b] Powered on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 954.969584] env[61957]: INFO nova.compute.manager [None req-d1519ec7-d973-47ee-af77-990ab0a419d4 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] [instance: 7793aa07-6aa2-459d-8a91-56c6b0412d3b] Took 7.89 seconds to spawn the instance on the hypervisor. [ 954.969801] env[61957]: DEBUG nova.compute.manager [None req-d1519ec7-d973-47ee-af77-990ab0a419d4 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] [instance: 7793aa07-6aa2-459d-8a91-56c6b0412d3b] Checking state {{(pid=61957) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 954.970765] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3877b48-46c6-448a-9622-329cb1b3754e {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.977468] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78e69013-bbd0-45c3-8c7a-667ec4768bc5 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.014575] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180597MB free_disk=141GB free_vcpus=48 pci_devices=None {{(pid=61957) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 955.015656] env[61957]: DEBUG oslo_concurrency.lockutils [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 955.158247] env[61957]: DEBUG oslo_vmware.api [None req-3141e8c0-2e34-4c8b-a8ac-0f12d351422c tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Task: {'id': task-1277814, 'name': CloneVM_Task} progress is 94%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.336139] env[61957]: DEBUG oslo_vmware.api [None req-7a8537d3-1c8b-4082-adcd-02831c20c149 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Task: {'id': task-1277815, 'name': PowerOffVM_Task, 'duration_secs': 0.244994} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.336496] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-7a8537d3-1c8b-4082-adcd-02831c20c149 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] [instance: e40fa112-4648-428f-a403-b3e3b8319ea8] Powered off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 955.336713] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-7a8537d3-1c8b-4082-adcd-02831c20c149 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] [instance: e40fa112-4648-428f-a403-b3e3b8319ea8] Unregistering the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 955.336999] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-08949c79-f4d0-4c25-b9e7-756ff31c9d25 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.346779] env[61957]: DEBUG oslo_vmware.api [None req-7678a123-82e2-420d-9405-a713c30b30cb tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': task-1277816, 'name': ReconfigVM_Task, 'duration_secs': 0.390466} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.347136] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-7678a123-82e2-420d-9405-a713c30b30cb tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2] Reconfigured VM instance instance-00000055 to attach disk [datastore2] 08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2/08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 955.347804] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-02a9f8ff-e8b5-47aa-b6db-034a84ac8807 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.354150] env[61957]: DEBUG oslo_vmware.api [None req-7678a123-82e2-420d-9405-a713c30b30cb tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Waiting for the task: (returnval){ [ 955.354150] env[61957]: value = "task-1277818" [ 955.354150] env[61957]: _type = "Task" [ 955.354150] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.361796] env[61957]: DEBUG oslo_vmware.api [None req-7678a123-82e2-420d-9405-a713c30b30cb tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': task-1277818, 'name': Rename_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.409887] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-7a8537d3-1c8b-4082-adcd-02831c20c149 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] [instance: e40fa112-4648-428f-a403-b3e3b8319ea8] Unregistered the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 955.410065] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-7a8537d3-1c8b-4082-adcd-02831c20c149 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] [instance: e40fa112-4648-428f-a403-b3e3b8319ea8] Deleting contents of the VM from datastore datastore2 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 955.410200] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-7a8537d3-1c8b-4082-adcd-02831c20c149 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Deleting the datastore file [datastore2] e40fa112-4648-428f-a403-b3e3b8319ea8 {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 955.411069] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-fee256a1-6925-421a-921b-1b53fb98a101 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.418276] env[61957]: DEBUG oslo_vmware.api [None req-7a8537d3-1c8b-4082-adcd-02831c20c149 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Waiting for the task: (returnval){ [ 955.418276] env[61957]: value = "task-1277819" [ 955.418276] env[61957]: _type = "Task" [ 955.418276] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.427190] env[61957]: DEBUG oslo_vmware.api [None req-7a8537d3-1c8b-4082-adcd-02831c20c149 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Task: {'id': task-1277819, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.450613] env[61957]: DEBUG oslo_concurrency.lockutils [None req-483cfd5f-2659-4d70-aee0-52ce6ba7fa9a tempest-ServerShowV247Test-393051144 tempest-ServerShowV247Test-393051144-project-member] Lock "0d0f2d34-de35-4e80-8d9f-12693add0786" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 17.493s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 955.496142] env[61957]: INFO nova.compute.manager [None req-d1519ec7-d973-47ee-af77-990ab0a419d4 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] [instance: 7793aa07-6aa2-459d-8a91-56c6b0412d3b] Took 28.24 seconds to build instance. [ 955.658957] env[61957]: DEBUG oslo_vmware.api [None req-3141e8c0-2e34-4c8b-a8ac-0f12d351422c tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Task: {'id': task-1277814, 'name': CloneVM_Task} progress is 94%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.690111] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e45875df-a662-4f28-854c-d5f2913f9fa0 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.697155] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad86ef51-2061-486a-879f-e81d78b3323e {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.729554] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a631ce2-a2aa-4fc2-a317-f5291feb9b73 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.737129] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5cac904-40aa-48e2-9fe9-c2eb3d0c7117 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.750280] env[61957]: DEBUG nova.compute.provider_tree [None req-680ab45d-0a61-469d-82c9-fe13d79ebbb0 tempest-InstanceActionsNegativeTestJSON-259524954 tempest-InstanceActionsNegativeTestJSON-259524954-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 955.864197] env[61957]: DEBUG oslo_vmware.api [None req-7678a123-82e2-420d-9405-a713c30b30cb tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': task-1277818, 'name': Rename_Task, 'duration_secs': 0.176065} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.864515] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-7678a123-82e2-420d-9405-a713c30b30cb tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2] Powering on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 955.864762] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-bcd8f39c-dc6d-4bc0-920e-5b9f8295635d {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.871685] env[61957]: DEBUG oslo_vmware.api [None req-7678a123-82e2-420d-9405-a713c30b30cb tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Waiting for the task: (returnval){ [ 955.871685] env[61957]: value = "task-1277820" [ 955.871685] env[61957]: _type = "Task" [ 955.871685] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.879912] env[61957]: DEBUG oslo_vmware.api [None req-7678a123-82e2-420d-9405-a713c30b30cb tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': task-1277820, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.900223] env[61957]: DEBUG oslo_concurrency.lockutils [None req-19f7f8b2-5996-490c-9f47-41f02815c525 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Acquiring lock "dbeeb200-70b9-4cb4-b5a4-182389d21918" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 955.900509] env[61957]: DEBUG oslo_concurrency.lockutils [None req-19f7f8b2-5996-490c-9f47-41f02815c525 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Lock "dbeeb200-70b9-4cb4-b5a4-182389d21918" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 955.900705] env[61957]: DEBUG nova.compute.manager [None req-19f7f8b2-5996-490c-9f47-41f02815c525 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: dbeeb200-70b9-4cb4-b5a4-182389d21918] Going to confirm migration 1 {{(pid=61957) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 955.927942] env[61957]: DEBUG oslo_vmware.api [None req-7a8537d3-1c8b-4082-adcd-02831c20c149 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Task: {'id': task-1277819, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.169255} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.928089] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-7a8537d3-1c8b-4082-adcd-02831c20c149 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Deleted the datastore file {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 955.928267] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-7a8537d3-1c8b-4082-adcd-02831c20c149 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] [instance: e40fa112-4648-428f-a403-b3e3b8319ea8] Deleted contents of the VM from datastore datastore2 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 955.928459] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-7a8537d3-1c8b-4082-adcd-02831c20c149 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] [instance: e40fa112-4648-428f-a403-b3e3b8319ea8] Instance destroyed {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 955.928636] env[61957]: INFO nova.compute.manager [None req-7a8537d3-1c8b-4082-adcd-02831c20c149 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] [instance: e40fa112-4648-428f-a403-b3e3b8319ea8] Took 1.16 seconds to destroy the instance on the hypervisor. [ 955.928887] env[61957]: DEBUG oslo.service.loopingcall [None req-7a8537d3-1c8b-4082-adcd-02831c20c149 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 955.929153] env[61957]: DEBUG nova.compute.manager [-] [instance: e40fa112-4648-428f-a403-b3e3b8319ea8] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 955.929223] env[61957]: DEBUG nova.network.neutron [-] [instance: e40fa112-4648-428f-a403-b3e3b8319ea8] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 956.001322] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d1519ec7-d973-47ee-af77-990ab0a419d4 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Lock "7793aa07-6aa2-459d-8a91-56c6b0412d3b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 29.753s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 956.157494] env[61957]: DEBUG oslo_vmware.api [None req-3141e8c0-2e34-4c8b-a8ac-0f12d351422c tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Task: {'id': task-1277814, 'name': CloneVM_Task} progress is 95%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.228597] env[61957]: DEBUG nova.compute.manager [req-e1eb0121-664c-4aa2-9d23-bbb78c03ed18 req-771152d3-860d-4487-ab7d-3d0433d8d5be service nova] [instance: e40fa112-4648-428f-a403-b3e3b8319ea8] Received event network-vif-deleted-c407d755-4caa-44d0-b9b2-8e59106cc23e {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 956.229258] env[61957]: INFO nova.compute.manager [req-e1eb0121-664c-4aa2-9d23-bbb78c03ed18 req-771152d3-860d-4487-ab7d-3d0433d8d5be service nova] [instance: e40fa112-4648-428f-a403-b3e3b8319ea8] Neutron deleted interface c407d755-4caa-44d0-b9b2-8e59106cc23e; detaching it from the instance and deleting it from the info cache [ 956.229258] env[61957]: DEBUG nova.network.neutron [req-e1eb0121-664c-4aa2-9d23-bbb78c03ed18 req-771152d3-860d-4487-ab7d-3d0433d8d5be service nova] [instance: e40fa112-4648-428f-a403-b3e3b8319ea8] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 956.253803] env[61957]: DEBUG nova.scheduler.client.report [None req-680ab45d-0a61-469d-82c9-fe13d79ebbb0 tempest-InstanceActionsNegativeTestJSON-259524954 tempest-InstanceActionsNegativeTestJSON-259524954-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 956.383943] env[61957]: DEBUG oslo_vmware.api [None req-7678a123-82e2-420d-9405-a713c30b30cb tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': task-1277820, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.511826] env[61957]: DEBUG oslo_concurrency.lockutils [None req-19f7f8b2-5996-490c-9f47-41f02815c525 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Acquiring lock "refresh_cache-dbeeb200-70b9-4cb4-b5a4-182389d21918" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 956.512037] env[61957]: DEBUG oslo_concurrency.lockutils [None req-19f7f8b2-5996-490c-9f47-41f02815c525 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Acquired lock "refresh_cache-dbeeb200-70b9-4cb4-b5a4-182389d21918" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 956.512243] env[61957]: DEBUG nova.network.neutron [None req-19f7f8b2-5996-490c-9f47-41f02815c525 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: dbeeb200-70b9-4cb4-b5a4-182389d21918] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 956.512456] env[61957]: DEBUG nova.objects.instance [None req-19f7f8b2-5996-490c-9f47-41f02815c525 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Lazy-loading 'info_cache' on Instance uuid dbeeb200-70b9-4cb4-b5a4-182389d21918 {{(pid=61957) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 956.585188] env[61957]: INFO nova.compute.manager [None req-7f4318a7-7ec4-4da1-9bbe-cec59770ca17 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] [instance: 7793aa07-6aa2-459d-8a91-56c6b0412d3b] Rescuing [ 956.585479] env[61957]: DEBUG oslo_concurrency.lockutils [None req-7f4318a7-7ec4-4da1-9bbe-cec59770ca17 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Acquiring lock "refresh_cache-7793aa07-6aa2-459d-8a91-56c6b0412d3b" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 956.585636] env[61957]: DEBUG oslo_concurrency.lockutils [None req-7f4318a7-7ec4-4da1-9bbe-cec59770ca17 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Acquired lock "refresh_cache-7793aa07-6aa2-459d-8a91-56c6b0412d3b" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 956.585805] env[61957]: DEBUG nova.network.neutron [None req-7f4318a7-7ec4-4da1-9bbe-cec59770ca17 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] [instance: 7793aa07-6aa2-459d-8a91-56c6b0412d3b] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 956.657764] env[61957]: DEBUG oslo_vmware.api [None req-3141e8c0-2e34-4c8b-a8ac-0f12d351422c tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Task: {'id': task-1277814, 'name': CloneVM_Task, 'duration_secs': 1.820169} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.658076] env[61957]: INFO nova.virt.vmwareapi.vmops [None req-3141e8c0-2e34-4c8b-a8ac-0f12d351422c tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] [instance: 736eabcf-70f8-4e1e-9830-270d9971fc9a] Created linked-clone VM from snapshot [ 956.658846] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-843733ad-eb4a-4484-b8e9-828c1d98877d {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.666498] env[61957]: DEBUG nova.virt.vmwareapi.images [None req-3141e8c0-2e34-4c8b-a8ac-0f12d351422c tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] [instance: 736eabcf-70f8-4e1e-9830-270d9971fc9a] Uploading image 0018d8e9-1d6e-4ca4-a224-31fc9e9f9d2e {{(pid=61957) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 956.691354] env[61957]: DEBUG oslo_vmware.rw_handles [None req-3141e8c0-2e34-4c8b-a8ac-0f12d351422c tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 956.691354] env[61957]: value = "vm-274573" [ 956.691354] env[61957]: _type = "VirtualMachine" [ 956.691354] env[61957]: }. {{(pid=61957) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 956.691662] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-327e7f26-9e8f-4882-bb49-1ffd1a8017e6 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.693686] env[61957]: DEBUG nova.network.neutron [-] [instance: e40fa112-4648-428f-a403-b3e3b8319ea8] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 956.698484] env[61957]: DEBUG oslo_vmware.rw_handles [None req-3141e8c0-2e34-4c8b-a8ac-0f12d351422c tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Lease: (returnval){ [ 956.698484] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52729d97-747b-e25b-cafa-96e062488aee" [ 956.698484] env[61957]: _type = "HttpNfcLease" [ 956.698484] env[61957]: } obtained for exporting VM: (result){ [ 956.698484] env[61957]: value = "vm-274573" [ 956.698484] env[61957]: _type = "VirtualMachine" [ 956.698484] env[61957]: }. {{(pid=61957) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 956.698768] env[61957]: DEBUG oslo_vmware.api [None req-3141e8c0-2e34-4c8b-a8ac-0f12d351422c tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Waiting for the lease: (returnval){ [ 956.698768] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52729d97-747b-e25b-cafa-96e062488aee" [ 956.698768] env[61957]: _type = "HttpNfcLease" [ 956.698768] env[61957]: } to be ready. {{(pid=61957) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 956.705023] env[61957]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 956.705023] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52729d97-747b-e25b-cafa-96e062488aee" [ 956.705023] env[61957]: _type = "HttpNfcLease" [ 956.705023] env[61957]: } is initializing. {{(pid=61957) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 956.732816] env[61957]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-394afd4a-0f59-4503-8fac-fe8f9b6ad8eb {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.741021] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02146190-bc4a-4136-bb6b-89b8a8583c77 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.769487] env[61957]: DEBUG oslo_concurrency.lockutils [None req-680ab45d-0a61-469d-82c9-fe13d79ebbb0 tempest-InstanceActionsNegativeTestJSON-259524954 tempest-InstanceActionsNegativeTestJSON-259524954-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.845s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 956.771580] env[61957]: DEBUG nova.compute.manager [req-e1eb0121-664c-4aa2-9d23-bbb78c03ed18 req-771152d3-860d-4487-ab7d-3d0433d8d5be service nova] [instance: e40fa112-4648-428f-a403-b3e3b8319ea8] Detach interface failed, port_id=c407d755-4caa-44d0-b9b2-8e59106cc23e, reason: Instance e40fa112-4648-428f-a403-b3e3b8319ea8 could not be found. {{(pid=61957) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 956.772191] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d1aa370b-6bd9-49b7-9a71-30e9793a2856 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.132s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 956.773642] env[61957]: INFO nova.compute.claims [None req-d1aa370b-6bd9-49b7-9a71-30e9793a2856 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] [instance: 00ecc8c2-398e-4a29-b19f-a2013985b481] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 956.794095] env[61957]: INFO nova.scheduler.client.report [None req-680ab45d-0a61-469d-82c9-fe13d79ebbb0 tempest-InstanceActionsNegativeTestJSON-259524954 tempest-InstanceActionsNegativeTestJSON-259524954-project-member] Deleted allocations for instance 7e2db5f6-1419-415e-b957-66cf2e3ec19d [ 956.882787] env[61957]: DEBUG oslo_vmware.api [None req-7678a123-82e2-420d-9405-a713c30b30cb tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': task-1277820, 'name': PowerOnVM_Task, 'duration_secs': 0.754196} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.883085] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-7678a123-82e2-420d-9405-a713c30b30cb tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2] Powered on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 956.883274] env[61957]: INFO nova.compute.manager [None req-7678a123-82e2-420d-9405-a713c30b30cb tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2] Took 7.41 seconds to spawn the instance on the hypervisor. [ 956.883451] env[61957]: DEBUG nova.compute.manager [None req-7678a123-82e2-420d-9405-a713c30b30cb tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2] Checking state {{(pid=61957) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 956.884298] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64dc7aaf-f1dd-4d62-9ac4-ffc76fd18a42 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.198243] env[61957]: INFO nova.compute.manager [-] [instance: e40fa112-4648-428f-a403-b3e3b8319ea8] Took 1.27 seconds to deallocate network for instance. [ 957.218419] env[61957]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 957.218419] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52729d97-747b-e25b-cafa-96e062488aee" [ 957.218419] env[61957]: _type = "HttpNfcLease" [ 957.218419] env[61957]: } is ready. {{(pid=61957) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 957.218716] env[61957]: DEBUG oslo_vmware.rw_handles [None req-3141e8c0-2e34-4c8b-a8ac-0f12d351422c tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 957.218716] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52729d97-747b-e25b-cafa-96e062488aee" [ 957.218716] env[61957]: _type = "HttpNfcLease" [ 957.218716] env[61957]: }. {{(pid=61957) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 957.219477] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e26fa97c-9ca2-41c0-8c29-6f3f7430588b {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.228444] env[61957]: DEBUG oslo_vmware.rw_handles [None req-3141e8c0-2e34-4c8b-a8ac-0f12d351422c tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52c996eb-d80c-6e44-ead7-48e0e52505b7/disk-0.vmdk from lease info. {{(pid=61957) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 957.228636] env[61957]: DEBUG oslo_vmware.rw_handles [None req-3141e8c0-2e34-4c8b-a8ac-0f12d351422c tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52c996eb-d80c-6e44-ead7-48e0e52505b7/disk-0.vmdk for reading. {{(pid=61957) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 957.304238] env[61957]: DEBUG oslo_concurrency.lockutils [None req-680ab45d-0a61-469d-82c9-fe13d79ebbb0 tempest-InstanceActionsNegativeTestJSON-259524954 tempest-InstanceActionsNegativeTestJSON-259524954-project-member] Lock "7e2db5f6-1419-415e-b957-66cf2e3ec19d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 18.279s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 957.333684] env[61957]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-08273222-2cc2-4922-b084-d088e18f2183 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.400510] env[61957]: INFO nova.compute.manager [None req-7678a123-82e2-420d-9405-a713c30b30cb tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2] Took 27.56 seconds to build instance. [ 957.501612] env[61957]: DEBUG nova.network.neutron [None req-7f4318a7-7ec4-4da1-9bbe-cec59770ca17 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] [instance: 7793aa07-6aa2-459d-8a91-56c6b0412d3b] Updating instance_info_cache with network_info: [{"id": "dfb387b2-224a-48bc-a581-17c9f58b213a", "address": "fa:16:3e:f4:50:7f", "network": {"id": "d55c72c6-107d-4965-b4e1-d18383813953", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-193304765-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "7758bdd10505455b8cecb8129d70ebd5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c24464bb-bb6b-43a2-bdcd-8086ad1a307f", "external-id": "nsx-vlan-transportzone-781", "segmentation_id": 781, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdfb387b2-22", "ovs_interfaceid": "dfb387b2-224a-48bc-a581-17c9f58b213a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 957.715504] env[61957]: DEBUG oslo_concurrency.lockutils [None req-7a8537d3-1c8b-4082-adcd-02831c20c149 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 957.890313] env[61957]: DEBUG nova.network.neutron [None req-19f7f8b2-5996-490c-9f47-41f02815c525 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: dbeeb200-70b9-4cb4-b5a4-182389d21918] Updating instance_info_cache with network_info: [{"id": "86553395-da58-4c44-b9f4-c67db304d3fe", "address": "fa:16:3e:a4:1b:84", "network": {"id": "6fbccc17-6674-4cd5-8f1b-aa73570baac7", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-394544811-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1dbdc699741a48af93c52068d88a5357", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "20e3f794-c7a3-4696-9488-ecf34c570ef9", "external-id": "nsx-vlan-transportzone-509", "segmentation_id": 509, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap86553395-da", "ovs_interfaceid": "86553395-da58-4c44-b9f4-c67db304d3fe", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 957.902785] env[61957]: DEBUG oslo_concurrency.lockutils [None req-7678a123-82e2-420d-9405-a713c30b30cb tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Lock "08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 29.629s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 958.006556] env[61957]: DEBUG oslo_concurrency.lockutils [None req-7f4318a7-7ec4-4da1-9bbe-cec59770ca17 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Releasing lock "refresh_cache-7793aa07-6aa2-459d-8a91-56c6b0412d3b" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 958.109302] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d019ed50-4fef-47fd-a61a-c283f4bb6896 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.118131] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7465945a-58c2-4b07-bb8c-9cc3357d53b6 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.154304] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78e71f14-2878-403b-b70d-d2b2303de1e6 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.162050] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbd817e2-cde0-4c61-a704-a41407d2e1b0 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.177653] env[61957]: DEBUG nova.compute.provider_tree [None req-d1aa370b-6bd9-49b7-9a71-30e9793a2856 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 958.394685] env[61957]: DEBUG oslo_concurrency.lockutils [None req-19f7f8b2-5996-490c-9f47-41f02815c525 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Releasing lock "refresh_cache-dbeeb200-70b9-4cb4-b5a4-182389d21918" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 958.394685] env[61957]: DEBUG nova.objects.instance [None req-19f7f8b2-5996-490c-9f47-41f02815c525 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Lazy-loading 'migration_context' on Instance uuid dbeeb200-70b9-4cb4-b5a4-182389d21918 {{(pid=61957) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 958.545211] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-7f4318a7-7ec4-4da1-9bbe-cec59770ca17 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] [instance: 7793aa07-6aa2-459d-8a91-56c6b0412d3b] Powering off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 958.546768] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c7c58472-9e15-4cc2-9459-8ac9a1c090e4 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.554517] env[61957]: DEBUG oslo_vmware.api [None req-7f4318a7-7ec4-4da1-9bbe-cec59770ca17 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Waiting for the task: (returnval){ [ 958.554517] env[61957]: value = "task-1277822" [ 958.554517] env[61957]: _type = "Task" [ 958.554517] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.569096] env[61957]: DEBUG oslo_vmware.api [None req-7f4318a7-7ec4-4da1-9bbe-cec59770ca17 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Task: {'id': task-1277822, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.680748] env[61957]: DEBUG nova.scheduler.client.report [None req-d1aa370b-6bd9-49b7-9a71-30e9793a2856 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 958.776615] env[61957]: DEBUG nova.compute.manager [req-f1811632-4f29-4a35-9caf-f65319baa9ac req-f261fd67-c6fb-478f-a509-ed9f31d908cf service nova] [instance: 08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2] Received event network-changed-f62a9f12-0b55-4138-a40e-b431855487fe {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 958.776926] env[61957]: DEBUG nova.compute.manager [req-f1811632-4f29-4a35-9caf-f65319baa9ac req-f261fd67-c6fb-478f-a509-ed9f31d908cf service nova] [instance: 08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2] Refreshing instance network info cache due to event network-changed-f62a9f12-0b55-4138-a40e-b431855487fe. {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 958.777272] env[61957]: DEBUG oslo_concurrency.lockutils [req-f1811632-4f29-4a35-9caf-f65319baa9ac req-f261fd67-c6fb-478f-a509-ed9f31d908cf service nova] Acquiring lock "refresh_cache-08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 958.777368] env[61957]: DEBUG oslo_concurrency.lockutils [req-f1811632-4f29-4a35-9caf-f65319baa9ac req-f261fd67-c6fb-478f-a509-ed9f31d908cf service nova] Acquired lock "refresh_cache-08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 958.777579] env[61957]: DEBUG nova.network.neutron [req-f1811632-4f29-4a35-9caf-f65319baa9ac req-f261fd67-c6fb-478f-a509-ed9f31d908cf service nova] [instance: 08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2] Refreshing network info cache for port f62a9f12-0b55-4138-a40e-b431855487fe {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 958.896801] env[61957]: DEBUG nova.objects.base [None req-19f7f8b2-5996-490c-9f47-41f02815c525 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Object Instance lazy-loaded attributes: info_cache,migration_context {{(pid=61957) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 958.897877] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-716eca7a-1f1d-4ced-b42b-c3a3255f9e90 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.927029] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d106ed2d-ff52-4df3-bc6b-4a2fc7083065 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.937089] env[61957]: DEBUG oslo_vmware.api [None req-19f7f8b2-5996-490c-9f47-41f02815c525 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Waiting for the task: (returnval){ [ 958.937089] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]5210a951-cb43-e9fe-4b65-33162c7752f2" [ 958.937089] env[61957]: _type = "Task" [ 958.937089] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.946816] env[61957]: DEBUG oslo_vmware.api [None req-19f7f8b2-5996-490c-9f47-41f02815c525 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]5210a951-cb43-e9fe-4b65-33162c7752f2, 'name': SearchDatastore_Task, 'duration_secs': 0.011252} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.947284] env[61957]: DEBUG oslo_concurrency.lockutils [None req-19f7f8b2-5996-490c-9f47-41f02815c525 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 959.070385] env[61957]: DEBUG oslo_vmware.api [None req-7f4318a7-7ec4-4da1-9bbe-cec59770ca17 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Task: {'id': task-1277822, 'name': PowerOffVM_Task, 'duration_secs': 0.240613} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.070385] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-7f4318a7-7ec4-4da1-9bbe-cec59770ca17 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] [instance: 7793aa07-6aa2-459d-8a91-56c6b0412d3b] Powered off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 959.070385] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b768cd8-9ab7-4524-9f07-bb8a6e1a549d {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.095133] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a68c978b-f98e-4106-b4d9-74df0f9be70f {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.130406] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-7f4318a7-7ec4-4da1-9bbe-cec59770ca17 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] [instance: 7793aa07-6aa2-459d-8a91-56c6b0412d3b] Powering off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 959.130959] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-91a21b44-ab10-43ba-97c7-cb60b5a7289d {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.139265] env[61957]: DEBUG oslo_vmware.api [None req-7f4318a7-7ec4-4da1-9bbe-cec59770ca17 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Waiting for the task: (returnval){ [ 959.139265] env[61957]: value = "task-1277823" [ 959.139265] env[61957]: _type = "Task" [ 959.139265] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.152175] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-7f4318a7-7ec4-4da1-9bbe-cec59770ca17 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] [instance: 7793aa07-6aa2-459d-8a91-56c6b0412d3b] VM already powered off {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 959.152175] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-7f4318a7-7ec4-4da1-9bbe-cec59770ca17 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] [instance: 7793aa07-6aa2-459d-8a91-56c6b0412d3b] Processing image 11c76a2c-f705-470a-ba9d-4657858bab38 {{(pid=61957) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 959.152175] env[61957]: DEBUG oslo_concurrency.lockutils [None req-7f4318a7-7ec4-4da1-9bbe-cec59770ca17 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 959.152175] env[61957]: DEBUG oslo_concurrency.lockutils [None req-7f4318a7-7ec4-4da1-9bbe-cec59770ca17 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 959.155987] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-7f4318a7-7ec4-4da1-9bbe-cec59770ca17 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 959.156793] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ac4696ae-7f2d-46f2-a85b-e3ee5a1e934f {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.170761] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-7f4318a7-7ec4-4da1-9bbe-cec59770ca17 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 959.170761] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-7f4318a7-7ec4-4da1-9bbe-cec59770ca17 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61957) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 959.170761] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ed70e602-23bc-41a6-894e-2162354353da {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.178400] env[61957]: DEBUG oslo_vmware.api [None req-7f4318a7-7ec4-4da1-9bbe-cec59770ca17 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Waiting for the task: (returnval){ [ 959.178400] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]5221f2c3-5675-e0b1-ed7a-a83a0ff91420" [ 959.178400] env[61957]: _type = "Task" [ 959.178400] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.187744] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d1aa370b-6bd9-49b7-9a71-30e9793a2856 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.415s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 959.188449] env[61957]: DEBUG nova.compute.manager [None req-d1aa370b-6bd9-49b7-9a71-30e9793a2856 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] [instance: 00ecc8c2-398e-4a29-b19f-a2013985b481] Start building networks asynchronously for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 959.191791] env[61957]: DEBUG oslo_vmware.api [None req-7f4318a7-7ec4-4da1-9bbe-cec59770ca17 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]5221f2c3-5675-e0b1-ed7a-a83a0ff91420, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.192213] env[61957]: DEBUG oslo_concurrency.lockutils [None req-9e60b82d-4fee-4f83-a3da-b48f03d27cc5 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.884s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 959.197019] env[61957]: INFO nova.compute.claims [None req-9e60b82d-4fee-4f83-a3da-b48f03d27cc5 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: d6c5c70d-86de-4dea-8b9e-76f321947a35] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 959.552954] env[61957]: DEBUG nova.network.neutron [req-f1811632-4f29-4a35-9caf-f65319baa9ac req-f261fd67-c6fb-478f-a509-ed9f31d908cf service nova] [instance: 08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2] Updated VIF entry in instance network info cache for port f62a9f12-0b55-4138-a40e-b431855487fe. {{(pid=61957) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 959.553395] env[61957]: DEBUG nova.network.neutron [req-f1811632-4f29-4a35-9caf-f65319baa9ac req-f261fd67-c6fb-478f-a509-ed9f31d908cf service nova] [instance: 08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2] Updating instance_info_cache with network_info: [{"id": "f62a9f12-0b55-4138-a40e-b431855487fe", "address": "fa:16:3e:af:48:d3", "network": {"id": "364cabcb-046d-4f91-b1ee-5dd3adfae6d7", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1432396967-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.219", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8713aa35bcb24b86ad0b58ca9fc991ba", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b7a73c01-1bb9-4612-a1a7-16d71b732e81", "external-id": "nsx-vlan-transportzone-711", "segmentation_id": 711, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf62a9f12-0b", "ovs_interfaceid": "f62a9f12-0b55-4138-a40e-b431855487fe", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 959.690790] env[61957]: DEBUG oslo_vmware.api [None req-7f4318a7-7ec4-4da1-9bbe-cec59770ca17 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]5221f2c3-5675-e0b1-ed7a-a83a0ff91420, 'name': SearchDatastore_Task, 'duration_secs': 0.015089} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.692236] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6554307b-cd75-44bb-8ea5-0f9fc170a40e {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.699433] env[61957]: DEBUG nova.compute.utils [None req-d1aa370b-6bd9-49b7-9a71-30e9793a2856 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Using /dev/sd instead of None {{(pid=61957) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 959.701203] env[61957]: DEBUG oslo_vmware.api [None req-7f4318a7-7ec4-4da1-9bbe-cec59770ca17 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Waiting for the task: (returnval){ [ 959.701203] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52ae700d-4cc7-9d42-9f7a-a744f828465f" [ 959.701203] env[61957]: _type = "Task" [ 959.701203] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.701782] env[61957]: DEBUG nova.compute.manager [None req-d1aa370b-6bd9-49b7-9a71-30e9793a2856 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] [instance: 00ecc8c2-398e-4a29-b19f-a2013985b481] Allocating IP information in the background. {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 959.701988] env[61957]: DEBUG nova.network.neutron [None req-d1aa370b-6bd9-49b7-9a71-30e9793a2856 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] [instance: 00ecc8c2-398e-4a29-b19f-a2013985b481] allocate_for_instance() {{(pid=61957) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 959.718761] env[61957]: DEBUG oslo_vmware.api [None req-7f4318a7-7ec4-4da1-9bbe-cec59770ca17 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52ae700d-4cc7-9d42-9f7a-a744f828465f, 'name': SearchDatastore_Task, 'duration_secs': 0.012762} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.719103] env[61957]: DEBUG oslo_concurrency.lockutils [None req-7f4318a7-7ec4-4da1-9bbe-cec59770ca17 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 959.719383] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-7f4318a7-7ec4-4da1-9bbe-cec59770ca17 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore2] 7793aa07-6aa2-459d-8a91-56c6b0412d3b/11c76a2c-f705-470a-ba9d-4657858bab38-rescue.vmdk. {{(pid=61957) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 959.719700] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f3117f33-5190-4b23-9b21-6e1e9107cd1f {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.728087] env[61957]: DEBUG oslo_vmware.api [None req-7f4318a7-7ec4-4da1-9bbe-cec59770ca17 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Waiting for the task: (returnval){ [ 959.728087] env[61957]: value = "task-1277824" [ 959.728087] env[61957]: _type = "Task" [ 959.728087] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.737827] env[61957]: DEBUG oslo_vmware.api [None req-7f4318a7-7ec4-4da1-9bbe-cec59770ca17 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Task: {'id': task-1277824, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.755231] env[61957]: DEBUG nova.policy [None req-d1aa370b-6bd9-49b7-9a71-30e9793a2856 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6653838d4b8a42618da8f8dc35362cdc', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f7c20ab3c822423fa837a03f066e4239', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61957) authorize /opt/stack/nova/nova/policy.py:203}} [ 960.058473] env[61957]: DEBUG oslo_concurrency.lockutils [req-f1811632-4f29-4a35-9caf-f65319baa9ac req-f261fd67-c6fb-478f-a509-ed9f31d908cf service nova] Releasing lock "refresh_cache-08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 960.100535] env[61957]: DEBUG nova.network.neutron [None req-d1aa370b-6bd9-49b7-9a71-30e9793a2856 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] [instance: 00ecc8c2-398e-4a29-b19f-a2013985b481] Successfully created port: 8bcd11cf-b802-45c5-b77b-4779291e581d {{(pid=61957) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 960.205924] env[61957]: DEBUG nova.compute.manager [None req-d1aa370b-6bd9-49b7-9a71-30e9793a2856 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] [instance: 00ecc8c2-398e-4a29-b19f-a2013985b481] Start building block device mappings for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 960.240796] env[61957]: DEBUG oslo_vmware.api [None req-7f4318a7-7ec4-4da1-9bbe-cec59770ca17 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Task: {'id': task-1277824, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.513259] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22d32ef9-8fac-46f5-8a43-de30ea9338a7 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.521385] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-647cb531-05d9-4335-b746-72261dd8af04 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.555914] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acdc3b09-c3e8-4ddb-9f98-034e31f50d81 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.563991] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a397c35-3ab8-456e-9408-dc2a3c2d57c0 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.578103] env[61957]: DEBUG nova.compute.provider_tree [None req-9e60b82d-4fee-4f83-a3da-b48f03d27cc5 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 960.739675] env[61957]: DEBUG oslo_vmware.api [None req-7f4318a7-7ec4-4da1-9bbe-cec59770ca17 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Task: {'id': task-1277824, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.596958} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.739999] env[61957]: INFO nova.virt.vmwareapi.ds_util [None req-7f4318a7-7ec4-4da1-9bbe-cec59770ca17 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore2] 7793aa07-6aa2-459d-8a91-56c6b0412d3b/11c76a2c-f705-470a-ba9d-4657858bab38-rescue.vmdk. [ 960.740977] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54baaf29-ee2e-42d8-bd6d-bad727cd0a14 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.766392] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-7f4318a7-7ec4-4da1-9bbe-cec59770ca17 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] [instance: 7793aa07-6aa2-459d-8a91-56c6b0412d3b] Reconfiguring VM instance instance-00000054 to attach disk [datastore2] 7793aa07-6aa2-459d-8a91-56c6b0412d3b/11c76a2c-f705-470a-ba9d-4657858bab38-rescue.vmdk or device None with type thin {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 960.767462] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9d97897e-1ad7-4b9c-89bb-d491d0d431b7 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.787518] env[61957]: DEBUG oslo_vmware.api [None req-7f4318a7-7ec4-4da1-9bbe-cec59770ca17 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Waiting for the task: (returnval){ [ 960.787518] env[61957]: value = "task-1277825" [ 960.787518] env[61957]: _type = "Task" [ 960.787518] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.796630] env[61957]: DEBUG oslo_vmware.api [None req-7f4318a7-7ec4-4da1-9bbe-cec59770ca17 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Task: {'id': task-1277825, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.081840] env[61957]: DEBUG nova.scheduler.client.report [None req-9e60b82d-4fee-4f83-a3da-b48f03d27cc5 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 961.220210] env[61957]: DEBUG nova.compute.manager [None req-d1aa370b-6bd9-49b7-9a71-30e9793a2856 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] [instance: 00ecc8c2-398e-4a29-b19f-a2013985b481] Start spawning the instance on the hypervisor. {{(pid=61957) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 961.248717] env[61957]: DEBUG nova.virt.hardware [None req-d1aa370b-6bd9-49b7-9a71-30e9793a2856 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T17:22:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T17:21:50Z,direct_url=,disk_format='vmdk',id=11c76a2c-f705-470a-ba9d-4657858bab38,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='81d3096d0f094373913fc3dc8f5c3c6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T17:21:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 961.248844] env[61957]: DEBUG nova.virt.hardware [None req-d1aa370b-6bd9-49b7-9a71-30e9793a2856 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Flavor limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 961.248919] env[61957]: DEBUG nova.virt.hardware [None req-d1aa370b-6bd9-49b7-9a71-30e9793a2856 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Image limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 961.249120] env[61957]: DEBUG nova.virt.hardware [None req-d1aa370b-6bd9-49b7-9a71-30e9793a2856 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Flavor pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 961.249506] env[61957]: DEBUG nova.virt.hardware [None req-d1aa370b-6bd9-49b7-9a71-30e9793a2856 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Image pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 961.249753] env[61957]: DEBUG nova.virt.hardware [None req-d1aa370b-6bd9-49b7-9a71-30e9793a2856 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 961.249998] env[61957]: DEBUG nova.virt.hardware [None req-d1aa370b-6bd9-49b7-9a71-30e9793a2856 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 961.251087] env[61957]: DEBUG nova.virt.hardware [None req-d1aa370b-6bd9-49b7-9a71-30e9793a2856 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 961.251087] env[61957]: DEBUG nova.virt.hardware [None req-d1aa370b-6bd9-49b7-9a71-30e9793a2856 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Got 1 possible topologies {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 961.251087] env[61957]: DEBUG nova.virt.hardware [None req-d1aa370b-6bd9-49b7-9a71-30e9793a2856 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 961.251087] env[61957]: DEBUG nova.virt.hardware [None req-d1aa370b-6bd9-49b7-9a71-30e9793a2856 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 961.251670] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-795ccc89-ce42-4bd4-8b36-de6fe2c76330 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.262074] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c072550-e045-4dc6-8077-39ddc1b563b8 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.297703] env[61957]: DEBUG oslo_vmware.api [None req-7f4318a7-7ec4-4da1-9bbe-cec59770ca17 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Task: {'id': task-1277825, 'name': ReconfigVM_Task, 'duration_secs': 0.302644} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.298037] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-7f4318a7-7ec4-4da1-9bbe-cec59770ca17 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] [instance: 7793aa07-6aa2-459d-8a91-56c6b0412d3b] Reconfigured VM instance instance-00000054 to attach disk [datastore2] 7793aa07-6aa2-459d-8a91-56c6b0412d3b/11c76a2c-f705-470a-ba9d-4657858bab38-rescue.vmdk or device None with type thin {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 961.298892] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7aecade8-8a81-4f66-bbb3-97b5b1cfa6a6 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.324316] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-56c0589f-0343-4580-9d8b-7aaa1448ee82 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.339893] env[61957]: DEBUG oslo_vmware.api [None req-7f4318a7-7ec4-4da1-9bbe-cec59770ca17 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Waiting for the task: (returnval){ [ 961.339893] env[61957]: value = "task-1277826" [ 961.339893] env[61957]: _type = "Task" [ 961.339893] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.348337] env[61957]: DEBUG oslo_vmware.api [None req-7f4318a7-7ec4-4da1-9bbe-cec59770ca17 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Task: {'id': task-1277826, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.587514] env[61957]: DEBUG oslo_concurrency.lockutils [None req-9e60b82d-4fee-4f83-a3da-b48f03d27cc5 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.395s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 961.587983] env[61957]: DEBUG nova.compute.manager [None req-9e60b82d-4fee-4f83-a3da-b48f03d27cc5 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: d6c5c70d-86de-4dea-8b9e-76f321947a35] Start building networks asynchronously for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 961.590650] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ce0965c3-bc7d-4a77-9d32-52f4d059d837 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.793s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 961.592217] env[61957]: INFO nova.compute.claims [None req-ce0965c3-bc7d-4a77-9d32-52f4d059d837 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] [instance: 964c8c9d-0721-4ad3-a146-fce1fd1d6513] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 961.636720] env[61957]: DEBUG nova.compute.manager [req-954c7e4b-9d02-4a4a-80a4-d3f54a26fbcd req-5f82e7f9-6356-4653-ae75-935d0ad85485 service nova] [instance: 00ecc8c2-398e-4a29-b19f-a2013985b481] Received event network-vif-plugged-8bcd11cf-b802-45c5-b77b-4779291e581d {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 961.637144] env[61957]: DEBUG oslo_concurrency.lockutils [req-954c7e4b-9d02-4a4a-80a4-d3f54a26fbcd req-5f82e7f9-6356-4653-ae75-935d0ad85485 service nova] Acquiring lock "00ecc8c2-398e-4a29-b19f-a2013985b481-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 961.637399] env[61957]: DEBUG oslo_concurrency.lockutils [req-954c7e4b-9d02-4a4a-80a4-d3f54a26fbcd req-5f82e7f9-6356-4653-ae75-935d0ad85485 service nova] Lock "00ecc8c2-398e-4a29-b19f-a2013985b481-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 961.637722] env[61957]: DEBUG oslo_concurrency.lockutils [req-954c7e4b-9d02-4a4a-80a4-d3f54a26fbcd req-5f82e7f9-6356-4653-ae75-935d0ad85485 service nova] Lock "00ecc8c2-398e-4a29-b19f-a2013985b481-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 961.637976] env[61957]: DEBUG nova.compute.manager [req-954c7e4b-9d02-4a4a-80a4-d3f54a26fbcd req-5f82e7f9-6356-4653-ae75-935d0ad85485 service nova] [instance: 00ecc8c2-398e-4a29-b19f-a2013985b481] No waiting events found dispatching network-vif-plugged-8bcd11cf-b802-45c5-b77b-4779291e581d {{(pid=61957) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 961.638209] env[61957]: WARNING nova.compute.manager [req-954c7e4b-9d02-4a4a-80a4-d3f54a26fbcd req-5f82e7f9-6356-4653-ae75-935d0ad85485 service nova] [instance: 00ecc8c2-398e-4a29-b19f-a2013985b481] Received unexpected event network-vif-plugged-8bcd11cf-b802-45c5-b77b-4779291e581d for instance with vm_state building and task_state spawning. [ 961.738554] env[61957]: DEBUG nova.network.neutron [None req-d1aa370b-6bd9-49b7-9a71-30e9793a2856 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] [instance: 00ecc8c2-398e-4a29-b19f-a2013985b481] Successfully updated port: 8bcd11cf-b802-45c5-b77b-4779291e581d {{(pid=61957) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 961.850958] env[61957]: DEBUG oslo_vmware.api [None req-7f4318a7-7ec4-4da1-9bbe-cec59770ca17 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Task: {'id': task-1277826, 'name': ReconfigVM_Task, 'duration_secs': 0.153899} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.850958] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-7f4318a7-7ec4-4da1-9bbe-cec59770ca17 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] [instance: 7793aa07-6aa2-459d-8a91-56c6b0412d3b] Powering on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 961.850958] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-672b96ae-ff2f-4ef2-a9d7-0b19e139cbb7 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.857805] env[61957]: DEBUG oslo_vmware.api [None req-7f4318a7-7ec4-4da1-9bbe-cec59770ca17 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Waiting for the task: (returnval){ [ 961.857805] env[61957]: value = "task-1277827" [ 961.857805] env[61957]: _type = "Task" [ 961.857805] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.865014] env[61957]: DEBUG oslo_vmware.api [None req-7f4318a7-7ec4-4da1-9bbe-cec59770ca17 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Task: {'id': task-1277827, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.096569] env[61957]: DEBUG nova.compute.utils [None req-9e60b82d-4fee-4f83-a3da-b48f03d27cc5 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Using /dev/sd instead of None {{(pid=61957) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 962.100271] env[61957]: DEBUG nova.compute.manager [None req-9e60b82d-4fee-4f83-a3da-b48f03d27cc5 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: d6c5c70d-86de-4dea-8b9e-76f321947a35] Allocating IP information in the background. {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 962.100474] env[61957]: DEBUG nova.network.neutron [None req-9e60b82d-4fee-4f83-a3da-b48f03d27cc5 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: d6c5c70d-86de-4dea-8b9e-76f321947a35] allocate_for_instance() {{(pid=61957) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 962.150594] env[61957]: DEBUG nova.policy [None req-9e60b82d-4fee-4f83-a3da-b48f03d27cc5 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '76e1754f96ed438ea0ed91ad337419f1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '975189ebb3cc4cdb9391880f0c9ba6ba', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61957) authorize /opt/stack/nova/nova/policy.py:203}} [ 962.241233] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d1aa370b-6bd9-49b7-9a71-30e9793a2856 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Acquiring lock "refresh_cache-00ecc8c2-398e-4a29-b19f-a2013985b481" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 962.241492] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d1aa370b-6bd9-49b7-9a71-30e9793a2856 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Acquired lock "refresh_cache-00ecc8c2-398e-4a29-b19f-a2013985b481" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 962.241600] env[61957]: DEBUG nova.network.neutron [None req-d1aa370b-6bd9-49b7-9a71-30e9793a2856 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] [instance: 00ecc8c2-398e-4a29-b19f-a2013985b481] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 962.372066] env[61957]: DEBUG oslo_vmware.api [None req-7f4318a7-7ec4-4da1-9bbe-cec59770ca17 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Task: {'id': task-1277827, 'name': PowerOnVM_Task, 'duration_secs': 0.402214} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.372359] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-7f4318a7-7ec4-4da1-9bbe-cec59770ca17 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] [instance: 7793aa07-6aa2-459d-8a91-56c6b0412d3b] Powered on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 962.376127] env[61957]: DEBUG nova.compute.manager [None req-7f4318a7-7ec4-4da1-9bbe-cec59770ca17 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] [instance: 7793aa07-6aa2-459d-8a91-56c6b0412d3b] Checking state {{(pid=61957) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 962.376998] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19cae805-0b6a-40ed-9c28-231a39d7e620 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.437548] env[61957]: DEBUG nova.network.neutron [None req-9e60b82d-4fee-4f83-a3da-b48f03d27cc5 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: d6c5c70d-86de-4dea-8b9e-76f321947a35] Successfully created port: 642fc1a5-237a-4bde-b28f-69b3f0f46a8a {{(pid=61957) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 962.601341] env[61957]: DEBUG nova.compute.manager [None req-9e60b82d-4fee-4f83-a3da-b48f03d27cc5 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: d6c5c70d-86de-4dea-8b9e-76f321947a35] Start building block device mappings for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 962.772950] env[61957]: DEBUG nova.network.neutron [None req-d1aa370b-6bd9-49b7-9a71-30e9793a2856 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] [instance: 00ecc8c2-398e-4a29-b19f-a2013985b481] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 962.848994] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8621974a-21ac-42cd-83c6-ebdd85978f30 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.859999] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-371322a8-0f38-449c-972d-1e8a8998a314 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.903194] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2896ad0c-5807-4591-a34a-6dd73df79f74 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.917288] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80d14d49-56bd-4adc-aa3e-a3a48e1f293f {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.931418] env[61957]: DEBUG nova.compute.provider_tree [None req-ce0965c3-bc7d-4a77-9d32-52f4d059d837 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 962.958636] env[61957]: DEBUG nova.network.neutron [None req-d1aa370b-6bd9-49b7-9a71-30e9793a2856 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] [instance: 00ecc8c2-398e-4a29-b19f-a2013985b481] Updating instance_info_cache with network_info: [{"id": "8bcd11cf-b802-45c5-b77b-4779291e581d", "address": "fa:16:3e:24:9b:92", "network": {"id": "cf378e9c-eaae-4168-bd60-9818fcfaf405", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1243326371-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f7c20ab3c822423fa837a03f066e4239", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8cb478a6-872c-4a90-a8db-526b374e82ce", "external-id": "nsx-vlan-transportzone-835", "segmentation_id": 835, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8bcd11cf-b8", "ovs_interfaceid": "8bcd11cf-b802-45c5-b77b-4779291e581d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 963.435378] env[61957]: DEBUG nova.scheduler.client.report [None req-ce0965c3-bc7d-4a77-9d32-52f4d059d837 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 963.446587] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d1aa370b-6bd9-49b7-9a71-30e9793a2856 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Releasing lock "refresh_cache-00ecc8c2-398e-4a29-b19f-a2013985b481" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 963.446960] env[61957]: DEBUG nova.compute.manager [None req-d1aa370b-6bd9-49b7-9a71-30e9793a2856 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] [instance: 00ecc8c2-398e-4a29-b19f-a2013985b481] Instance network_info: |[{"id": "8bcd11cf-b802-45c5-b77b-4779291e581d", "address": "fa:16:3e:24:9b:92", "network": {"id": "cf378e9c-eaae-4168-bd60-9818fcfaf405", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1243326371-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f7c20ab3c822423fa837a03f066e4239", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8cb478a6-872c-4a90-a8db-526b374e82ce", "external-id": "nsx-vlan-transportzone-835", "segmentation_id": 835, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8bcd11cf-b8", "ovs_interfaceid": "8bcd11cf-b802-45c5-b77b-4779291e581d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 963.447410] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-d1aa370b-6bd9-49b7-9a71-30e9793a2856 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] [instance: 00ecc8c2-398e-4a29-b19f-a2013985b481] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:24:9b:92', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8cb478a6-872c-4a90-a8db-526b374e82ce', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8bcd11cf-b802-45c5-b77b-4779291e581d', 'vif_model': 'vmxnet3'}] {{(pid=61957) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 963.454878] env[61957]: DEBUG oslo.service.loopingcall [None req-d1aa370b-6bd9-49b7-9a71-30e9793a2856 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 963.455754] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 00ecc8c2-398e-4a29-b19f-a2013985b481] Creating VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 963.455996] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c6326691-bb62-42e6-a25f-f185d551af03 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.476603] env[61957]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 963.476603] env[61957]: value = "task-1277828" [ 963.476603] env[61957]: _type = "Task" [ 963.476603] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.485479] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277828, 'name': CreateVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.612959] env[61957]: DEBUG nova.compute.manager [None req-9e60b82d-4fee-4f83-a3da-b48f03d27cc5 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: d6c5c70d-86de-4dea-8b9e-76f321947a35] Start spawning the instance on the hypervisor. {{(pid=61957) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 963.641756] env[61957]: DEBUG nova.virt.hardware [None req-9e60b82d-4fee-4f83-a3da-b48f03d27cc5 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T17:22:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='aa8666c91040afd6c19df2759aa0a462',container_format='bare',created_at=2024-10-12T17:29:51Z,direct_url=,disk_format='vmdk',id=fa797977-e3da-4f16-8445-e20c97e9e8ae,min_disk=1,min_ram=0,name='tempest-test-snap-1948603291',owner='975189ebb3cc4cdb9391880f0c9ba6ba',properties=ImageMetaProps,protected=,size=21334016,status='active',tags=,updated_at=2024-10-12T17:30:06Z,virtual_size=,visibility=), allow threads: False {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 963.641756] env[61957]: DEBUG nova.virt.hardware [None req-9e60b82d-4fee-4f83-a3da-b48f03d27cc5 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Flavor limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 963.641891] env[61957]: DEBUG nova.virt.hardware [None req-9e60b82d-4fee-4f83-a3da-b48f03d27cc5 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Image limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 963.642034] env[61957]: DEBUG nova.virt.hardware [None req-9e60b82d-4fee-4f83-a3da-b48f03d27cc5 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Flavor pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 963.642196] env[61957]: DEBUG nova.virt.hardware [None req-9e60b82d-4fee-4f83-a3da-b48f03d27cc5 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Image pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 963.642361] env[61957]: DEBUG nova.virt.hardware [None req-9e60b82d-4fee-4f83-a3da-b48f03d27cc5 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 963.642614] env[61957]: DEBUG nova.virt.hardware [None req-9e60b82d-4fee-4f83-a3da-b48f03d27cc5 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 963.642786] env[61957]: DEBUG nova.virt.hardware [None req-9e60b82d-4fee-4f83-a3da-b48f03d27cc5 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 963.642958] env[61957]: DEBUG nova.virt.hardware [None req-9e60b82d-4fee-4f83-a3da-b48f03d27cc5 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Got 1 possible topologies {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 963.643138] env[61957]: DEBUG nova.virt.hardware [None req-9e60b82d-4fee-4f83-a3da-b48f03d27cc5 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 963.643317] env[61957]: DEBUG nova.virt.hardware [None req-9e60b82d-4fee-4f83-a3da-b48f03d27cc5 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 963.644213] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54691dd5-2b20-4e14-aedb-ab39638c9ac6 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.653009] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-717cedca-543e-4803-963f-3ccd62ab22b6 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.699629] env[61957]: DEBUG nova.compute.manager [req-8a351fd2-7004-4a9f-a338-8b781b0e478e req-4c5a09ba-91c8-4057-80cb-ed7696d3d2cd service nova] [instance: 00ecc8c2-398e-4a29-b19f-a2013985b481] Received event network-changed-8bcd11cf-b802-45c5-b77b-4779291e581d {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 963.699762] env[61957]: DEBUG nova.compute.manager [req-8a351fd2-7004-4a9f-a338-8b781b0e478e req-4c5a09ba-91c8-4057-80cb-ed7696d3d2cd service nova] [instance: 00ecc8c2-398e-4a29-b19f-a2013985b481] Refreshing instance network info cache due to event network-changed-8bcd11cf-b802-45c5-b77b-4779291e581d. {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 963.699959] env[61957]: DEBUG oslo_concurrency.lockutils [req-8a351fd2-7004-4a9f-a338-8b781b0e478e req-4c5a09ba-91c8-4057-80cb-ed7696d3d2cd service nova] Acquiring lock "refresh_cache-00ecc8c2-398e-4a29-b19f-a2013985b481" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 963.700212] env[61957]: DEBUG oslo_concurrency.lockutils [req-8a351fd2-7004-4a9f-a338-8b781b0e478e req-4c5a09ba-91c8-4057-80cb-ed7696d3d2cd service nova] Acquired lock "refresh_cache-00ecc8c2-398e-4a29-b19f-a2013985b481" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 963.700291] env[61957]: DEBUG nova.network.neutron [req-8a351fd2-7004-4a9f-a338-8b781b0e478e req-4c5a09ba-91c8-4057-80cb-ed7696d3d2cd service nova] [instance: 00ecc8c2-398e-4a29-b19f-a2013985b481] Refreshing network info cache for port 8bcd11cf-b802-45c5-b77b-4779291e581d {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 963.883744] env[61957]: DEBUG nova.compute.manager [req-12e6a1cd-f7e8-4eb6-921b-911913788d1c req-c2aa4d62-1ce3-4c11-90d8-0d44c7af188d service nova] [instance: d6c5c70d-86de-4dea-8b9e-76f321947a35] Received event network-vif-plugged-642fc1a5-237a-4bde-b28f-69b3f0f46a8a {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 963.884052] env[61957]: DEBUG oslo_concurrency.lockutils [req-12e6a1cd-f7e8-4eb6-921b-911913788d1c req-c2aa4d62-1ce3-4c11-90d8-0d44c7af188d service nova] Acquiring lock "d6c5c70d-86de-4dea-8b9e-76f321947a35-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 963.884278] env[61957]: DEBUG oslo_concurrency.lockutils [req-12e6a1cd-f7e8-4eb6-921b-911913788d1c req-c2aa4d62-1ce3-4c11-90d8-0d44c7af188d service nova] Lock "d6c5c70d-86de-4dea-8b9e-76f321947a35-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 963.884475] env[61957]: DEBUG oslo_concurrency.lockutils [req-12e6a1cd-f7e8-4eb6-921b-911913788d1c req-c2aa4d62-1ce3-4c11-90d8-0d44c7af188d service nova] Lock "d6c5c70d-86de-4dea-8b9e-76f321947a35-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 963.885143] env[61957]: DEBUG nova.compute.manager [req-12e6a1cd-f7e8-4eb6-921b-911913788d1c req-c2aa4d62-1ce3-4c11-90d8-0d44c7af188d service nova] [instance: d6c5c70d-86de-4dea-8b9e-76f321947a35] No waiting events found dispatching network-vif-plugged-642fc1a5-237a-4bde-b28f-69b3f0f46a8a {{(pid=61957) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 963.885781] env[61957]: WARNING nova.compute.manager [req-12e6a1cd-f7e8-4eb6-921b-911913788d1c req-c2aa4d62-1ce3-4c11-90d8-0d44c7af188d service nova] [instance: d6c5c70d-86de-4dea-8b9e-76f321947a35] Received unexpected event network-vif-plugged-642fc1a5-237a-4bde-b28f-69b3f0f46a8a for instance with vm_state building and task_state spawning. [ 963.940525] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ce0965c3-bc7d-4a77-9d32-52f4d059d837 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.349s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 963.940962] env[61957]: DEBUG nova.compute.manager [None req-ce0965c3-bc7d-4a77-9d32-52f4d059d837 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] [instance: 964c8c9d-0721-4ad3-a146-fce1fd1d6513] Start building networks asynchronously for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 963.944070] env[61957]: DEBUG oslo_concurrency.lockutils [None req-cedd08d9-20a0-42aa-9f1f-33859d216b7f tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.693s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 963.945694] env[61957]: INFO nova.compute.claims [None req-cedd08d9-20a0-42aa-9f1f-33859d216b7f tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 98929ed3-d420-45cd-9cde-9738b2ac8251] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 963.986869] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277828, 'name': CreateVM_Task, 'duration_secs': 0.439798} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.987075] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 00ecc8c2-398e-4a29-b19f-a2013985b481] Created VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 963.987745] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d1aa370b-6bd9-49b7-9a71-30e9793a2856 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 963.988809] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d1aa370b-6bd9-49b7-9a71-30e9793a2856 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Acquired lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 963.988809] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d1aa370b-6bd9-49b7-9a71-30e9793a2856 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 963.988809] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f67cb2aa-126c-4468-89a9-993409381251 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.994082] env[61957]: DEBUG oslo_vmware.api [None req-d1aa370b-6bd9-49b7-9a71-30e9793a2856 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Waiting for the task: (returnval){ [ 963.994082] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52ec4aa1-8f90-4806-0444-8a2a289ce9e0" [ 963.994082] env[61957]: _type = "Task" [ 963.994082] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.002646] env[61957]: DEBUG oslo_vmware.api [None req-d1aa370b-6bd9-49b7-9a71-30e9793a2856 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52ec4aa1-8f90-4806-0444-8a2a289ce9e0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.145322] env[61957]: DEBUG oslo_concurrency.lockutils [None req-731394e9-d0a3-46c3-8b02-41ef72d369f5 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Acquiring lock "f65458b2-0db6-4d98-8c82-b27b6945a3fd" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 964.145548] env[61957]: DEBUG oslo_concurrency.lockutils [None req-731394e9-d0a3-46c3-8b02-41ef72d369f5 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Lock "f65458b2-0db6-4d98-8c82-b27b6945a3fd" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 964.411927] env[61957]: DEBUG nova.network.neutron [req-8a351fd2-7004-4a9f-a338-8b781b0e478e req-4c5a09ba-91c8-4057-80cb-ed7696d3d2cd service nova] [instance: 00ecc8c2-398e-4a29-b19f-a2013985b481] Updated VIF entry in instance network info cache for port 8bcd11cf-b802-45c5-b77b-4779291e581d. {{(pid=61957) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 964.412526] env[61957]: DEBUG nova.network.neutron [req-8a351fd2-7004-4a9f-a338-8b781b0e478e req-4c5a09ba-91c8-4057-80cb-ed7696d3d2cd service nova] [instance: 00ecc8c2-398e-4a29-b19f-a2013985b481] Updating instance_info_cache with network_info: [{"id": "8bcd11cf-b802-45c5-b77b-4779291e581d", "address": "fa:16:3e:24:9b:92", "network": {"id": "cf378e9c-eaae-4168-bd60-9818fcfaf405", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1243326371-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f7c20ab3c822423fa837a03f066e4239", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8cb478a6-872c-4a90-a8db-526b374e82ce", "external-id": "nsx-vlan-transportzone-835", "segmentation_id": 835, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8bcd11cf-b8", "ovs_interfaceid": "8bcd11cf-b802-45c5-b77b-4779291e581d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 964.443131] env[61957]: DEBUG nova.network.neutron [None req-9e60b82d-4fee-4f83-a3da-b48f03d27cc5 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: d6c5c70d-86de-4dea-8b9e-76f321947a35] Successfully updated port: 642fc1a5-237a-4bde-b28f-69b3f0f46a8a {{(pid=61957) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 964.451505] env[61957]: DEBUG nova.compute.utils [None req-ce0965c3-bc7d-4a77-9d32-52f4d059d837 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Using /dev/sd instead of None {{(pid=61957) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 964.460205] env[61957]: DEBUG nova.compute.manager [None req-ce0965c3-bc7d-4a77-9d32-52f4d059d837 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] [instance: 964c8c9d-0721-4ad3-a146-fce1fd1d6513] Allocating IP information in the background. {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 964.460205] env[61957]: DEBUG nova.network.neutron [None req-ce0965c3-bc7d-4a77-9d32-52f4d059d837 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] [instance: 964c8c9d-0721-4ad3-a146-fce1fd1d6513] allocate_for_instance() {{(pid=61957) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 964.505561] env[61957]: DEBUG oslo_vmware.api [None req-d1aa370b-6bd9-49b7-9a71-30e9793a2856 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52ec4aa1-8f90-4806-0444-8a2a289ce9e0, 'name': SearchDatastore_Task, 'duration_secs': 0.022615} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.507149] env[61957]: DEBUG nova.policy [None req-ce0965c3-bc7d-4a77-9d32-52f4d059d837 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd0405e9c38cb4012a0212284be7e9aed', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1673c854272f4b14a623a73b8f4b687b', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61957) authorize /opt/stack/nova/nova/policy.py:203}} [ 964.508809] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d1aa370b-6bd9-49b7-9a71-30e9793a2856 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Releasing lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 964.509071] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-d1aa370b-6bd9-49b7-9a71-30e9793a2856 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] [instance: 00ecc8c2-398e-4a29-b19f-a2013985b481] Processing image 11c76a2c-f705-470a-ba9d-4657858bab38 {{(pid=61957) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 964.509334] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d1aa370b-6bd9-49b7-9a71-30e9793a2856 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 964.509507] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d1aa370b-6bd9-49b7-9a71-30e9793a2856 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Acquired lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 964.509701] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-d1aa370b-6bd9-49b7-9a71-30e9793a2856 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 964.510015] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a2577409-81f7-43dd-8bf1-fcab11d06c48 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.520111] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-d1aa370b-6bd9-49b7-9a71-30e9793a2856 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 964.520364] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-d1aa370b-6bd9-49b7-9a71-30e9793a2856 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61957) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 964.521167] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-08531818-d7a5-408f-bd56-a4c9af3b9a83 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.526617] env[61957]: DEBUG oslo_vmware.api [None req-d1aa370b-6bd9-49b7-9a71-30e9793a2856 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Waiting for the task: (returnval){ [ 964.526617] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52796b2e-c525-d105-c116-602f9fb207e7" [ 964.526617] env[61957]: _type = "Task" [ 964.526617] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.535349] env[61957]: DEBUG oslo_vmware.api [None req-d1aa370b-6bd9-49b7-9a71-30e9793a2856 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52796b2e-c525-d105-c116-602f9fb207e7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.647505] env[61957]: DEBUG nova.compute.manager [None req-731394e9-d0a3-46c3-8b02-41ef72d369f5 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] [instance: f65458b2-0db6-4d98-8c82-b27b6945a3fd] Starting instance... {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 964.827190] env[61957]: DEBUG nova.network.neutron [None req-ce0965c3-bc7d-4a77-9d32-52f4d059d837 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] [instance: 964c8c9d-0721-4ad3-a146-fce1fd1d6513] Successfully created port: 57483094-30a7-4f77-8c6e-25dbd1db61c7 {{(pid=61957) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 964.915298] env[61957]: DEBUG oslo_concurrency.lockutils [req-8a351fd2-7004-4a9f-a338-8b781b0e478e req-4c5a09ba-91c8-4057-80cb-ed7696d3d2cd service nova] Releasing lock "refresh_cache-00ecc8c2-398e-4a29-b19f-a2013985b481" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 964.946341] env[61957]: DEBUG oslo_concurrency.lockutils [None req-9e60b82d-4fee-4f83-a3da-b48f03d27cc5 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Acquiring lock "refresh_cache-d6c5c70d-86de-4dea-8b9e-76f321947a35" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 964.947445] env[61957]: DEBUG oslo_concurrency.lockutils [None req-9e60b82d-4fee-4f83-a3da-b48f03d27cc5 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Acquired lock "refresh_cache-d6c5c70d-86de-4dea-8b9e-76f321947a35" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 964.947445] env[61957]: DEBUG nova.network.neutron [None req-9e60b82d-4fee-4f83-a3da-b48f03d27cc5 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: d6c5c70d-86de-4dea-8b9e-76f321947a35] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 964.961283] env[61957]: DEBUG nova.compute.manager [None req-ce0965c3-bc7d-4a77-9d32-52f4d059d837 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] [instance: 964c8c9d-0721-4ad3-a146-fce1fd1d6513] Start building block device mappings for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 965.051199] env[61957]: DEBUG oslo_vmware.api [None req-d1aa370b-6bd9-49b7-9a71-30e9793a2856 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52796b2e-c525-d105-c116-602f9fb207e7, 'name': SearchDatastore_Task, 'duration_secs': 0.013276} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.052501] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cc9f605e-2b26-4ab1-9342-7dfee53a15b6 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.062337] env[61957]: DEBUG oslo_vmware.api [None req-d1aa370b-6bd9-49b7-9a71-30e9793a2856 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Waiting for the task: (returnval){ [ 965.062337] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52e228bf-0db7-17f9-b5f7-3dd28a9b09c9" [ 965.062337] env[61957]: _type = "Task" [ 965.062337] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.071819] env[61957]: DEBUG oslo_vmware.api [None req-d1aa370b-6bd9-49b7-9a71-30e9793a2856 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52e228bf-0db7-17f9-b5f7-3dd28a9b09c9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.168116] env[61957]: DEBUG oslo_concurrency.lockutils [None req-731394e9-d0a3-46c3-8b02-41ef72d369f5 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 965.250651] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23894de9-7bc0-4679-8672-d508c12de469 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.259956] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de982863-48ab-4143-af57-0043b8b173c7 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.293580] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6314a1c0-2622-4f7a-8dee-1ec227f75088 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.301584] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a8ea717-dbe3-4f57-a63f-a3fce160d270 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.315888] env[61957]: DEBUG nova.compute.provider_tree [None req-cedd08d9-20a0-42aa-9f1f-33859d216b7f tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 965.488608] env[61957]: DEBUG nova.network.neutron [None req-9e60b82d-4fee-4f83-a3da-b48f03d27cc5 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: d6c5c70d-86de-4dea-8b9e-76f321947a35] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 965.574276] env[61957]: DEBUG oslo_vmware.api [None req-d1aa370b-6bd9-49b7-9a71-30e9793a2856 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52e228bf-0db7-17f9-b5f7-3dd28a9b09c9, 'name': SearchDatastore_Task, 'duration_secs': 0.023893} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.574554] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d1aa370b-6bd9-49b7-9a71-30e9793a2856 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Releasing lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 965.574965] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-d1aa370b-6bd9-49b7-9a71-30e9793a2856 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore1] 00ecc8c2-398e-4a29-b19f-a2013985b481/00ecc8c2-398e-4a29-b19f-a2013985b481.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 965.575175] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8bbf1de0-0eee-4f43-a835-0b97a150e3e9 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.582596] env[61957]: DEBUG oslo_vmware.api [None req-d1aa370b-6bd9-49b7-9a71-30e9793a2856 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Waiting for the task: (returnval){ [ 965.582596] env[61957]: value = "task-1277829" [ 965.582596] env[61957]: _type = "Task" [ 965.582596] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.590816] env[61957]: DEBUG oslo_vmware.api [None req-d1aa370b-6bd9-49b7-9a71-30e9793a2856 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Task: {'id': task-1277829, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.641810] env[61957]: DEBUG nova.network.neutron [None req-9e60b82d-4fee-4f83-a3da-b48f03d27cc5 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: d6c5c70d-86de-4dea-8b9e-76f321947a35] Updating instance_info_cache with network_info: [{"id": "642fc1a5-237a-4bde-b28f-69b3f0f46a8a", "address": "fa:16:3e:36:b9:44", "network": {"id": "dc7d4c3b-1f91-4937-a2b0-1084e5c478f0", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1637370260-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "975189ebb3cc4cdb9391880f0c9ba6ba", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b71230ae-e879-4384-88ce-fe64c86fce22", "external-id": "nsx-vlan-transportzone-473", "segmentation_id": 473, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap642fc1a5-23", "ovs_interfaceid": "642fc1a5-237a-4bde-b28f-69b3f0f46a8a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 965.725546] env[61957]: DEBUG nova.compute.manager [req-4b934be2-80f9-43a3-af7d-0ec06ec94157 req-8a15376e-ad65-45e6-b4ac-e66288505a65 service nova] [instance: d6c5c70d-86de-4dea-8b9e-76f321947a35] Received event network-changed-642fc1a5-237a-4bde-b28f-69b3f0f46a8a {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 965.726278] env[61957]: DEBUG nova.compute.manager [req-4b934be2-80f9-43a3-af7d-0ec06ec94157 req-8a15376e-ad65-45e6-b4ac-e66288505a65 service nova] [instance: d6c5c70d-86de-4dea-8b9e-76f321947a35] Refreshing instance network info cache due to event network-changed-642fc1a5-237a-4bde-b28f-69b3f0f46a8a. {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 965.726499] env[61957]: DEBUG oslo_concurrency.lockutils [req-4b934be2-80f9-43a3-af7d-0ec06ec94157 req-8a15376e-ad65-45e6-b4ac-e66288505a65 service nova] Acquiring lock "refresh_cache-d6c5c70d-86de-4dea-8b9e-76f321947a35" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 965.819650] env[61957]: DEBUG nova.scheduler.client.report [None req-cedd08d9-20a0-42aa-9f1f-33859d216b7f tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 965.977122] env[61957]: DEBUG nova.compute.manager [None req-ce0965c3-bc7d-4a77-9d32-52f4d059d837 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] [instance: 964c8c9d-0721-4ad3-a146-fce1fd1d6513] Start spawning the instance on the hypervisor. {{(pid=61957) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 966.004289] env[61957]: DEBUG nova.virt.hardware [None req-ce0965c3-bc7d-4a77-9d32-52f4d059d837 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T17:22:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T17:21:50Z,direct_url=,disk_format='vmdk',id=11c76a2c-f705-470a-ba9d-4657858bab38,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='81d3096d0f094373913fc3dc8f5c3c6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T17:21:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 966.004583] env[61957]: DEBUG nova.virt.hardware [None req-ce0965c3-bc7d-4a77-9d32-52f4d059d837 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Flavor limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 966.004728] env[61957]: DEBUG nova.virt.hardware [None req-ce0965c3-bc7d-4a77-9d32-52f4d059d837 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Image limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 966.004924] env[61957]: DEBUG nova.virt.hardware [None req-ce0965c3-bc7d-4a77-9d32-52f4d059d837 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Flavor pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 966.005116] env[61957]: DEBUG nova.virt.hardware [None req-ce0965c3-bc7d-4a77-9d32-52f4d059d837 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Image pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 966.005233] env[61957]: DEBUG nova.virt.hardware [None req-ce0965c3-bc7d-4a77-9d32-52f4d059d837 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 966.005449] env[61957]: DEBUG nova.virt.hardware [None req-ce0965c3-bc7d-4a77-9d32-52f4d059d837 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 966.005618] env[61957]: DEBUG nova.virt.hardware [None req-ce0965c3-bc7d-4a77-9d32-52f4d059d837 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 966.005793] env[61957]: DEBUG nova.virt.hardware [None req-ce0965c3-bc7d-4a77-9d32-52f4d059d837 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Got 1 possible topologies {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 966.005963] env[61957]: DEBUG nova.virt.hardware [None req-ce0965c3-bc7d-4a77-9d32-52f4d059d837 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 966.006156] env[61957]: DEBUG nova.virt.hardware [None req-ce0965c3-bc7d-4a77-9d32-52f4d059d837 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 966.007146] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8951062c-a65f-4484-8313-8ef7110f0d88 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.016558] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28e9b198-fa3a-4248-96ea-885b2ed80caa {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.093158] env[61957]: DEBUG oslo_vmware.api [None req-d1aa370b-6bd9-49b7-9a71-30e9793a2856 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Task: {'id': task-1277829, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.144629] env[61957]: DEBUG oslo_concurrency.lockutils [None req-9e60b82d-4fee-4f83-a3da-b48f03d27cc5 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Releasing lock "refresh_cache-d6c5c70d-86de-4dea-8b9e-76f321947a35" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 966.144986] env[61957]: DEBUG nova.compute.manager [None req-9e60b82d-4fee-4f83-a3da-b48f03d27cc5 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: d6c5c70d-86de-4dea-8b9e-76f321947a35] Instance network_info: |[{"id": "642fc1a5-237a-4bde-b28f-69b3f0f46a8a", "address": "fa:16:3e:36:b9:44", "network": {"id": "dc7d4c3b-1f91-4937-a2b0-1084e5c478f0", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1637370260-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "975189ebb3cc4cdb9391880f0c9ba6ba", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b71230ae-e879-4384-88ce-fe64c86fce22", "external-id": "nsx-vlan-transportzone-473", "segmentation_id": 473, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap642fc1a5-23", "ovs_interfaceid": "642fc1a5-237a-4bde-b28f-69b3f0f46a8a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 966.145350] env[61957]: DEBUG oslo_concurrency.lockutils [req-4b934be2-80f9-43a3-af7d-0ec06ec94157 req-8a15376e-ad65-45e6-b4ac-e66288505a65 service nova] Acquired lock "refresh_cache-d6c5c70d-86de-4dea-8b9e-76f321947a35" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 966.145532] env[61957]: DEBUG nova.network.neutron [req-4b934be2-80f9-43a3-af7d-0ec06ec94157 req-8a15376e-ad65-45e6-b4ac-e66288505a65 service nova] [instance: d6c5c70d-86de-4dea-8b9e-76f321947a35] Refreshing network info cache for port 642fc1a5-237a-4bde-b28f-69b3f0f46a8a {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 966.146895] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-9e60b82d-4fee-4f83-a3da-b48f03d27cc5 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: d6c5c70d-86de-4dea-8b9e-76f321947a35] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:36:b9:44', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'b71230ae-e879-4384-88ce-fe64c86fce22', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '642fc1a5-237a-4bde-b28f-69b3f0f46a8a', 'vif_model': 'vmxnet3'}] {{(pid=61957) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 966.156496] env[61957]: DEBUG oslo.service.loopingcall [None req-9e60b82d-4fee-4f83-a3da-b48f03d27cc5 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 966.160510] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d6c5c70d-86de-4dea-8b9e-76f321947a35] Creating VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 966.162231] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-35c1ebba-d97c-4f72-b3fa-40d0032a606f {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.183494] env[61957]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 966.183494] env[61957]: value = "task-1277830" [ 966.183494] env[61957]: _type = "Task" [ 966.183494] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.192686] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277830, 'name': CreateVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.239229] env[61957]: DEBUG oslo_vmware.rw_handles [None req-3141e8c0-2e34-4c8b-a8ac-0f12d351422c tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52c996eb-d80c-6e44-ead7-48e0e52505b7/disk-0.vmdk. {{(pid=61957) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 966.240487] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b73ec3e-81d6-4ca8-b74d-fdcb5614165a {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.250454] env[61957]: DEBUG oslo_vmware.rw_handles [None req-3141e8c0-2e34-4c8b-a8ac-0f12d351422c tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52c996eb-d80c-6e44-ead7-48e0e52505b7/disk-0.vmdk is in state: ready. {{(pid=61957) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 966.250815] env[61957]: ERROR oslo_vmware.rw_handles [None req-3141e8c0-2e34-4c8b-a8ac-0f12d351422c tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52c996eb-d80c-6e44-ead7-48e0e52505b7/disk-0.vmdk due to incomplete transfer. [ 966.251058] env[61957]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-f0e0ce06-9b93-4be0-b934-d5b8003cbf45 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.258836] env[61957]: DEBUG oslo_vmware.rw_handles [None req-3141e8c0-2e34-4c8b-a8ac-0f12d351422c tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52c996eb-d80c-6e44-ead7-48e0e52505b7/disk-0.vmdk. {{(pid=61957) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 966.259147] env[61957]: DEBUG nova.virt.vmwareapi.images [None req-3141e8c0-2e34-4c8b-a8ac-0f12d351422c tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] [instance: 736eabcf-70f8-4e1e-9830-270d9971fc9a] Uploaded image 0018d8e9-1d6e-4ca4-a224-31fc9e9f9d2e to the Glance image server {{(pid=61957) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 966.262497] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-3141e8c0-2e34-4c8b-a8ac-0f12d351422c tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] [instance: 736eabcf-70f8-4e1e-9830-270d9971fc9a] Destroying the VM {{(pid=61957) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 966.262497] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-a21c0c73-b32d-479a-b671-1c7d2ca27597 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.268971] env[61957]: DEBUG oslo_vmware.api [None req-3141e8c0-2e34-4c8b-a8ac-0f12d351422c tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Waiting for the task: (returnval){ [ 966.268971] env[61957]: value = "task-1277831" [ 966.268971] env[61957]: _type = "Task" [ 966.268971] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.277208] env[61957]: DEBUG oslo_vmware.api [None req-3141e8c0-2e34-4c8b-a8ac-0f12d351422c tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Task: {'id': task-1277831, 'name': Destroy_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.324835] env[61957]: DEBUG oslo_concurrency.lockutils [None req-cedd08d9-20a0-42aa-9f1f-33859d216b7f tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.381s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 966.325444] env[61957]: DEBUG nova.compute.manager [None req-cedd08d9-20a0-42aa-9f1f-33859d216b7f tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 98929ed3-d420-45cd-9cde-9738b2ac8251] Start building networks asynchronously for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 966.328552] env[61957]: DEBUG oslo_concurrency.lockutils [None req-5dd00032-4725-4b73-a1a7-7ef95ae8be93 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 12.155s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 966.328948] env[61957]: DEBUG nova.objects.instance [None req-5dd00032-4725-4b73-a1a7-7ef95ae8be93 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Lazy-loading 'resources' on Instance uuid 50e09db7-9f8f-452b-8232-bd7473e9e63a {{(pid=61957) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 966.510889] env[61957]: DEBUG nova.network.neutron [req-4b934be2-80f9-43a3-af7d-0ec06ec94157 req-8a15376e-ad65-45e6-b4ac-e66288505a65 service nova] [instance: d6c5c70d-86de-4dea-8b9e-76f321947a35] Updated VIF entry in instance network info cache for port 642fc1a5-237a-4bde-b28f-69b3f0f46a8a. {{(pid=61957) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 966.511281] env[61957]: DEBUG nova.network.neutron [req-4b934be2-80f9-43a3-af7d-0ec06ec94157 req-8a15376e-ad65-45e6-b4ac-e66288505a65 service nova] [instance: d6c5c70d-86de-4dea-8b9e-76f321947a35] Updating instance_info_cache with network_info: [{"id": "642fc1a5-237a-4bde-b28f-69b3f0f46a8a", "address": "fa:16:3e:36:b9:44", "network": {"id": "dc7d4c3b-1f91-4937-a2b0-1084e5c478f0", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1637370260-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "975189ebb3cc4cdb9391880f0c9ba6ba", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b71230ae-e879-4384-88ce-fe64c86fce22", "external-id": "nsx-vlan-transportzone-473", "segmentation_id": 473, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap642fc1a5-23", "ovs_interfaceid": "642fc1a5-237a-4bde-b28f-69b3f0f46a8a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 966.549400] env[61957]: DEBUG nova.network.neutron [None req-ce0965c3-bc7d-4a77-9d32-52f4d059d837 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] [instance: 964c8c9d-0721-4ad3-a146-fce1fd1d6513] Successfully updated port: 57483094-30a7-4f77-8c6e-25dbd1db61c7 {{(pid=61957) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 966.592567] env[61957]: DEBUG oslo_vmware.api [None req-d1aa370b-6bd9-49b7-9a71-30e9793a2856 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Task: {'id': task-1277829, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.537327} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.592777] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-d1aa370b-6bd9-49b7-9a71-30e9793a2856 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore1] 00ecc8c2-398e-4a29-b19f-a2013985b481/00ecc8c2-398e-4a29-b19f-a2013985b481.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 966.593009] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-d1aa370b-6bd9-49b7-9a71-30e9793a2856 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] [instance: 00ecc8c2-398e-4a29-b19f-a2013985b481] Extending root virtual disk to 1048576 {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 966.593276] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0f962815-e568-48dc-8aad-424654bd82d2 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.598587] env[61957]: DEBUG oslo_vmware.api [None req-d1aa370b-6bd9-49b7-9a71-30e9793a2856 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Waiting for the task: (returnval){ [ 966.598587] env[61957]: value = "task-1277832" [ 966.598587] env[61957]: _type = "Task" [ 966.598587] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.606341] env[61957]: DEBUG oslo_vmware.api [None req-d1aa370b-6bd9-49b7-9a71-30e9793a2856 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Task: {'id': task-1277832, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.693113] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277830, 'name': CreateVM_Task, 'duration_secs': 0.406303} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.693301] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d6c5c70d-86de-4dea-8b9e-76f321947a35] Created VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 966.694033] env[61957]: DEBUG oslo_concurrency.lockutils [None req-9e60b82d-4fee-4f83-a3da-b48f03d27cc5 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fa797977-e3da-4f16-8445-e20c97e9e8ae" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 966.694208] env[61957]: DEBUG oslo_concurrency.lockutils [None req-9e60b82d-4fee-4f83-a3da-b48f03d27cc5 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fa797977-e3da-4f16-8445-e20c97e9e8ae" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 966.694568] env[61957]: DEBUG oslo_concurrency.lockutils [None req-9e60b82d-4fee-4f83-a3da-b48f03d27cc5 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/fa797977-e3da-4f16-8445-e20c97e9e8ae" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 966.694822] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8253d451-c328-4033-9ea1-09735c53cc1a {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.699905] env[61957]: DEBUG oslo_vmware.api [None req-9e60b82d-4fee-4f83-a3da-b48f03d27cc5 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Waiting for the task: (returnval){ [ 966.699905] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]526b492b-7f59-2fd4-35cb-febb177da10f" [ 966.699905] env[61957]: _type = "Task" [ 966.699905] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.708339] env[61957]: DEBUG oslo_vmware.api [None req-9e60b82d-4fee-4f83-a3da-b48f03d27cc5 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]526b492b-7f59-2fd4-35cb-febb177da10f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.778788] env[61957]: DEBUG oslo_vmware.api [None req-3141e8c0-2e34-4c8b-a8ac-0f12d351422c tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Task: {'id': task-1277831, 'name': Destroy_Task, 'duration_secs': 0.309127} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.779125] env[61957]: INFO nova.virt.vmwareapi.vm_util [None req-3141e8c0-2e34-4c8b-a8ac-0f12d351422c tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] [instance: 736eabcf-70f8-4e1e-9830-270d9971fc9a] Destroyed the VM [ 966.779373] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-3141e8c0-2e34-4c8b-a8ac-0f12d351422c tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] [instance: 736eabcf-70f8-4e1e-9830-270d9971fc9a] Deleting Snapshot of the VM instance {{(pid=61957) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 966.779651] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-25e74042-a573-401b-bac2-5864a3f55db9 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.786243] env[61957]: DEBUG oslo_vmware.api [None req-3141e8c0-2e34-4c8b-a8ac-0f12d351422c tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Waiting for the task: (returnval){ [ 966.786243] env[61957]: value = "task-1277833" [ 966.786243] env[61957]: _type = "Task" [ 966.786243] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.793840] env[61957]: DEBUG oslo_vmware.api [None req-3141e8c0-2e34-4c8b-a8ac-0f12d351422c tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Task: {'id': task-1277833, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.832146] env[61957]: DEBUG nova.compute.utils [None req-cedd08d9-20a0-42aa-9f1f-33859d216b7f tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Using /dev/sd instead of None {{(pid=61957) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 966.835628] env[61957]: DEBUG nova.compute.manager [None req-cedd08d9-20a0-42aa-9f1f-33859d216b7f tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 98929ed3-d420-45cd-9cde-9738b2ac8251] Allocating IP information in the background. {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 966.835805] env[61957]: DEBUG nova.network.neutron [None req-cedd08d9-20a0-42aa-9f1f-33859d216b7f tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 98929ed3-d420-45cd-9cde-9738b2ac8251] allocate_for_instance() {{(pid=61957) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 966.879281] env[61957]: DEBUG nova.policy [None req-cedd08d9-20a0-42aa-9f1f-33859d216b7f tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '85d4f8bb3aa64e7285aac42d14087ff1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e16f6dce3f0e44fb96de516f17d4c6f5', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61957) authorize /opt/stack/nova/nova/policy.py:203}} [ 967.013888] env[61957]: DEBUG oslo_concurrency.lockutils [req-4b934be2-80f9-43a3-af7d-0ec06ec94157 req-8a15376e-ad65-45e6-b4ac-e66288505a65 service nova] Releasing lock "refresh_cache-d6c5c70d-86de-4dea-8b9e-76f321947a35" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 967.056596] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ce0965c3-bc7d-4a77-9d32-52f4d059d837 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Acquiring lock "refresh_cache-964c8c9d-0721-4ad3-a146-fce1fd1d6513" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 967.056742] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ce0965c3-bc7d-4a77-9d32-52f4d059d837 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Acquired lock "refresh_cache-964c8c9d-0721-4ad3-a146-fce1fd1d6513" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 967.056892] env[61957]: DEBUG nova.network.neutron [None req-ce0965c3-bc7d-4a77-9d32-52f4d059d837 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] [instance: 964c8c9d-0721-4ad3-a146-fce1fd1d6513] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 967.077549] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6d61c91-2a04-4098-a7ce-0dc343f09355 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.087724] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c77c805a-a14e-4648-aa38-fdb4a5274a42 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.121429] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd18e07b-76b6-4066-b4fd-877b54906e3d {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.128678] env[61957]: DEBUG oslo_vmware.api [None req-d1aa370b-6bd9-49b7-9a71-30e9793a2856 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Task: {'id': task-1277832, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.218229} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.131281] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-d1aa370b-6bd9-49b7-9a71-30e9793a2856 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] [instance: 00ecc8c2-398e-4a29-b19f-a2013985b481] Extended root virtual disk {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 967.133011] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9d25421-4d4b-49b3-9d71-835039dd662f {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.136436] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8b90157-3f81-4055-8eae-517fb5819c26 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.142038] env[61957]: DEBUG nova.network.neutron [None req-cedd08d9-20a0-42aa-9f1f-33859d216b7f tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 98929ed3-d420-45cd-9cde-9738b2ac8251] Successfully created port: bcb69e39-31e7-4ef0-837a-fa6d08bd0449 {{(pid=61957) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 967.155545] env[61957]: DEBUG nova.compute.provider_tree [None req-5dd00032-4725-4b73-a1a7-7ef95ae8be93 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 967.173311] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-d1aa370b-6bd9-49b7-9a71-30e9793a2856 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] [instance: 00ecc8c2-398e-4a29-b19f-a2013985b481] Reconfiguring VM instance instance-00000056 to attach disk [datastore1] 00ecc8c2-398e-4a29-b19f-a2013985b481/00ecc8c2-398e-4a29-b19f-a2013985b481.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 967.176884] env[61957]: DEBUG nova.scheduler.client.report [None req-5dd00032-4725-4b73-a1a7-7ef95ae8be93 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 967.178063] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d0385370-0f3f-48b8-a404-1d20d1f0a777 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.198226] env[61957]: DEBUG oslo_vmware.api [None req-d1aa370b-6bd9-49b7-9a71-30e9793a2856 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Waiting for the task: (returnval){ [ 967.198226] env[61957]: value = "task-1277834" [ 967.198226] env[61957]: _type = "Task" [ 967.198226] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.210018] env[61957]: DEBUG oslo_vmware.api [None req-d1aa370b-6bd9-49b7-9a71-30e9793a2856 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Task: {'id': task-1277834, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.213423] env[61957]: DEBUG oslo_concurrency.lockutils [None req-9e60b82d-4fee-4f83-a3da-b48f03d27cc5 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fa797977-e3da-4f16-8445-e20c97e9e8ae" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 967.213689] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-9e60b82d-4fee-4f83-a3da-b48f03d27cc5 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: d6c5c70d-86de-4dea-8b9e-76f321947a35] Processing image fa797977-e3da-4f16-8445-e20c97e9e8ae {{(pid=61957) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 967.213932] env[61957]: DEBUG oslo_concurrency.lockutils [None req-9e60b82d-4fee-4f83-a3da-b48f03d27cc5 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fa797977-e3da-4f16-8445-e20c97e9e8ae/fa797977-e3da-4f16-8445-e20c97e9e8ae.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 967.214092] env[61957]: DEBUG oslo_concurrency.lockutils [None req-9e60b82d-4fee-4f83-a3da-b48f03d27cc5 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fa797977-e3da-4f16-8445-e20c97e9e8ae/fa797977-e3da-4f16-8445-e20c97e9e8ae.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 967.214278] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-9e60b82d-4fee-4f83-a3da-b48f03d27cc5 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 967.214522] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d74d667d-3de3-4093-8d22-68a81f898184 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.223835] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-9e60b82d-4fee-4f83-a3da-b48f03d27cc5 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 967.224452] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-9e60b82d-4fee-4f83-a3da-b48f03d27cc5 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61957) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 967.224881] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c85286ff-badb-4f3f-b778-9ef749d387a4 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.232320] env[61957]: DEBUG oslo_vmware.api [None req-9e60b82d-4fee-4f83-a3da-b48f03d27cc5 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Waiting for the task: (returnval){ [ 967.232320] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]523d3667-c41e-3029-0715-20231513f3e1" [ 967.232320] env[61957]: _type = "Task" [ 967.232320] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.240355] env[61957]: DEBUG oslo_vmware.api [None req-9e60b82d-4fee-4f83-a3da-b48f03d27cc5 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]523d3667-c41e-3029-0715-20231513f3e1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.296157] env[61957]: DEBUG oslo_vmware.api [None req-3141e8c0-2e34-4c8b-a8ac-0f12d351422c tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Task: {'id': task-1277833, 'name': RemoveSnapshot_Task, 'duration_secs': 0.39379} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.296341] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-3141e8c0-2e34-4c8b-a8ac-0f12d351422c tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] [instance: 736eabcf-70f8-4e1e-9830-270d9971fc9a] Deleted Snapshot of the VM instance {{(pid=61957) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 967.296605] env[61957]: DEBUG nova.compute.manager [None req-3141e8c0-2e34-4c8b-a8ac-0f12d351422c tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] [instance: 736eabcf-70f8-4e1e-9830-270d9971fc9a] Checking state {{(pid=61957) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 967.297386] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35dee132-57c2-4fc9-9f5a-8156e6857d95 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.338858] env[61957]: DEBUG nova.compute.manager [None req-cedd08d9-20a0-42aa-9f1f-33859d216b7f tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 98929ed3-d420-45cd-9cde-9738b2ac8251] Start building block device mappings for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 967.594952] env[61957]: DEBUG nova.network.neutron [None req-ce0965c3-bc7d-4a77-9d32-52f4d059d837 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] [instance: 964c8c9d-0721-4ad3-a146-fce1fd1d6513] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 967.694574] env[61957]: DEBUG oslo_concurrency.lockutils [None req-5dd00032-4725-4b73-a1a7-7ef95ae8be93 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.366s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 967.697392] env[61957]: DEBUG oslo_concurrency.lockutils [None req-7c0e7a82-e685-47c5-90d8-2de75a11cf75 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 13.518s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 967.697684] env[61957]: DEBUG nova.objects.instance [None req-7c0e7a82-e685-47c5-90d8-2de75a11cf75 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Lazy-loading 'resources' on Instance uuid 2e32aae5-edd7-402f-98ad-75a93d26f7a1 {{(pid=61957) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 967.709026] env[61957]: DEBUG oslo_vmware.api [None req-d1aa370b-6bd9-49b7-9a71-30e9793a2856 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Task: {'id': task-1277834, 'name': ReconfigVM_Task, 'duration_secs': 0.431136} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.709818] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-d1aa370b-6bd9-49b7-9a71-30e9793a2856 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] [instance: 00ecc8c2-398e-4a29-b19f-a2013985b481] Reconfigured VM instance instance-00000056 to attach disk [datastore1] 00ecc8c2-398e-4a29-b19f-a2013985b481/00ecc8c2-398e-4a29-b19f-a2013985b481.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 967.710533] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-707fc2ec-8b67-43a9-b1b7-e6b74e0f12af {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.716968] env[61957]: DEBUG oslo_vmware.api [None req-d1aa370b-6bd9-49b7-9a71-30e9793a2856 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Waiting for the task: (returnval){ [ 967.716968] env[61957]: value = "task-1277835" [ 967.716968] env[61957]: _type = "Task" [ 967.716968] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.720122] env[61957]: INFO nova.scheduler.client.report [None req-5dd00032-4725-4b73-a1a7-7ef95ae8be93 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Deleted allocations for instance 50e09db7-9f8f-452b-8232-bd7473e9e63a [ 967.731585] env[61957]: DEBUG oslo_vmware.api [None req-d1aa370b-6bd9-49b7-9a71-30e9793a2856 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Task: {'id': task-1277835, 'name': Rename_Task} progress is 6%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.741410] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-9e60b82d-4fee-4f83-a3da-b48f03d27cc5 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: d6c5c70d-86de-4dea-8b9e-76f321947a35] Preparing fetch location {{(pid=61957) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 967.741536] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-9e60b82d-4fee-4f83-a3da-b48f03d27cc5 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: d6c5c70d-86de-4dea-8b9e-76f321947a35] Fetch image to [datastore1] OSTACK_IMG_25d5b1c0-08b2-417d-b7dc-5ba4b7379530/OSTACK_IMG_25d5b1c0-08b2-417d-b7dc-5ba4b7379530.vmdk {{(pid=61957) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 967.742796] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-9e60b82d-4fee-4f83-a3da-b48f03d27cc5 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: d6c5c70d-86de-4dea-8b9e-76f321947a35] Downloading stream optimized image fa797977-e3da-4f16-8445-e20c97e9e8ae to [datastore1] OSTACK_IMG_25d5b1c0-08b2-417d-b7dc-5ba4b7379530/OSTACK_IMG_25d5b1c0-08b2-417d-b7dc-5ba4b7379530.vmdk on the data store datastore1 as vApp {{(pid=61957) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 967.742796] env[61957]: DEBUG nova.virt.vmwareapi.images [None req-9e60b82d-4fee-4f83-a3da-b48f03d27cc5 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: d6c5c70d-86de-4dea-8b9e-76f321947a35] Downloading image file data fa797977-e3da-4f16-8445-e20c97e9e8ae to the ESX as VM named 'OSTACK_IMG_25d5b1c0-08b2-417d-b7dc-5ba4b7379530' {{(pid=61957) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 967.756855] env[61957]: DEBUG nova.compute.manager [req-0437497e-ae8a-4e71-8948-a5b3d912035d req-090c5110-879d-4658-8f59-0980302ca99c service nova] [instance: 964c8c9d-0721-4ad3-a146-fce1fd1d6513] Received event network-vif-plugged-57483094-30a7-4f77-8c6e-25dbd1db61c7 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 967.757117] env[61957]: DEBUG oslo_concurrency.lockutils [req-0437497e-ae8a-4e71-8948-a5b3d912035d req-090c5110-879d-4658-8f59-0980302ca99c service nova] Acquiring lock "964c8c9d-0721-4ad3-a146-fce1fd1d6513-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 967.757346] env[61957]: DEBUG oslo_concurrency.lockutils [req-0437497e-ae8a-4e71-8948-a5b3d912035d req-090c5110-879d-4658-8f59-0980302ca99c service nova] Lock "964c8c9d-0721-4ad3-a146-fce1fd1d6513-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 967.757540] env[61957]: DEBUG oslo_concurrency.lockutils [req-0437497e-ae8a-4e71-8948-a5b3d912035d req-090c5110-879d-4658-8f59-0980302ca99c service nova] Lock "964c8c9d-0721-4ad3-a146-fce1fd1d6513-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 967.757710] env[61957]: DEBUG nova.compute.manager [req-0437497e-ae8a-4e71-8948-a5b3d912035d req-090c5110-879d-4658-8f59-0980302ca99c service nova] [instance: 964c8c9d-0721-4ad3-a146-fce1fd1d6513] No waiting events found dispatching network-vif-plugged-57483094-30a7-4f77-8c6e-25dbd1db61c7 {{(pid=61957) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 967.757897] env[61957]: WARNING nova.compute.manager [req-0437497e-ae8a-4e71-8948-a5b3d912035d req-090c5110-879d-4658-8f59-0980302ca99c service nova] [instance: 964c8c9d-0721-4ad3-a146-fce1fd1d6513] Received unexpected event network-vif-plugged-57483094-30a7-4f77-8c6e-25dbd1db61c7 for instance with vm_state building and task_state spawning. [ 967.758086] env[61957]: DEBUG nova.compute.manager [req-0437497e-ae8a-4e71-8948-a5b3d912035d req-090c5110-879d-4658-8f59-0980302ca99c service nova] [instance: 964c8c9d-0721-4ad3-a146-fce1fd1d6513] Received event network-changed-57483094-30a7-4f77-8c6e-25dbd1db61c7 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 967.758260] env[61957]: DEBUG nova.compute.manager [req-0437497e-ae8a-4e71-8948-a5b3d912035d req-090c5110-879d-4658-8f59-0980302ca99c service nova] [instance: 964c8c9d-0721-4ad3-a146-fce1fd1d6513] Refreshing instance network info cache due to event network-changed-57483094-30a7-4f77-8c6e-25dbd1db61c7. {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 967.758453] env[61957]: DEBUG oslo_concurrency.lockutils [req-0437497e-ae8a-4e71-8948-a5b3d912035d req-090c5110-879d-4658-8f59-0980302ca99c service nova] Acquiring lock "refresh_cache-964c8c9d-0721-4ad3-a146-fce1fd1d6513" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 967.768894] env[61957]: DEBUG nova.network.neutron [None req-ce0965c3-bc7d-4a77-9d32-52f4d059d837 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] [instance: 964c8c9d-0721-4ad3-a146-fce1fd1d6513] Updating instance_info_cache with network_info: [{"id": "57483094-30a7-4f77-8c6e-25dbd1db61c7", "address": "fa:16:3e:55:82:02", "network": {"id": "82fd2357-53b6-4903-963e-a2866736643e", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-612681178-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1673c854272f4b14a623a73b8f4b687b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6eb7e3e9-5cc2-40f1-a6eb-f70f06531667", "external-id": "nsx-vlan-transportzone-938", "segmentation_id": 938, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap57483094-30", "ovs_interfaceid": "57483094-30a7-4f77-8c6e-25dbd1db61c7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 967.809961] env[61957]: INFO nova.compute.manager [None req-3141e8c0-2e34-4c8b-a8ac-0f12d351422c tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] [instance: 736eabcf-70f8-4e1e-9830-270d9971fc9a] Shelve offloading [ 967.813767] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-3141e8c0-2e34-4c8b-a8ac-0f12d351422c tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] [instance: 736eabcf-70f8-4e1e-9830-270d9971fc9a] Powering off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 967.814057] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d8fceb2b-c5c2-4b48-a3ba-c95a259078f2 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.822182] env[61957]: DEBUG oslo_vmware.api [None req-3141e8c0-2e34-4c8b-a8ac-0f12d351422c tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Waiting for the task: (returnval){ [ 967.822182] env[61957]: value = "task-1277836" [ 967.822182] env[61957]: _type = "Task" [ 967.822182] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.827122] env[61957]: DEBUG oslo_vmware.rw_handles [None req-9e60b82d-4fee-4f83-a3da-b48f03d27cc5 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 967.827122] env[61957]: value = "resgroup-9" [ 967.827122] env[61957]: _type = "ResourcePool" [ 967.827122] env[61957]: }. {{(pid=61957) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 967.827564] env[61957]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-f03dc476-c110-4660-b801-7d4bbe387f18 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.849886] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-3141e8c0-2e34-4c8b-a8ac-0f12d351422c tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] [instance: 736eabcf-70f8-4e1e-9830-270d9971fc9a] VM already powered off {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 967.850103] env[61957]: DEBUG nova.compute.manager [None req-3141e8c0-2e34-4c8b-a8ac-0f12d351422c tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] [instance: 736eabcf-70f8-4e1e-9830-270d9971fc9a] Checking state {{(pid=61957) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 967.850869] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3235ec3f-1b75-4298-b7ef-09072f98f491 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.854634] env[61957]: DEBUG oslo_vmware.rw_handles [None req-9e60b82d-4fee-4f83-a3da-b48f03d27cc5 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Lease: (returnval){ [ 967.854634] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]528d7146-c376-c017-472d-a8a5a89a7e92" [ 967.854634] env[61957]: _type = "HttpNfcLease" [ 967.854634] env[61957]: } obtained for vApp import into resource pool (val){ [ 967.854634] env[61957]: value = "resgroup-9" [ 967.854634] env[61957]: _type = "ResourcePool" [ 967.854634] env[61957]: }. {{(pid=61957) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 967.855051] env[61957]: DEBUG oslo_vmware.api [None req-9e60b82d-4fee-4f83-a3da-b48f03d27cc5 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Waiting for the lease: (returnval){ [ 967.855051] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]528d7146-c376-c017-472d-a8a5a89a7e92" [ 967.855051] env[61957]: _type = "HttpNfcLease" [ 967.855051] env[61957]: } to be ready. {{(pid=61957) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 967.859765] env[61957]: DEBUG oslo_concurrency.lockutils [None req-3141e8c0-2e34-4c8b-a8ac-0f12d351422c tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Acquiring lock "refresh_cache-736eabcf-70f8-4e1e-9830-270d9971fc9a" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 967.859933] env[61957]: DEBUG oslo_concurrency.lockutils [None req-3141e8c0-2e34-4c8b-a8ac-0f12d351422c tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Acquired lock "refresh_cache-736eabcf-70f8-4e1e-9830-270d9971fc9a" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 967.860198] env[61957]: DEBUG nova.network.neutron [None req-3141e8c0-2e34-4c8b-a8ac-0f12d351422c tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] [instance: 736eabcf-70f8-4e1e-9830-270d9971fc9a] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 967.864481] env[61957]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 967.864481] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]528d7146-c376-c017-472d-a8a5a89a7e92" [ 967.864481] env[61957]: _type = "HttpNfcLease" [ 967.864481] env[61957]: } is initializing. {{(pid=61957) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 968.231982] env[61957]: DEBUG oslo_vmware.api [None req-d1aa370b-6bd9-49b7-9a71-30e9793a2856 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Task: {'id': task-1277835, 'name': Rename_Task, 'duration_secs': 0.127974} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.232515] env[61957]: DEBUG oslo_concurrency.lockutils [None req-5dd00032-4725-4b73-a1a7-7ef95ae8be93 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Lock "50e09db7-9f8f-452b-8232-bd7473e9e63a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 17.925s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 968.236365] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-d1aa370b-6bd9-49b7-9a71-30e9793a2856 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] [instance: 00ecc8c2-398e-4a29-b19f-a2013985b481] Powering on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 968.236617] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-74da87a0-827a-49bd-8bac-ae99069da619 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.243607] env[61957]: DEBUG oslo_vmware.api [None req-d1aa370b-6bd9-49b7-9a71-30e9793a2856 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Waiting for the task: (returnval){ [ 968.243607] env[61957]: value = "task-1277838" [ 968.243607] env[61957]: _type = "Task" [ 968.243607] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.254632] env[61957]: DEBUG oslo_vmware.api [None req-d1aa370b-6bd9-49b7-9a71-30e9793a2856 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Task: {'id': task-1277838, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.273102] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ce0965c3-bc7d-4a77-9d32-52f4d059d837 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Releasing lock "refresh_cache-964c8c9d-0721-4ad3-a146-fce1fd1d6513" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 968.273427] env[61957]: DEBUG nova.compute.manager [None req-ce0965c3-bc7d-4a77-9d32-52f4d059d837 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] [instance: 964c8c9d-0721-4ad3-a146-fce1fd1d6513] Instance network_info: |[{"id": "57483094-30a7-4f77-8c6e-25dbd1db61c7", "address": "fa:16:3e:55:82:02", "network": {"id": "82fd2357-53b6-4903-963e-a2866736643e", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-612681178-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1673c854272f4b14a623a73b8f4b687b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6eb7e3e9-5cc2-40f1-a6eb-f70f06531667", "external-id": "nsx-vlan-transportzone-938", "segmentation_id": 938, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap57483094-30", "ovs_interfaceid": "57483094-30a7-4f77-8c6e-25dbd1db61c7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 968.273736] env[61957]: DEBUG oslo_concurrency.lockutils [req-0437497e-ae8a-4e71-8948-a5b3d912035d req-090c5110-879d-4658-8f59-0980302ca99c service nova] Acquired lock "refresh_cache-964c8c9d-0721-4ad3-a146-fce1fd1d6513" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 968.273916] env[61957]: DEBUG nova.network.neutron [req-0437497e-ae8a-4e71-8948-a5b3d912035d req-090c5110-879d-4658-8f59-0980302ca99c service nova] [instance: 964c8c9d-0721-4ad3-a146-fce1fd1d6513] Refreshing network info cache for port 57483094-30a7-4f77-8c6e-25dbd1db61c7 {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 968.275332] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-ce0965c3-bc7d-4a77-9d32-52f4d059d837 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] [instance: 964c8c9d-0721-4ad3-a146-fce1fd1d6513] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:55:82:02', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '6eb7e3e9-5cc2-40f1-a6eb-f70f06531667', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '57483094-30a7-4f77-8c6e-25dbd1db61c7', 'vif_model': 'vmxnet3'}] {{(pid=61957) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 968.284176] env[61957]: DEBUG oslo.service.loopingcall [None req-ce0965c3-bc7d-4a77-9d32-52f4d059d837 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 968.287966] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 964c8c9d-0721-4ad3-a146-fce1fd1d6513] Creating VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 968.288722] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9eeb92dc-d966-448f-9c5a-5a6228f6676f {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.312703] env[61957]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 968.312703] env[61957]: value = "task-1277839" [ 968.312703] env[61957]: _type = "Task" [ 968.312703] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.322563] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277839, 'name': CreateVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.357433] env[61957]: DEBUG nova.compute.manager [None req-cedd08d9-20a0-42aa-9f1f-33859d216b7f tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 98929ed3-d420-45cd-9cde-9738b2ac8251] Start spawning the instance on the hypervisor. {{(pid=61957) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 968.369476] env[61957]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 968.369476] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]528d7146-c376-c017-472d-a8a5a89a7e92" [ 968.369476] env[61957]: _type = "HttpNfcLease" [ 968.369476] env[61957]: } is ready. {{(pid=61957) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 968.369841] env[61957]: DEBUG oslo_vmware.rw_handles [None req-9e60b82d-4fee-4f83-a3da-b48f03d27cc5 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 968.369841] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]528d7146-c376-c017-472d-a8a5a89a7e92" [ 968.369841] env[61957]: _type = "HttpNfcLease" [ 968.369841] env[61957]: }. {{(pid=61957) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 968.371164] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3c89cfc-8b84-4aa2-bd54-2ee6a47b13a7 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.383658] env[61957]: DEBUG oslo_vmware.rw_handles [None req-9e60b82d-4fee-4f83-a3da-b48f03d27cc5 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/529cb135-3cd2-7b9e-3c08-7c200fc40980/disk-0.vmdk from lease info. {{(pid=61957) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 968.383906] env[61957]: DEBUG oslo_vmware.rw_handles [None req-9e60b82d-4fee-4f83-a3da-b48f03d27cc5 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Creating HTTP connection to write to file with size = 21334016 and URL = https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/529cb135-3cd2-7b9e-3c08-7c200fc40980/disk-0.vmdk. {{(pid=61957) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 968.387301] env[61957]: DEBUG nova.virt.hardware [None req-cedd08d9-20a0-42aa-9f1f-33859d216b7f tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T17:22:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T17:21:50Z,direct_url=,disk_format='vmdk',id=11c76a2c-f705-470a-ba9d-4657858bab38,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='81d3096d0f094373913fc3dc8f5c3c6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T17:21:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 968.387998] env[61957]: DEBUG nova.virt.hardware [None req-cedd08d9-20a0-42aa-9f1f-33859d216b7f tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Flavor limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 968.387998] env[61957]: DEBUG nova.virt.hardware [None req-cedd08d9-20a0-42aa-9f1f-33859d216b7f tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Image limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 968.387998] env[61957]: DEBUG nova.virt.hardware [None req-cedd08d9-20a0-42aa-9f1f-33859d216b7f tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Flavor pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 968.387998] env[61957]: DEBUG nova.virt.hardware [None req-cedd08d9-20a0-42aa-9f1f-33859d216b7f tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Image pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 968.388397] env[61957]: DEBUG nova.virt.hardware [None req-cedd08d9-20a0-42aa-9f1f-33859d216b7f tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 968.388446] env[61957]: DEBUG nova.virt.hardware [None req-cedd08d9-20a0-42aa-9f1f-33859d216b7f tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 968.388611] env[61957]: DEBUG nova.virt.hardware [None req-cedd08d9-20a0-42aa-9f1f-33859d216b7f tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 968.388770] env[61957]: DEBUG nova.virt.hardware [None req-cedd08d9-20a0-42aa-9f1f-33859d216b7f tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Got 1 possible topologies {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 968.388934] env[61957]: DEBUG nova.virt.hardware [None req-cedd08d9-20a0-42aa-9f1f-33859d216b7f tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 968.389137] env[61957]: DEBUG nova.virt.hardware [None req-cedd08d9-20a0-42aa-9f1f-33859d216b7f tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 968.389950] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b20e343-761b-47ba-b135-664fa2358863 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.465873] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50956a0e-1a5d-4431-80c6-92b080b50ff7 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.473948] env[61957]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-897fc476-5f0b-4319-9c6f-35d6006bd218 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.551724] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abf9daee-3f50-4f94-8110-7b59417038fb {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.560330] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b186506-75f0-4126-83d6-a0341852e856 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.595047] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebf054ea-5943-4f1f-afc7-02089f11f892 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.605638] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d46619d7-9690-4e68-91c0-f6d383a46e8a {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.620155] env[61957]: DEBUG nova.compute.provider_tree [None req-7c0e7a82-e685-47c5-90d8-2de75a11cf75 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 968.759369] env[61957]: DEBUG oslo_vmware.api [None req-d1aa370b-6bd9-49b7-9a71-30e9793a2856 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Task: {'id': task-1277838, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.825478] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277839, 'name': CreateVM_Task, 'duration_secs': 0.333774} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.827463] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 964c8c9d-0721-4ad3-a146-fce1fd1d6513] Created VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 968.828291] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ce0965c3-bc7d-4a77-9d32-52f4d059d837 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 968.828518] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ce0965c3-bc7d-4a77-9d32-52f4d059d837 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Acquired lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 968.828861] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ce0965c3-bc7d-4a77-9d32-52f4d059d837 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 968.829178] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d91a7e11-832e-4bad-8040-aa82ec4f8657 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.834687] env[61957]: DEBUG oslo_vmware.api [None req-ce0965c3-bc7d-4a77-9d32-52f4d059d837 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Waiting for the task: (returnval){ [ 968.834687] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52242ee7-a395-5ff1-37ac-eba0a06c8f7e" [ 968.834687] env[61957]: _type = "Task" [ 968.834687] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.840765] env[61957]: DEBUG nova.network.neutron [None req-3141e8c0-2e34-4c8b-a8ac-0f12d351422c tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] [instance: 736eabcf-70f8-4e1e-9830-270d9971fc9a] Updating instance_info_cache with network_info: [{"id": "95763a65-6383-409f-a9d6-1f37aaf44d34", "address": "fa:16:3e:64:07:83", "network": {"id": "b303ad01-ef9b-4119-9722-3be27ffb30bd", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-96064402-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.145", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "55742b1d4fbb401ea61f990ecaa2def6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a8b99a46-3e7f-4ef1-9e45-58e6cd17f210", "external-id": "nsx-vlan-transportzone-704", "segmentation_id": 704, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap95763a65-63", "ovs_interfaceid": "95763a65-6383-409f-a9d6-1f37aaf44d34", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 968.845120] env[61957]: DEBUG nova.network.neutron [None req-cedd08d9-20a0-42aa-9f1f-33859d216b7f tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 98929ed3-d420-45cd-9cde-9738b2ac8251] Successfully updated port: bcb69e39-31e7-4ef0-837a-fa6d08bd0449 {{(pid=61957) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 968.852557] env[61957]: DEBUG oslo_vmware.api [None req-ce0965c3-bc7d-4a77-9d32-52f4d059d837 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52242ee7-a395-5ff1-37ac-eba0a06c8f7e, 'name': SearchDatastore_Task} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.855693] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ce0965c3-bc7d-4a77-9d32-52f4d059d837 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Releasing lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 968.856078] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-ce0965c3-bc7d-4a77-9d32-52f4d059d837 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] [instance: 964c8c9d-0721-4ad3-a146-fce1fd1d6513] Processing image 11c76a2c-f705-470a-ba9d-4657858bab38 {{(pid=61957) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 968.856320] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ce0965c3-bc7d-4a77-9d32-52f4d059d837 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 968.856480] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ce0965c3-bc7d-4a77-9d32-52f4d059d837 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Acquired lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 968.856663] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-ce0965c3-bc7d-4a77-9d32-52f4d059d837 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 968.858063] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2e469580-fb5e-431c-a934-cbb6cb9f6a5c {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.868544] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-ce0965c3-bc7d-4a77-9d32-52f4d059d837 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 968.868784] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-ce0965c3-bc7d-4a77-9d32-52f4d059d837 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61957) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 968.869862] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-24ce4dac-8794-40b2-b2f3-e5b0f8518488 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.879391] env[61957]: DEBUG oslo_vmware.api [None req-ce0965c3-bc7d-4a77-9d32-52f4d059d837 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Waiting for the task: (returnval){ [ 968.879391] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52f1c8b6-0da5-7024-8103-8023cd4b4ebb" [ 968.879391] env[61957]: _type = "Task" [ 968.879391] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.887542] env[61957]: DEBUG oslo_vmware.api [None req-ce0965c3-bc7d-4a77-9d32-52f4d059d837 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52f1c8b6-0da5-7024-8103-8023cd4b4ebb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.050302] env[61957]: DEBUG nova.network.neutron [req-0437497e-ae8a-4e71-8948-a5b3d912035d req-090c5110-879d-4658-8f59-0980302ca99c service nova] [instance: 964c8c9d-0721-4ad3-a146-fce1fd1d6513] Updated VIF entry in instance network info cache for port 57483094-30a7-4f77-8c6e-25dbd1db61c7. {{(pid=61957) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 969.050751] env[61957]: DEBUG nova.network.neutron [req-0437497e-ae8a-4e71-8948-a5b3d912035d req-090c5110-879d-4658-8f59-0980302ca99c service nova] [instance: 964c8c9d-0721-4ad3-a146-fce1fd1d6513] Updating instance_info_cache with network_info: [{"id": "57483094-30a7-4f77-8c6e-25dbd1db61c7", "address": "fa:16:3e:55:82:02", "network": {"id": "82fd2357-53b6-4903-963e-a2866736643e", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-612681178-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1673c854272f4b14a623a73b8f4b687b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6eb7e3e9-5cc2-40f1-a6eb-f70f06531667", "external-id": "nsx-vlan-transportzone-938", "segmentation_id": 938, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap57483094-30", "ovs_interfaceid": "57483094-30a7-4f77-8c6e-25dbd1db61c7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 969.124057] env[61957]: DEBUG nova.scheduler.client.report [None req-7c0e7a82-e685-47c5-90d8-2de75a11cf75 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 969.173078] env[61957]: DEBUG oslo_concurrency.lockutils [None req-6a0c88f8-9f89-4186-9d19-09e1aa8c2fe1 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Acquiring lock "526ff179-62a6-4763-ab25-797617c4ed57" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 969.173388] env[61957]: DEBUG oslo_concurrency.lockutils [None req-6a0c88f8-9f89-4186-9d19-09e1aa8c2fe1 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Lock "526ff179-62a6-4763-ab25-797617c4ed57" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 969.210294] env[61957]: DEBUG oslo_vmware.rw_handles [None req-9e60b82d-4fee-4f83-a3da-b48f03d27cc5 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Completed reading data from the image iterator. {{(pid=61957) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 969.210591] env[61957]: DEBUG oslo_vmware.rw_handles [None req-9e60b82d-4fee-4f83-a3da-b48f03d27cc5 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/529cb135-3cd2-7b9e-3c08-7c200fc40980/disk-0.vmdk. {{(pid=61957) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 969.211675] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da4eac12-b6f4-42ed-a94d-9a0c99f66aec {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.219019] env[61957]: DEBUG oslo_vmware.rw_handles [None req-9e60b82d-4fee-4f83-a3da-b48f03d27cc5 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/529cb135-3cd2-7b9e-3c08-7c200fc40980/disk-0.vmdk is in state: ready. {{(pid=61957) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 969.219019] env[61957]: DEBUG oslo_vmware.rw_handles [None req-9e60b82d-4fee-4f83-a3da-b48f03d27cc5 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Releasing lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/529cb135-3cd2-7b9e-3c08-7c200fc40980/disk-0.vmdk. {{(pid=61957) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 969.219148] env[61957]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-32b6baa5-e99b-4b11-af1e-59e4034c7abe {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.255287] env[61957]: DEBUG oslo_vmware.api [None req-d1aa370b-6bd9-49b7-9a71-30e9793a2856 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Task: {'id': task-1277838, 'name': PowerOnVM_Task, 'duration_secs': 0.540788} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.255539] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-d1aa370b-6bd9-49b7-9a71-30e9793a2856 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] [instance: 00ecc8c2-398e-4a29-b19f-a2013985b481] Powered on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 969.256089] env[61957]: INFO nova.compute.manager [None req-d1aa370b-6bd9-49b7-9a71-30e9793a2856 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] [instance: 00ecc8c2-398e-4a29-b19f-a2013985b481] Took 8.04 seconds to spawn the instance on the hypervisor. [ 969.256089] env[61957]: DEBUG nova.compute.manager [None req-d1aa370b-6bd9-49b7-9a71-30e9793a2856 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] [instance: 00ecc8c2-398e-4a29-b19f-a2013985b481] Checking state {{(pid=61957) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 969.256792] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e6f221b-91c4-4063-92da-3d6897c2edce {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.343576] env[61957]: DEBUG oslo_concurrency.lockutils [None req-3141e8c0-2e34-4c8b-a8ac-0f12d351422c tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Releasing lock "refresh_cache-736eabcf-70f8-4e1e-9830-270d9971fc9a" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 969.354408] env[61957]: DEBUG oslo_concurrency.lockutils [None req-cedd08d9-20a0-42aa-9f1f-33859d216b7f tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Acquiring lock "refresh_cache-98929ed3-d420-45cd-9cde-9738b2ac8251" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 969.354869] env[61957]: DEBUG oslo_concurrency.lockutils [None req-cedd08d9-20a0-42aa-9f1f-33859d216b7f tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Acquired lock "refresh_cache-98929ed3-d420-45cd-9cde-9738b2ac8251" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 969.354869] env[61957]: DEBUG nova.network.neutron [None req-cedd08d9-20a0-42aa-9f1f-33859d216b7f tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 98929ed3-d420-45cd-9cde-9738b2ac8251] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 969.372363] env[61957]: DEBUG oslo_vmware.rw_handles [None req-9e60b82d-4fee-4f83-a3da-b48f03d27cc5 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Closed VMDK write handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/529cb135-3cd2-7b9e-3c08-7c200fc40980/disk-0.vmdk. {{(pid=61957) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 969.372657] env[61957]: INFO nova.virt.vmwareapi.images [None req-9e60b82d-4fee-4f83-a3da-b48f03d27cc5 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: d6c5c70d-86de-4dea-8b9e-76f321947a35] Downloaded image file data fa797977-e3da-4f16-8445-e20c97e9e8ae [ 969.373641] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-754b7619-6bd2-4f53-aaf6-d0b392df6664 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.393050] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-087e027f-484d-4e6a-9ca8-3ff5331d8b28 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.399934] env[61957]: DEBUG oslo_vmware.api [None req-ce0965c3-bc7d-4a77-9d32-52f4d059d837 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52f1c8b6-0da5-7024-8103-8023cd4b4ebb, 'name': SearchDatastore_Task, 'duration_secs': 0.009398} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.400753] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d922b69f-1e39-4bf6-b7ea-d27c48f9058f {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.405956] env[61957]: DEBUG oslo_vmware.api [None req-ce0965c3-bc7d-4a77-9d32-52f4d059d837 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Waiting for the task: (returnval){ [ 969.405956] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52cf31fe-dcc3-770e-848e-7c860da468d0" [ 969.405956] env[61957]: _type = "Task" [ 969.405956] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.414170] env[61957]: DEBUG oslo_vmware.api [None req-ce0965c3-bc7d-4a77-9d32-52f4d059d837 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52cf31fe-dcc3-770e-848e-7c860da468d0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.419549] env[61957]: INFO nova.virt.vmwareapi.images [None req-9e60b82d-4fee-4f83-a3da-b48f03d27cc5 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: d6c5c70d-86de-4dea-8b9e-76f321947a35] The imported VM was unregistered [ 969.421900] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-9e60b82d-4fee-4f83-a3da-b48f03d27cc5 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: d6c5c70d-86de-4dea-8b9e-76f321947a35] Caching image {{(pid=61957) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 969.422162] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-9e60b82d-4fee-4f83-a3da-b48f03d27cc5 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Creating directory with path [datastore1] devstack-image-cache_base/fa797977-e3da-4f16-8445-e20c97e9e8ae {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 969.422449] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d7c1d578-e8d0-40fb-9724-fc59ff9e89a0 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.440065] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-9e60b82d-4fee-4f83-a3da-b48f03d27cc5 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Created directory with path [datastore1] devstack-image-cache_base/fa797977-e3da-4f16-8445-e20c97e9e8ae {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 969.440270] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-9e60b82d-4fee-4f83-a3da-b48f03d27cc5 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Moving virtual disk from [datastore1] OSTACK_IMG_25d5b1c0-08b2-417d-b7dc-5ba4b7379530/OSTACK_IMG_25d5b1c0-08b2-417d-b7dc-5ba4b7379530.vmdk to [datastore1] devstack-image-cache_base/fa797977-e3da-4f16-8445-e20c97e9e8ae/fa797977-e3da-4f16-8445-e20c97e9e8ae.vmdk. {{(pid=61957) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 969.440523] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-c96c9c05-d9e1-4f60-81a0-6840d056e4ae {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.447805] env[61957]: DEBUG oslo_vmware.api [None req-9e60b82d-4fee-4f83-a3da-b48f03d27cc5 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Waiting for the task: (returnval){ [ 969.447805] env[61957]: value = "task-1277841" [ 969.447805] env[61957]: _type = "Task" [ 969.447805] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.458135] env[61957]: DEBUG oslo_vmware.api [None req-9e60b82d-4fee-4f83-a3da-b48f03d27cc5 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': task-1277841, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.554248] env[61957]: DEBUG oslo_concurrency.lockutils [req-0437497e-ae8a-4e71-8948-a5b3d912035d req-090c5110-879d-4658-8f59-0980302ca99c service nova] Releasing lock "refresh_cache-964c8c9d-0721-4ad3-a146-fce1fd1d6513" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 969.630302] env[61957]: DEBUG oslo_concurrency.lockutils [None req-7c0e7a82-e685-47c5-90d8-2de75a11cf75 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.933s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 969.632702] env[61957]: DEBUG oslo_concurrency.lockutils [None req-5145e111-afa1-43c9-8275-d20aceeb3d2b tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.213s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 969.635115] env[61957]: INFO nova.compute.claims [None req-5145e111-afa1-43c9-8275-d20aceeb3d2b tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 6f45c7c8-de07-4d21-bccc-e4afe3c8ad3d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 969.678326] env[61957]: INFO nova.compute.manager [None req-6a0c88f8-9f89-4186-9d19-09e1aa8c2fe1 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] [instance: 526ff179-62a6-4763-ab25-797617c4ed57] Detaching volume 6c0df615-4d1a-45c6-9f48-ab1813e515b1 [ 969.686231] env[61957]: INFO nova.scheduler.client.report [None req-7c0e7a82-e685-47c5-90d8-2de75a11cf75 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Deleted allocations for instance 2e32aae5-edd7-402f-98ad-75a93d26f7a1 [ 969.737169] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-3141e8c0-2e34-4c8b-a8ac-0f12d351422c tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] [instance: 736eabcf-70f8-4e1e-9830-270d9971fc9a] Destroying instance {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 969.737169] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25db34f6-e89d-4373-a6bd-f8366078fc8e {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.741184] env[61957]: INFO nova.virt.block_device [None req-6a0c88f8-9f89-4186-9d19-09e1aa8c2fe1 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] [instance: 526ff179-62a6-4763-ab25-797617c4ed57] Attempting to driver detach volume 6c0df615-4d1a-45c6-9f48-ab1813e515b1 from mountpoint /dev/sdb [ 969.741659] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-6a0c88f8-9f89-4186-9d19-09e1aa8c2fe1 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] [instance: 526ff179-62a6-4763-ab25-797617c4ed57] Volume detach. Driver type: vmdk {{(pid=61957) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 969.742057] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-6a0c88f8-9f89-4186-9d19-09e1aa8c2fe1 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] [instance: 526ff179-62a6-4763-ab25-797617c4ed57] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-274553', 'volume_id': '6c0df615-4d1a-45c6-9f48-ab1813e515b1', 'name': 'volume-6c0df615-4d1a-45c6-9f48-ab1813e515b1', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '526ff179-62a6-4763-ab25-797617c4ed57', 'attached_at': '', 'detached_at': '', 'volume_id': '6c0df615-4d1a-45c6-9f48-ab1813e515b1', 'serial': '6c0df615-4d1a-45c6-9f48-ab1813e515b1'} {{(pid=61957) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 969.743109] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecc5f4b4-3628-4d43-a35d-db659fa15ac4 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.752559] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-3141e8c0-2e34-4c8b-a8ac-0f12d351422c tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] [instance: 736eabcf-70f8-4e1e-9830-270d9971fc9a] Unregistering the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 969.773876] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-48b752a1-181e-4587-bfe9-e88c90f3b668 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.780450] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-406c58e2-6493-4513-b50f-85dbd105adac {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.787539] env[61957]: INFO nova.compute.manager [None req-d1aa370b-6bd9-49b7-9a71-30e9793a2856 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] [instance: 00ecc8c2-398e-4a29-b19f-a2013985b481] Took 27.17 seconds to build instance. [ 969.791657] env[61957]: DEBUG nova.compute.manager [req-0c98b8b7-df38-43eb-9cbd-c408239a0442 req-be60acd4-6af5-46f2-bc9c-0e3d55b78ea5 service nova] [instance: 98929ed3-d420-45cd-9cde-9738b2ac8251] Received event network-vif-plugged-bcb69e39-31e7-4ef0-837a-fa6d08bd0449 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 969.791899] env[61957]: DEBUG oslo_concurrency.lockutils [req-0c98b8b7-df38-43eb-9cbd-c408239a0442 req-be60acd4-6af5-46f2-bc9c-0e3d55b78ea5 service nova] Acquiring lock "98929ed3-d420-45cd-9cde-9738b2ac8251-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 969.792185] env[61957]: DEBUG oslo_concurrency.lockutils [req-0c98b8b7-df38-43eb-9cbd-c408239a0442 req-be60acd4-6af5-46f2-bc9c-0e3d55b78ea5 service nova] Lock "98929ed3-d420-45cd-9cde-9738b2ac8251-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 969.792369] env[61957]: DEBUG oslo_concurrency.lockutils [req-0c98b8b7-df38-43eb-9cbd-c408239a0442 req-be60acd4-6af5-46f2-bc9c-0e3d55b78ea5 service nova] Lock "98929ed3-d420-45cd-9cde-9738b2ac8251-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 969.792556] env[61957]: DEBUG nova.compute.manager [req-0c98b8b7-df38-43eb-9cbd-c408239a0442 req-be60acd4-6af5-46f2-bc9c-0e3d55b78ea5 service nova] [instance: 98929ed3-d420-45cd-9cde-9738b2ac8251] No waiting events found dispatching network-vif-plugged-bcb69e39-31e7-4ef0-837a-fa6d08bd0449 {{(pid=61957) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 969.792740] env[61957]: WARNING nova.compute.manager [req-0c98b8b7-df38-43eb-9cbd-c408239a0442 req-be60acd4-6af5-46f2-bc9c-0e3d55b78ea5 service nova] [instance: 98929ed3-d420-45cd-9cde-9738b2ac8251] Received unexpected event network-vif-plugged-bcb69e39-31e7-4ef0-837a-fa6d08bd0449 for instance with vm_state building and task_state spawning. [ 969.792911] env[61957]: DEBUG nova.compute.manager [req-0c98b8b7-df38-43eb-9cbd-c408239a0442 req-be60acd4-6af5-46f2-bc9c-0e3d55b78ea5 service nova] [instance: 98929ed3-d420-45cd-9cde-9738b2ac8251] Received event network-changed-bcb69e39-31e7-4ef0-837a-fa6d08bd0449 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 969.793192] env[61957]: DEBUG nova.compute.manager [req-0c98b8b7-df38-43eb-9cbd-c408239a0442 req-be60acd4-6af5-46f2-bc9c-0e3d55b78ea5 service nova] [instance: 98929ed3-d420-45cd-9cde-9738b2ac8251] Refreshing instance network info cache due to event network-changed-bcb69e39-31e7-4ef0-837a-fa6d08bd0449. {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 969.793382] env[61957]: DEBUG oslo_concurrency.lockutils [req-0c98b8b7-df38-43eb-9cbd-c408239a0442 req-be60acd4-6af5-46f2-bc9c-0e3d55b78ea5 service nova] Acquiring lock "refresh_cache-98929ed3-d420-45cd-9cde-9738b2ac8251" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 969.796717] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-197dd5d4-c184-4b29-85e6-9116f5683520 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.820694] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0217beb6-07c3-44dd-919e-9fadb09f49b0 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.837701] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-6a0c88f8-9f89-4186-9d19-09e1aa8c2fe1 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] The volume has not been displaced from its original location: [datastore1] volume-6c0df615-4d1a-45c6-9f48-ab1813e515b1/volume-6c0df615-4d1a-45c6-9f48-ab1813e515b1.vmdk. No consolidation needed. {{(pid=61957) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 969.843298] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-6a0c88f8-9f89-4186-9d19-09e1aa8c2fe1 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] [instance: 526ff179-62a6-4763-ab25-797617c4ed57] Reconfiguring VM instance instance-00000041 to detach disk 2001 {{(pid=61957) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 969.844029] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-10ea565f-858e-4761-9352-453b5a3ed5a9 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.866877] env[61957]: DEBUG oslo_vmware.api [None req-6a0c88f8-9f89-4186-9d19-09e1aa8c2fe1 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Waiting for the task: (returnval){ [ 969.866877] env[61957]: value = "task-1277843" [ 969.866877] env[61957]: _type = "Task" [ 969.866877] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.876453] env[61957]: DEBUG oslo_vmware.api [None req-6a0c88f8-9f89-4186-9d19-09e1aa8c2fe1 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Task: {'id': task-1277843, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.877890] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-3141e8c0-2e34-4c8b-a8ac-0f12d351422c tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] [instance: 736eabcf-70f8-4e1e-9830-270d9971fc9a] Unregistered the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 969.878248] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-3141e8c0-2e34-4c8b-a8ac-0f12d351422c tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] [instance: 736eabcf-70f8-4e1e-9830-270d9971fc9a] Deleting contents of the VM from datastore datastore1 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 969.878385] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-3141e8c0-2e34-4c8b-a8ac-0f12d351422c tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Deleting the datastore file [datastore1] 736eabcf-70f8-4e1e-9830-270d9971fc9a {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 969.878658] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3de58cd1-d169-4446-a483-4d5dfc77e746 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.887890] env[61957]: DEBUG oslo_vmware.api [None req-3141e8c0-2e34-4c8b-a8ac-0f12d351422c tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Waiting for the task: (returnval){ [ 969.887890] env[61957]: value = "task-1277844" [ 969.887890] env[61957]: _type = "Task" [ 969.887890] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.903588] env[61957]: DEBUG nova.network.neutron [None req-cedd08d9-20a0-42aa-9f1f-33859d216b7f tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 98929ed3-d420-45cd-9cde-9738b2ac8251] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 969.917369] env[61957]: DEBUG oslo_vmware.api [None req-3141e8c0-2e34-4c8b-a8ac-0f12d351422c tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Task: {'id': task-1277844, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.922665] env[61957]: DEBUG oslo_vmware.api [None req-ce0965c3-bc7d-4a77-9d32-52f4d059d837 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52cf31fe-dcc3-770e-848e-7c860da468d0, 'name': SearchDatastore_Task, 'duration_secs': 0.021306} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.925717] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ce0965c3-bc7d-4a77-9d32-52f4d059d837 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Releasing lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 969.925717] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-ce0965c3-bc7d-4a77-9d32-52f4d059d837 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore1] 964c8c9d-0721-4ad3-a146-fce1fd1d6513/964c8c9d-0721-4ad3-a146-fce1fd1d6513.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 969.925717] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2d075a3b-8464-4318-b543-084738c5be17 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.933213] env[61957]: DEBUG oslo_vmware.api [None req-ce0965c3-bc7d-4a77-9d32-52f4d059d837 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Waiting for the task: (returnval){ [ 969.933213] env[61957]: value = "task-1277845" [ 969.933213] env[61957]: _type = "Task" [ 969.933213] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.946366] env[61957]: DEBUG oslo_vmware.api [None req-ce0965c3-bc7d-4a77-9d32-52f4d059d837 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Task: {'id': task-1277845, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.964797] env[61957]: DEBUG oslo_vmware.api [None req-9e60b82d-4fee-4f83-a3da-b48f03d27cc5 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': task-1277841, 'name': MoveVirtualDisk_Task} progress is 21%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.147526] env[61957]: DEBUG nova.network.neutron [None req-cedd08d9-20a0-42aa-9f1f-33859d216b7f tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 98929ed3-d420-45cd-9cde-9738b2ac8251] Updating instance_info_cache with network_info: [{"id": "bcb69e39-31e7-4ef0-837a-fa6d08bd0449", "address": "fa:16:3e:d4:d9:2a", "network": {"id": "efe08823-51fe-4620-a849-5345e262cc71", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1035436819-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e16f6dce3f0e44fb96de516f17d4c6f5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "667a2e97-c1be-421d-9941-6b84c2629b43", "external-id": "nsx-vlan-transportzone-484", "segmentation_id": 484, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbcb69e39-31", "ovs_interfaceid": "bcb69e39-31e7-4ef0-837a-fa6d08bd0449", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 970.197633] env[61957]: DEBUG oslo_concurrency.lockutils [None req-7c0e7a82-e685-47c5-90d8-2de75a11cf75 tempest-MultipleCreateTestJSON-436867553 tempest-MultipleCreateTestJSON-436867553-project-member] Lock "2e32aae5-edd7-402f-98ad-75a93d26f7a1" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 19.810s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 970.290227] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d1aa370b-6bd9-49b7-9a71-30e9793a2856 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Lock "00ecc8c2-398e-4a29-b19f-a2013985b481" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 28.678s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 970.376931] env[61957]: DEBUG oslo_vmware.api [None req-6a0c88f8-9f89-4186-9d19-09e1aa8c2fe1 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Task: {'id': task-1277843, 'name': ReconfigVM_Task, 'duration_secs': 0.339909} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.377268] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-6a0c88f8-9f89-4186-9d19-09e1aa8c2fe1 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] [instance: 526ff179-62a6-4763-ab25-797617c4ed57] Reconfigured VM instance instance-00000041 to detach disk 2001 {{(pid=61957) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 970.382456] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e5d05260-1021-4950-8219-60a66e7a48c6 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.398318] env[61957]: DEBUG oslo_concurrency.lockutils [None req-06b74b35-57d5-4a4b-8055-77adbdba299a tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Acquiring lock "00ecc8c2-398e-4a29-b19f-a2013985b481" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 970.398633] env[61957]: DEBUG oslo_concurrency.lockutils [None req-06b74b35-57d5-4a4b-8055-77adbdba299a tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Lock "00ecc8c2-398e-4a29-b19f-a2013985b481" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 970.398843] env[61957]: DEBUG oslo_concurrency.lockutils [None req-06b74b35-57d5-4a4b-8055-77adbdba299a tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Acquiring lock "00ecc8c2-398e-4a29-b19f-a2013985b481-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 970.399031] env[61957]: DEBUG oslo_concurrency.lockutils [None req-06b74b35-57d5-4a4b-8055-77adbdba299a tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Lock "00ecc8c2-398e-4a29-b19f-a2013985b481-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 970.399202] env[61957]: DEBUG oslo_concurrency.lockutils [None req-06b74b35-57d5-4a4b-8055-77adbdba299a tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Lock "00ecc8c2-398e-4a29-b19f-a2013985b481-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 970.405080] env[61957]: DEBUG oslo_vmware.api [None req-3141e8c0-2e34-4c8b-a8ac-0f12d351422c tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Task: {'id': task-1277844, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.405396] env[61957]: DEBUG oslo_vmware.api [None req-6a0c88f8-9f89-4186-9d19-09e1aa8c2fe1 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Waiting for the task: (returnval){ [ 970.405396] env[61957]: value = "task-1277846" [ 970.405396] env[61957]: _type = "Task" [ 970.405396] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.405895] env[61957]: INFO nova.compute.manager [None req-06b74b35-57d5-4a4b-8055-77adbdba299a tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] [instance: 00ecc8c2-398e-4a29-b19f-a2013985b481] Terminating instance [ 970.410696] env[61957]: DEBUG nova.compute.manager [None req-06b74b35-57d5-4a4b-8055-77adbdba299a tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] [instance: 00ecc8c2-398e-4a29-b19f-a2013985b481] Start destroying the instance on the hypervisor. {{(pid=61957) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 970.410905] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-06b74b35-57d5-4a4b-8055-77adbdba299a tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] [instance: 00ecc8c2-398e-4a29-b19f-a2013985b481] Destroying instance {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 970.411695] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-435f601e-c5e7-4d99-a69e-d114a4a39a59 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.421444] env[61957]: DEBUG oslo_vmware.api [None req-6a0c88f8-9f89-4186-9d19-09e1aa8c2fe1 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Task: {'id': task-1277846, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.423765] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-06b74b35-57d5-4a4b-8055-77adbdba299a tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] [instance: 00ecc8c2-398e-4a29-b19f-a2013985b481] Powering off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 970.424082] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ed4cac52-0758-4a37-9a30-94f8e9a8faf0 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.432491] env[61957]: DEBUG oslo_vmware.api [None req-06b74b35-57d5-4a4b-8055-77adbdba299a tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Waiting for the task: (returnval){ [ 970.432491] env[61957]: value = "task-1277847" [ 970.432491] env[61957]: _type = "Task" [ 970.432491] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.443722] env[61957]: DEBUG oslo_vmware.api [None req-ce0965c3-bc7d-4a77-9d32-52f4d059d837 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Task: {'id': task-1277845, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.446759] env[61957]: DEBUG oslo_vmware.api [None req-06b74b35-57d5-4a4b-8055-77adbdba299a tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Task: {'id': task-1277847, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.459918] env[61957]: DEBUG oslo_vmware.api [None req-9e60b82d-4fee-4f83-a3da-b48f03d27cc5 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': task-1277841, 'name': MoveVirtualDisk_Task} progress is 43%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.650076] env[61957]: DEBUG oslo_concurrency.lockutils [None req-cedd08d9-20a0-42aa-9f1f-33859d216b7f tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Releasing lock "refresh_cache-98929ed3-d420-45cd-9cde-9738b2ac8251" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 970.650504] env[61957]: DEBUG nova.compute.manager [None req-cedd08d9-20a0-42aa-9f1f-33859d216b7f tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 98929ed3-d420-45cd-9cde-9738b2ac8251] Instance network_info: |[{"id": "bcb69e39-31e7-4ef0-837a-fa6d08bd0449", "address": "fa:16:3e:d4:d9:2a", "network": {"id": "efe08823-51fe-4620-a849-5345e262cc71", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1035436819-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e16f6dce3f0e44fb96de516f17d4c6f5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "667a2e97-c1be-421d-9941-6b84c2629b43", "external-id": "nsx-vlan-transportzone-484", "segmentation_id": 484, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbcb69e39-31", "ovs_interfaceid": "bcb69e39-31e7-4ef0-837a-fa6d08bd0449", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 970.650905] env[61957]: DEBUG oslo_concurrency.lockutils [req-0c98b8b7-df38-43eb-9cbd-c408239a0442 req-be60acd4-6af5-46f2-bc9c-0e3d55b78ea5 service nova] Acquired lock "refresh_cache-98929ed3-d420-45cd-9cde-9738b2ac8251" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 970.651132] env[61957]: DEBUG nova.network.neutron [req-0c98b8b7-df38-43eb-9cbd-c408239a0442 req-be60acd4-6af5-46f2-bc9c-0e3d55b78ea5 service nova] [instance: 98929ed3-d420-45cd-9cde-9738b2ac8251] Refreshing network info cache for port bcb69e39-31e7-4ef0-837a-fa6d08bd0449 {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 970.652700] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-cedd08d9-20a0-42aa-9f1f-33859d216b7f tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 98929ed3-d420-45cd-9cde-9738b2ac8251] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d4:d9:2a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '667a2e97-c1be-421d-9941-6b84c2629b43', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'bcb69e39-31e7-4ef0-837a-fa6d08bd0449', 'vif_model': 'vmxnet3'}] {{(pid=61957) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 970.661624] env[61957]: DEBUG oslo.service.loopingcall [None req-cedd08d9-20a0-42aa-9f1f-33859d216b7f tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 970.661624] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 98929ed3-d420-45cd-9cde-9738b2ac8251] Creating VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 970.661624] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-be177b64-60d9-44cb-b9ef-cfecf7da3924 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.687355] env[61957]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 970.687355] env[61957]: value = "task-1277848" [ 970.687355] env[61957]: _type = "Task" [ 970.687355] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.703032] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277848, 'name': CreateVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.907781] env[61957]: DEBUG oslo_vmware.api [None req-3141e8c0-2e34-4c8b-a8ac-0f12d351422c tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Task: {'id': task-1277844, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.920709] env[61957]: DEBUG oslo_vmware.api [None req-6a0c88f8-9f89-4186-9d19-09e1aa8c2fe1 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Task: {'id': task-1277846, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.926667] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5852dfcf-4eba-49fb-83a3-c49eca7e6d2c {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.943402] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bf8f319-ab4f-4f58-b4f1-2d4443aeebe2 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.950075] env[61957]: DEBUG oslo_vmware.api [None req-06b74b35-57d5-4a4b-8055-77adbdba299a tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Task: {'id': task-1277847, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.983968] env[61957]: DEBUG oslo_vmware.api [None req-ce0965c3-bc7d-4a77-9d32-52f4d059d837 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Task: {'id': task-1277845, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.985539] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95f7f25b-3c60-481d-a642-9dd57ad44091 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.991257] env[61957]: DEBUG oslo_vmware.api [None req-9e60b82d-4fee-4f83-a3da-b48f03d27cc5 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': task-1277841, 'name': MoveVirtualDisk_Task} progress is 66%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.998381] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0c82fe8-9195-47b1-bd79-8ff300f0b7a5 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.014188] env[61957]: DEBUG nova.compute.provider_tree [None req-5145e111-afa1-43c9-8275-d20aceeb3d2b tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 971.203235] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277848, 'name': CreateVM_Task} progress is 25%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.408771] env[61957]: DEBUG oslo_vmware.api [None req-3141e8c0-2e34-4c8b-a8ac-0f12d351422c tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Task: {'id': task-1277844, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.420850] env[61957]: DEBUG oslo_vmware.api [None req-6a0c88f8-9f89-4186-9d19-09e1aa8c2fe1 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Task: {'id': task-1277846, 'name': ReconfigVM_Task, 'duration_secs': 0.929319} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.421240] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-6a0c88f8-9f89-4186-9d19-09e1aa8c2fe1 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] [instance: 526ff179-62a6-4763-ab25-797617c4ed57] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-274553', 'volume_id': '6c0df615-4d1a-45c6-9f48-ab1813e515b1', 'name': 'volume-6c0df615-4d1a-45c6-9f48-ab1813e515b1', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '526ff179-62a6-4763-ab25-797617c4ed57', 'attached_at': '', 'detached_at': '', 'volume_id': '6c0df615-4d1a-45c6-9f48-ab1813e515b1', 'serial': '6c0df615-4d1a-45c6-9f48-ab1813e515b1'} {{(pid=61957) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 971.431498] env[61957]: DEBUG nova.network.neutron [req-0c98b8b7-df38-43eb-9cbd-c408239a0442 req-be60acd4-6af5-46f2-bc9c-0e3d55b78ea5 service nova] [instance: 98929ed3-d420-45cd-9cde-9738b2ac8251] Updated VIF entry in instance network info cache for port bcb69e39-31e7-4ef0-837a-fa6d08bd0449. {{(pid=61957) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 971.431915] env[61957]: DEBUG nova.network.neutron [req-0c98b8b7-df38-43eb-9cbd-c408239a0442 req-be60acd4-6af5-46f2-bc9c-0e3d55b78ea5 service nova] [instance: 98929ed3-d420-45cd-9cde-9738b2ac8251] Updating instance_info_cache with network_info: [{"id": "bcb69e39-31e7-4ef0-837a-fa6d08bd0449", "address": "fa:16:3e:d4:d9:2a", "network": {"id": "efe08823-51fe-4620-a849-5345e262cc71", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1035436819-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e16f6dce3f0e44fb96de516f17d4c6f5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "667a2e97-c1be-421d-9941-6b84c2629b43", "external-id": "nsx-vlan-transportzone-484", "segmentation_id": 484, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbcb69e39-31", "ovs_interfaceid": "bcb69e39-31e7-4ef0-837a-fa6d08bd0449", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 971.447199] env[61957]: DEBUG oslo_vmware.api [None req-06b74b35-57d5-4a4b-8055-77adbdba299a tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Task: {'id': task-1277847, 'name': PowerOffVM_Task, 'duration_secs': 0.942414} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.448901] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-06b74b35-57d5-4a4b-8055-77adbdba299a tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] [instance: 00ecc8c2-398e-4a29-b19f-a2013985b481] Powered off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 971.449101] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-06b74b35-57d5-4a4b-8055-77adbdba299a tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] [instance: 00ecc8c2-398e-4a29-b19f-a2013985b481] Unregistering the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 971.449715] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2e202b29-53cd-4554-9b72-0e9273c3f24f {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.459537] env[61957]: DEBUG oslo_vmware.api [None req-ce0965c3-bc7d-4a77-9d32-52f4d059d837 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Task: {'id': task-1277845, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.470531] env[61957]: DEBUG oslo_vmware.api [None req-9e60b82d-4fee-4f83-a3da-b48f03d27cc5 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': task-1277841, 'name': MoveVirtualDisk_Task} progress is 85%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.520139] env[61957]: DEBUG nova.scheduler.client.report [None req-5145e111-afa1-43c9-8275-d20aceeb3d2b tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 971.560176] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-06b74b35-57d5-4a4b-8055-77adbdba299a tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] [instance: 00ecc8c2-398e-4a29-b19f-a2013985b481] Unregistered the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 971.560176] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-06b74b35-57d5-4a4b-8055-77adbdba299a tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] [instance: 00ecc8c2-398e-4a29-b19f-a2013985b481] Deleting contents of the VM from datastore datastore1 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 971.560176] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-06b74b35-57d5-4a4b-8055-77adbdba299a tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Deleting the datastore file [datastore1] 00ecc8c2-398e-4a29-b19f-a2013985b481 {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 971.560176] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d43ae93b-07a9-40d4-807c-29b2b1d63269 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.568928] env[61957]: DEBUG oslo_vmware.api [None req-06b74b35-57d5-4a4b-8055-77adbdba299a tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Waiting for the task: (returnval){ [ 971.568928] env[61957]: value = "task-1277850" [ 971.568928] env[61957]: _type = "Task" [ 971.568928] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.581212] env[61957]: DEBUG oslo_vmware.api [None req-06b74b35-57d5-4a4b-8055-77adbdba299a tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Task: {'id': task-1277850, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.709209] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277848, 'name': CreateVM_Task, 'duration_secs': 0.862642} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.709209] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 98929ed3-d420-45cd-9cde-9738b2ac8251] Created VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 971.712051] env[61957]: DEBUG oslo_concurrency.lockutils [None req-cedd08d9-20a0-42aa-9f1f-33859d216b7f tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 971.712051] env[61957]: DEBUG oslo_concurrency.lockutils [None req-cedd08d9-20a0-42aa-9f1f-33859d216b7f tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Acquired lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 971.712051] env[61957]: DEBUG oslo_concurrency.lockutils [None req-cedd08d9-20a0-42aa-9f1f-33859d216b7f tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 971.712051] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2297b540-d181-4413-a156-2f2291e52916 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.722033] env[61957]: DEBUG oslo_vmware.api [None req-cedd08d9-20a0-42aa-9f1f-33859d216b7f tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Waiting for the task: (returnval){ [ 971.722033] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52646921-963b-adc7-21b3-164388c8d630" [ 971.722033] env[61957]: _type = "Task" [ 971.722033] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.731673] env[61957]: DEBUG oslo_vmware.api [None req-cedd08d9-20a0-42aa-9f1f-33859d216b7f tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52646921-963b-adc7-21b3-164388c8d630, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.908492] env[61957]: DEBUG oslo_vmware.api [None req-3141e8c0-2e34-4c8b-a8ac-0f12d351422c tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Task: {'id': task-1277844, 'name': DeleteDatastoreFile_Task, 'duration_secs': 1.802786} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.908794] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-3141e8c0-2e34-4c8b-a8ac-0f12d351422c tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Deleted the datastore file {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 971.909049] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-3141e8c0-2e34-4c8b-a8ac-0f12d351422c tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] [instance: 736eabcf-70f8-4e1e-9830-270d9971fc9a] Deleted contents of the VM from datastore datastore1 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 971.909363] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-3141e8c0-2e34-4c8b-a8ac-0f12d351422c tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] [instance: 736eabcf-70f8-4e1e-9830-270d9971fc9a] Instance destroyed {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 971.939224] env[61957]: DEBUG oslo_concurrency.lockutils [req-0c98b8b7-df38-43eb-9cbd-c408239a0442 req-be60acd4-6af5-46f2-bc9c-0e3d55b78ea5 service nova] Releasing lock "refresh_cache-98929ed3-d420-45cd-9cde-9738b2ac8251" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 971.940234] env[61957]: DEBUG nova.compute.manager [req-0c98b8b7-df38-43eb-9cbd-c408239a0442 req-be60acd4-6af5-46f2-bc9c-0e3d55b78ea5 service nova] [instance: 736eabcf-70f8-4e1e-9830-270d9971fc9a] Received event network-vif-unplugged-95763a65-6383-409f-a9d6-1f37aaf44d34 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 971.940234] env[61957]: DEBUG oslo_concurrency.lockutils [req-0c98b8b7-df38-43eb-9cbd-c408239a0442 req-be60acd4-6af5-46f2-bc9c-0e3d55b78ea5 service nova] Acquiring lock "736eabcf-70f8-4e1e-9830-270d9971fc9a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 971.940234] env[61957]: DEBUG oslo_concurrency.lockutils [req-0c98b8b7-df38-43eb-9cbd-c408239a0442 req-be60acd4-6af5-46f2-bc9c-0e3d55b78ea5 service nova] Lock "736eabcf-70f8-4e1e-9830-270d9971fc9a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 971.940486] env[61957]: DEBUG oslo_concurrency.lockutils [req-0c98b8b7-df38-43eb-9cbd-c408239a0442 req-be60acd4-6af5-46f2-bc9c-0e3d55b78ea5 service nova] Lock "736eabcf-70f8-4e1e-9830-270d9971fc9a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 971.940486] env[61957]: DEBUG nova.compute.manager [req-0c98b8b7-df38-43eb-9cbd-c408239a0442 req-be60acd4-6af5-46f2-bc9c-0e3d55b78ea5 service nova] [instance: 736eabcf-70f8-4e1e-9830-270d9971fc9a] No waiting events found dispatching network-vif-unplugged-95763a65-6383-409f-a9d6-1f37aaf44d34 {{(pid=61957) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 971.940860] env[61957]: WARNING nova.compute.manager [req-0c98b8b7-df38-43eb-9cbd-c408239a0442 req-be60acd4-6af5-46f2-bc9c-0e3d55b78ea5 service nova] [instance: 736eabcf-70f8-4e1e-9830-270d9971fc9a] Received unexpected event network-vif-unplugged-95763a65-6383-409f-a9d6-1f37aaf44d34 for instance with vm_state shelved and task_state shelving_offloading. [ 971.941079] env[61957]: DEBUG nova.compute.manager [req-0c98b8b7-df38-43eb-9cbd-c408239a0442 req-be60acd4-6af5-46f2-bc9c-0e3d55b78ea5 service nova] [instance: 736eabcf-70f8-4e1e-9830-270d9971fc9a] Received event network-changed-95763a65-6383-409f-a9d6-1f37aaf44d34 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 971.942270] env[61957]: DEBUG nova.compute.manager [req-0c98b8b7-df38-43eb-9cbd-c408239a0442 req-be60acd4-6af5-46f2-bc9c-0e3d55b78ea5 service nova] [instance: 736eabcf-70f8-4e1e-9830-270d9971fc9a] Refreshing instance network info cache due to event network-changed-95763a65-6383-409f-a9d6-1f37aaf44d34. {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 971.942270] env[61957]: DEBUG oslo_concurrency.lockutils [req-0c98b8b7-df38-43eb-9cbd-c408239a0442 req-be60acd4-6af5-46f2-bc9c-0e3d55b78ea5 service nova] Acquiring lock "refresh_cache-736eabcf-70f8-4e1e-9830-270d9971fc9a" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 971.942270] env[61957]: DEBUG oslo_concurrency.lockutils [req-0c98b8b7-df38-43eb-9cbd-c408239a0442 req-be60acd4-6af5-46f2-bc9c-0e3d55b78ea5 service nova] Acquired lock "refresh_cache-736eabcf-70f8-4e1e-9830-270d9971fc9a" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 971.942270] env[61957]: DEBUG nova.network.neutron [req-0c98b8b7-df38-43eb-9cbd-c408239a0442 req-be60acd4-6af5-46f2-bc9c-0e3d55b78ea5 service nova] [instance: 736eabcf-70f8-4e1e-9830-270d9971fc9a] Refreshing network info cache for port 95763a65-6383-409f-a9d6-1f37aaf44d34 {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 971.944114] env[61957]: INFO nova.scheduler.client.report [None req-3141e8c0-2e34-4c8b-a8ac-0f12d351422c tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Deleted allocations for instance 736eabcf-70f8-4e1e-9830-270d9971fc9a [ 971.959234] env[61957]: DEBUG oslo_vmware.api [None req-ce0965c3-bc7d-4a77-9d32-52f4d059d837 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Task: {'id': task-1277845, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.972778] env[61957]: DEBUG oslo_vmware.api [None req-9e60b82d-4fee-4f83-a3da-b48f03d27cc5 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': task-1277841, 'name': MoveVirtualDisk_Task} progress is 100%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.987165] env[61957]: DEBUG nova.objects.instance [None req-6a0c88f8-9f89-4186-9d19-09e1aa8c2fe1 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Lazy-loading 'flavor' on Instance uuid 526ff179-62a6-4763-ab25-797617c4ed57 {{(pid=61957) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 972.027103] env[61957]: DEBUG oslo_concurrency.lockutils [None req-5145e111-afa1-43c9-8275-d20aceeb3d2b tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.394s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 972.027756] env[61957]: DEBUG nova.compute.manager [None req-5145e111-afa1-43c9-8275-d20aceeb3d2b tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 6f45c7c8-de07-4d21-bccc-e4afe3c8ad3d] Start building networks asynchronously for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 972.031959] env[61957]: DEBUG oslo_concurrency.lockutils [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 17.017s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 972.082633] env[61957]: DEBUG oslo_vmware.api [None req-06b74b35-57d5-4a4b-8055-77adbdba299a tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Task: {'id': task-1277850, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.235093] env[61957]: DEBUG oslo_vmware.api [None req-cedd08d9-20a0-42aa-9f1f-33859d216b7f tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52646921-963b-adc7-21b3-164388c8d630, 'name': SearchDatastore_Task, 'duration_secs': 0.072425} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.235093] env[61957]: DEBUG oslo_concurrency.lockutils [None req-cedd08d9-20a0-42aa-9f1f-33859d216b7f tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Releasing lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 972.235093] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-cedd08d9-20a0-42aa-9f1f-33859d216b7f tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 98929ed3-d420-45cd-9cde-9738b2ac8251] Processing image 11c76a2c-f705-470a-ba9d-4657858bab38 {{(pid=61957) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 972.235872] env[61957]: DEBUG oslo_concurrency.lockutils [None req-cedd08d9-20a0-42aa-9f1f-33859d216b7f tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 972.235872] env[61957]: DEBUG oslo_concurrency.lockutils [None req-cedd08d9-20a0-42aa-9f1f-33859d216b7f tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Acquired lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 972.235872] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-cedd08d9-20a0-42aa-9f1f-33859d216b7f tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 972.236042] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-773983a2-d24c-4ddb-8b29-6b4e5a601788 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.245545] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-cedd08d9-20a0-42aa-9f1f-33859d216b7f tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 972.245545] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-cedd08d9-20a0-42aa-9f1f-33859d216b7f tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61957) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 972.247138] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-916cd2d5-ee58-4bb9-a947-7d7206f5b845 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.252065] env[61957]: DEBUG oslo_vmware.api [None req-cedd08d9-20a0-42aa-9f1f-33859d216b7f tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Waiting for the task: (returnval){ [ 972.252065] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52be4ac4-4dfc-8cdf-be58-6f061306b4b0" [ 972.252065] env[61957]: _type = "Task" [ 972.252065] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.259986] env[61957]: DEBUG oslo_vmware.api [None req-cedd08d9-20a0-42aa-9f1f-33859d216b7f tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52be4ac4-4dfc-8cdf-be58-6f061306b4b0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.449942] env[61957]: DEBUG oslo_concurrency.lockutils [None req-3141e8c0-2e34-4c8b-a8ac-0f12d351422c tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 972.462030] env[61957]: DEBUG oslo_vmware.api [None req-ce0965c3-bc7d-4a77-9d32-52f4d059d837 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Task: {'id': task-1277845, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.292708} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.464662] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-ce0965c3-bc7d-4a77-9d32-52f4d059d837 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore1] 964c8c9d-0721-4ad3-a146-fce1fd1d6513/964c8c9d-0721-4ad3-a146-fce1fd1d6513.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 972.464902] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-ce0965c3-bc7d-4a77-9d32-52f4d059d837 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] [instance: 964c8c9d-0721-4ad3-a146-fce1fd1d6513] Extending root virtual disk to 1048576 {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 972.465175] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-99966177-e605-440a-9a23-ebd2eaa3b783 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.472667] env[61957]: DEBUG oslo_vmware.api [None req-9e60b82d-4fee-4f83-a3da-b48f03d27cc5 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': task-1277841, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.762716} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.473585] env[61957]: INFO nova.virt.vmwareapi.ds_util [None req-9e60b82d-4fee-4f83-a3da-b48f03d27cc5 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Moved virtual disk from [datastore1] OSTACK_IMG_25d5b1c0-08b2-417d-b7dc-5ba4b7379530/OSTACK_IMG_25d5b1c0-08b2-417d-b7dc-5ba4b7379530.vmdk to [datastore1] devstack-image-cache_base/fa797977-e3da-4f16-8445-e20c97e9e8ae/fa797977-e3da-4f16-8445-e20c97e9e8ae.vmdk. [ 972.473804] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-9e60b82d-4fee-4f83-a3da-b48f03d27cc5 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: d6c5c70d-86de-4dea-8b9e-76f321947a35] Cleaning up location [datastore1] OSTACK_IMG_25d5b1c0-08b2-417d-b7dc-5ba4b7379530 {{(pid=61957) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 972.473974] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-9e60b82d-4fee-4f83-a3da-b48f03d27cc5 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Deleting the datastore file [datastore1] OSTACK_IMG_25d5b1c0-08b2-417d-b7dc-5ba4b7379530 {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 972.474608] env[61957]: DEBUG oslo_vmware.api [None req-ce0965c3-bc7d-4a77-9d32-52f4d059d837 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Waiting for the task: (returnval){ [ 972.474608] env[61957]: value = "task-1277851" [ 972.474608] env[61957]: _type = "Task" [ 972.474608] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.474608] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5e9a8568-f451-4338-bae6-7de1544bfe1d {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.484302] env[61957]: DEBUG oslo_vmware.api [None req-ce0965c3-bc7d-4a77-9d32-52f4d059d837 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Task: {'id': task-1277851, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.485534] env[61957]: DEBUG oslo_vmware.api [None req-9e60b82d-4fee-4f83-a3da-b48f03d27cc5 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Waiting for the task: (returnval){ [ 972.485534] env[61957]: value = "task-1277852" [ 972.485534] env[61957]: _type = "Task" [ 972.485534] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.494911] env[61957]: DEBUG oslo_vmware.api [None req-9e60b82d-4fee-4f83-a3da-b48f03d27cc5 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': task-1277852, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.538234] env[61957]: DEBUG nova.compute.utils [None req-5145e111-afa1-43c9-8275-d20aceeb3d2b tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Using /dev/sd instead of None {{(pid=61957) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 972.546664] env[61957]: DEBUG nova.compute.manager [None req-5145e111-afa1-43c9-8275-d20aceeb3d2b tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 6f45c7c8-de07-4d21-bccc-e4afe3c8ad3d] Allocating IP information in the background. {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 972.547326] env[61957]: DEBUG nova.network.neutron [None req-5145e111-afa1-43c9-8275-d20aceeb3d2b tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 6f45c7c8-de07-4d21-bccc-e4afe3c8ad3d] allocate_for_instance() {{(pid=61957) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 972.580276] env[61957]: DEBUG oslo_vmware.api [None req-06b74b35-57d5-4a4b-8055-77adbdba299a tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Task: {'id': task-1277850, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.993115} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.581168] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-06b74b35-57d5-4a4b-8055-77adbdba299a tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Deleted the datastore file {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 972.581168] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-06b74b35-57d5-4a4b-8055-77adbdba299a tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] [instance: 00ecc8c2-398e-4a29-b19f-a2013985b481] Deleted contents of the VM from datastore datastore1 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 972.581168] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-06b74b35-57d5-4a4b-8055-77adbdba299a tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] [instance: 00ecc8c2-398e-4a29-b19f-a2013985b481] Instance destroyed {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 972.581168] env[61957]: INFO nova.compute.manager [None req-06b74b35-57d5-4a4b-8055-77adbdba299a tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] [instance: 00ecc8c2-398e-4a29-b19f-a2013985b481] Took 2.17 seconds to destroy the instance on the hypervisor. [ 972.581388] env[61957]: DEBUG oslo.service.loopingcall [None req-06b74b35-57d5-4a4b-8055-77adbdba299a tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 972.581629] env[61957]: DEBUG nova.compute.manager [-] [instance: 00ecc8c2-398e-4a29-b19f-a2013985b481] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 972.581629] env[61957]: DEBUG nova.network.neutron [-] [instance: 00ecc8c2-398e-4a29-b19f-a2013985b481] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 972.670907] env[61957]: DEBUG nova.policy [None req-5145e111-afa1-43c9-8275-d20aceeb3d2b tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e2f13689624b483d903ce12ef290db8b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '65fd71bcd1bf41238a9cc3a5d6dd4924', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61957) authorize /opt/stack/nova/nova/policy.py:203}} [ 972.766925] env[61957]: DEBUG oslo_vmware.api [None req-cedd08d9-20a0-42aa-9f1f-33859d216b7f tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52be4ac4-4dfc-8cdf-be58-6f061306b4b0, 'name': SearchDatastore_Task, 'duration_secs': 0.044904} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.768096] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2d6751d6-1299-437c-bb85-ded899a98d60 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.776402] env[61957]: DEBUG oslo_vmware.api [None req-cedd08d9-20a0-42aa-9f1f-33859d216b7f tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Waiting for the task: (returnval){ [ 972.776402] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52b2c1ed-6742-def8-57f2-9143055a1638" [ 972.776402] env[61957]: _type = "Task" [ 972.776402] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.787042] env[61957]: DEBUG oslo_vmware.api [None req-cedd08d9-20a0-42aa-9f1f-33859d216b7f tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52b2c1ed-6742-def8-57f2-9143055a1638, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.985792] env[61957]: DEBUG oslo_vmware.api [None req-ce0965c3-bc7d-4a77-9d32-52f4d059d837 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Task: {'id': task-1277851, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.061402} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.986586] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-ce0965c3-bc7d-4a77-9d32-52f4d059d837 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] [instance: 964c8c9d-0721-4ad3-a146-fce1fd1d6513] Extended root virtual disk {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 972.990077] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01c1f418-8772-423d-9267-148cdb2fde32 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.997761] env[61957]: DEBUG oslo_concurrency.lockutils [None req-6a0c88f8-9f89-4186-9d19-09e1aa8c2fe1 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Lock "526ff179-62a6-4763-ab25-797617c4ed57" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.824s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 972.999438] env[61957]: DEBUG oslo_vmware.api [None req-9e60b82d-4fee-4f83-a3da-b48f03d27cc5 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': task-1277852, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.096331} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.009395] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-9e60b82d-4fee-4f83-a3da-b48f03d27cc5 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Deleted the datastore file {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 973.010061] env[61957]: DEBUG oslo_concurrency.lockutils [None req-9e60b82d-4fee-4f83-a3da-b48f03d27cc5 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fa797977-e3da-4f16-8445-e20c97e9e8ae/fa797977-e3da-4f16-8445-e20c97e9e8ae.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 973.010606] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-9e60b82d-4fee-4f83-a3da-b48f03d27cc5 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/fa797977-e3da-4f16-8445-e20c97e9e8ae/fa797977-e3da-4f16-8445-e20c97e9e8ae.vmdk to [datastore1] d6c5c70d-86de-4dea-8b9e-76f321947a35/d6c5c70d-86de-4dea-8b9e-76f321947a35.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 973.020889] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-ce0965c3-bc7d-4a77-9d32-52f4d059d837 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] [instance: 964c8c9d-0721-4ad3-a146-fce1fd1d6513] Reconfiguring VM instance instance-00000058 to attach disk [datastore1] 964c8c9d-0721-4ad3-a146-fce1fd1d6513/964c8c9d-0721-4ad3-a146-fce1fd1d6513.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 973.025017] env[61957]: DEBUG nova.network.neutron [req-0c98b8b7-df38-43eb-9cbd-c408239a0442 req-be60acd4-6af5-46f2-bc9c-0e3d55b78ea5 service nova] [instance: 736eabcf-70f8-4e1e-9830-270d9971fc9a] Updated VIF entry in instance network info cache for port 95763a65-6383-409f-a9d6-1f37aaf44d34. {{(pid=61957) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 973.025017] env[61957]: DEBUG nova.network.neutron [req-0c98b8b7-df38-43eb-9cbd-c408239a0442 req-be60acd4-6af5-46f2-bc9c-0e3d55b78ea5 service nova] [instance: 736eabcf-70f8-4e1e-9830-270d9971fc9a] Updating instance_info_cache with network_info: [{"id": "95763a65-6383-409f-a9d6-1f37aaf44d34", "address": "fa:16:3e:64:07:83", "network": {"id": "b303ad01-ef9b-4119-9722-3be27ffb30bd", "bridge": null, "label": "tempest-AttachVolumeShelveTestJSON-96064402-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.145", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "55742b1d4fbb401ea61f990ecaa2def6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap95763a65-63", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 973.025017] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6c42f244-3eed-4b29-ad43-db9114c38c61 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.026890] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7745bcb4-73f2-408a-a5f5-b6544bb5adaf {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.042868] env[61957]: DEBUG oslo_concurrency.lockutils [req-0c98b8b7-df38-43eb-9cbd-c408239a0442 req-be60acd4-6af5-46f2-bc9c-0e3d55b78ea5 service nova] Releasing lock "refresh_cache-736eabcf-70f8-4e1e-9830-270d9971fc9a" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 973.050019] env[61957]: DEBUG oslo_vmware.api [None req-ce0965c3-bc7d-4a77-9d32-52f4d059d837 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Waiting for the task: (returnval){ [ 973.050019] env[61957]: value = "task-1277854" [ 973.050019] env[61957]: _type = "Task" [ 973.050019] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.050019] env[61957]: DEBUG nova.compute.manager [None req-5145e111-afa1-43c9-8275-d20aceeb3d2b tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 6f45c7c8-de07-4d21-bccc-e4afe3c8ad3d] Start building block device mappings for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 973.053921] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Applying migration context for instance dbeeb200-70b9-4cb4-b5a4-182389d21918 as it has an incoming, in-progress migration f4269166-72d4-48ca-b806-e1e29520eb19. Migration status is confirming {{(pid=61957) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 973.056209] env[61957]: INFO nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] [instance: dbeeb200-70b9-4cb4-b5a4-182389d21918] Updating resource usage from migration f4269166-72d4-48ca-b806-e1e29520eb19 [ 973.058988] env[61957]: DEBUG oslo_vmware.api [None req-9e60b82d-4fee-4f83-a3da-b48f03d27cc5 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Waiting for the task: (returnval){ [ 973.058988] env[61957]: value = "task-1277853" [ 973.058988] env[61957]: _type = "Task" [ 973.058988] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.077369] env[61957]: DEBUG oslo_vmware.api [None req-9e60b82d-4fee-4f83-a3da-b48f03d27cc5 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': task-1277853, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.078152] env[61957]: DEBUG oslo_vmware.api [None req-ce0965c3-bc7d-4a77-9d32-52f4d059d837 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Task: {'id': task-1277854, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.084694] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Instance 526ff179-62a6-4763-ab25-797617c4ed57 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61957) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 973.084694] env[61957]: WARNING nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Instance e40fa112-4648-428f-a403-b3e3b8319ea8 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 973.084694] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Instance 1a0327c2-5671-4970-9db7-c7cc912d8678 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61957) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 973.084694] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Instance d4674fbc-47b1-42d1-aaba-e86d46c51e8f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61957) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 973.084694] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Instance 9b0053fe-420e-4b92-86a8-ed44f6a6ec49 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61957) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 973.084694] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Migration f4269166-72d4-48ca-b806-e1e29520eb19 is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61957) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 973.084972] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Instance dbeeb200-70b9-4cb4-b5a4-182389d21918 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61957) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 973.084972] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Instance 7793aa07-6aa2-459d-8a91-56c6b0412d3b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61957) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 973.084972] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Instance 08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61957) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 973.085318] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Instance 00ecc8c2-398e-4a29-b19f-a2013985b481 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61957) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 973.085318] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Instance d6c5c70d-86de-4dea-8b9e-76f321947a35 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61957) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 973.085318] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Instance 964c8c9d-0721-4ad3-a146-fce1fd1d6513 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61957) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 973.085445] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Instance 98929ed3-d420-45cd-9cde-9738b2ac8251 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61957) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 973.085483] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Instance 6f45c7c8-de07-4d21-bccc-e4afe3c8ad3d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61957) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 973.213665] env[61957]: DEBUG nova.network.neutron [None req-5145e111-afa1-43c9-8275-d20aceeb3d2b tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 6f45c7c8-de07-4d21-bccc-e4afe3c8ad3d] Successfully created port: e0a600d6-e6d6-4598-8f1d-647ba04dd492 {{(pid=61957) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 973.292159] env[61957]: DEBUG oslo_vmware.api [None req-cedd08d9-20a0-42aa-9f1f-33859d216b7f tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52b2c1ed-6742-def8-57f2-9143055a1638, 'name': SearchDatastore_Task, 'duration_secs': 0.012759} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.292534] env[61957]: DEBUG oslo_concurrency.lockutils [None req-cedd08d9-20a0-42aa-9f1f-33859d216b7f tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Releasing lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 973.295040] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-cedd08d9-20a0-42aa-9f1f-33859d216b7f tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore1] 98929ed3-d420-45cd-9cde-9738b2ac8251/98929ed3-d420-45cd-9cde-9738b2ac8251.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 973.295040] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-08d7d3c7-7744-49ca-8bea-028eb6c05066 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.305057] env[61957]: DEBUG oslo_vmware.api [None req-cedd08d9-20a0-42aa-9f1f-33859d216b7f tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Waiting for the task: (returnval){ [ 973.305057] env[61957]: value = "task-1277855" [ 973.305057] env[61957]: _type = "Task" [ 973.305057] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.316762] env[61957]: DEBUG oslo_vmware.api [None req-cedd08d9-20a0-42aa-9f1f-33859d216b7f tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1277855, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.444168] env[61957]: DEBUG oslo_concurrency.lockutils [None req-76247214-76e0-4102-8741-73979bd5e95a tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Acquiring lock "526ff179-62a6-4763-ab25-797617c4ed57" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 973.444571] env[61957]: DEBUG oslo_concurrency.lockutils [None req-76247214-76e0-4102-8741-73979bd5e95a tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Lock "526ff179-62a6-4763-ab25-797617c4ed57" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 973.444706] env[61957]: DEBUG nova.compute.manager [None req-76247214-76e0-4102-8741-73979bd5e95a tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] [instance: 526ff179-62a6-4763-ab25-797617c4ed57] Checking state {{(pid=61957) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 973.445734] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13983440-2dc5-4ed3-8a3d-26df03094400 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.455559] env[61957]: DEBUG nova.compute.manager [None req-76247214-76e0-4102-8741-73979bd5e95a tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] [instance: 526ff179-62a6-4763-ab25-797617c4ed57] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61957) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 973.456194] env[61957]: DEBUG nova.objects.instance [None req-76247214-76e0-4102-8741-73979bd5e95a tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Lazy-loading 'flavor' on Instance uuid 526ff179-62a6-4763-ab25-797617c4ed57 {{(pid=61957) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 973.538757] env[61957]: DEBUG nova.compute.manager [req-3d081be4-a7b6-447a-bd13-6e8b2d08fb91 req-a05ba36a-62a2-430a-a935-51fe9a65dbda service nova] [instance: 00ecc8c2-398e-4a29-b19f-a2013985b481] Received event network-vif-deleted-8bcd11cf-b802-45c5-b77b-4779291e581d {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 973.538982] env[61957]: INFO nova.compute.manager [req-3d081be4-a7b6-447a-bd13-6e8b2d08fb91 req-a05ba36a-62a2-430a-a935-51fe9a65dbda service nova] [instance: 00ecc8c2-398e-4a29-b19f-a2013985b481] Neutron deleted interface 8bcd11cf-b802-45c5-b77b-4779291e581d; detaching it from the instance and deleting it from the info cache [ 973.539177] env[61957]: DEBUG nova.network.neutron [req-3d081be4-a7b6-447a-bd13-6e8b2d08fb91 req-a05ba36a-62a2-430a-a935-51fe9a65dbda service nova] [instance: 00ecc8c2-398e-4a29-b19f-a2013985b481] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 973.566668] env[61957]: DEBUG oslo_vmware.api [None req-ce0965c3-bc7d-4a77-9d32-52f4d059d837 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Task: {'id': task-1277854, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.578489] env[61957]: DEBUG oslo_vmware.api [None req-9e60b82d-4fee-4f83-a3da-b48f03d27cc5 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': task-1277853, 'name': CopyVirtualDisk_Task} progress is 21%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.590498] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Instance f65458b2-0db6-4d98-8c82-b27b6945a3fd has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61957) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 973.590866] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Total usable vcpus: 48, total allocated vcpus: 13 {{(pid=61957) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 973.591044] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=3072MB phys_disk=200GB used_disk=13GB total_vcpus=48 used_vcpus=13 pci_stats=[] {{(pid=61957) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 973.650809] env[61957]: DEBUG nova.network.neutron [-] [instance: 00ecc8c2-398e-4a29-b19f-a2013985b481] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 973.811061] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4399eb7-fc74-46b2-8053-839454e73b5a {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.821861] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9062dd04-c885-42ae-bae8-fa123a6345df {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.825340] env[61957]: DEBUG oslo_vmware.api [None req-cedd08d9-20a0-42aa-9f1f-33859d216b7f tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1277855, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.857501] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-832615ea-0c49-4e72-9fa4-6bb0216df49b {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.866626] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-703222cf-043f-4665-9cd1-63ee3c292f0b {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.883091] env[61957]: DEBUG nova.compute.provider_tree [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 973.962628] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-76247214-76e0-4102-8741-73979bd5e95a tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] [instance: 526ff179-62a6-4763-ab25-797617c4ed57] Powering off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 973.962926] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-79966f1c-f900-4134-95e6-711a38928554 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.972480] env[61957]: DEBUG oslo_vmware.api [None req-76247214-76e0-4102-8741-73979bd5e95a tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Waiting for the task: (returnval){ [ 973.972480] env[61957]: value = "task-1277856" [ 973.972480] env[61957]: _type = "Task" [ 973.972480] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.983426] env[61957]: DEBUG oslo_vmware.api [None req-76247214-76e0-4102-8741-73979bd5e95a tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Task: {'id': task-1277856, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.042346] env[61957]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2cbe7a10-bcac-4f79-8f07-b21fb1633b2b {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.061948] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1dd056ef-88c2-4aba-95d1-bed7c7a74965 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.079561] env[61957]: DEBUG nova.compute.manager [None req-5145e111-afa1-43c9-8275-d20aceeb3d2b tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 6f45c7c8-de07-4d21-bccc-e4afe3c8ad3d] Start spawning the instance on the hypervisor. {{(pid=61957) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 974.093857] env[61957]: DEBUG oslo_vmware.api [None req-ce0965c3-bc7d-4a77-9d32-52f4d059d837 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Task: {'id': task-1277854, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.106302] env[61957]: DEBUG nova.compute.manager [req-3d081be4-a7b6-447a-bd13-6e8b2d08fb91 req-a05ba36a-62a2-430a-a935-51fe9a65dbda service nova] [instance: 00ecc8c2-398e-4a29-b19f-a2013985b481] Detach interface failed, port_id=8bcd11cf-b802-45c5-b77b-4779291e581d, reason: Instance 00ecc8c2-398e-4a29-b19f-a2013985b481 could not be found. {{(pid=61957) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 974.109921] env[61957]: DEBUG oslo_vmware.api [None req-9e60b82d-4fee-4f83-a3da-b48f03d27cc5 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': task-1277853, 'name': CopyVirtualDisk_Task} progress is 43%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.130028] env[61957]: DEBUG nova.virt.hardware [None req-5145e111-afa1-43c9-8275-d20aceeb3d2b tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T17:22:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T17:21:50Z,direct_url=,disk_format='vmdk',id=11c76a2c-f705-470a-ba9d-4657858bab38,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='81d3096d0f094373913fc3dc8f5c3c6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T17:21:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 974.130174] env[61957]: DEBUG nova.virt.hardware [None req-5145e111-afa1-43c9-8275-d20aceeb3d2b tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Flavor limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 974.130340] env[61957]: DEBUG nova.virt.hardware [None req-5145e111-afa1-43c9-8275-d20aceeb3d2b tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Image limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 974.130536] env[61957]: DEBUG nova.virt.hardware [None req-5145e111-afa1-43c9-8275-d20aceeb3d2b tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Flavor pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 974.130703] env[61957]: DEBUG nova.virt.hardware [None req-5145e111-afa1-43c9-8275-d20aceeb3d2b tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Image pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 974.130858] env[61957]: DEBUG nova.virt.hardware [None req-5145e111-afa1-43c9-8275-d20aceeb3d2b tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 974.131213] env[61957]: DEBUG nova.virt.hardware [None req-5145e111-afa1-43c9-8275-d20aceeb3d2b tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 974.131302] env[61957]: DEBUG nova.virt.hardware [None req-5145e111-afa1-43c9-8275-d20aceeb3d2b tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 974.131578] env[61957]: DEBUG nova.virt.hardware [None req-5145e111-afa1-43c9-8275-d20aceeb3d2b tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Got 1 possible topologies {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 974.131791] env[61957]: DEBUG nova.virt.hardware [None req-5145e111-afa1-43c9-8275-d20aceeb3d2b tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 974.131995] env[61957]: DEBUG nova.virt.hardware [None req-5145e111-afa1-43c9-8275-d20aceeb3d2b tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 974.133063] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7138076-d3bf-426b-99d6-2acbfe646b78 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.143984] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c7607eb-d7af-4ec1-8c8c-d6214281d0c4 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.155427] env[61957]: INFO nova.compute.manager [-] [instance: 00ecc8c2-398e-4a29-b19f-a2013985b481] Took 1.57 seconds to deallocate network for instance. [ 974.318527] env[61957]: DEBUG oslo_vmware.api [None req-cedd08d9-20a0-42aa-9f1f-33859d216b7f tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1277855, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.389175] env[61957]: DEBUG nova.scheduler.client.report [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 974.485987] env[61957]: DEBUG oslo_vmware.api [None req-76247214-76e0-4102-8741-73979bd5e95a tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Task: {'id': task-1277856, 'name': PowerOffVM_Task, 'duration_secs': 0.242908} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.486318] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-76247214-76e0-4102-8741-73979bd5e95a tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] [instance: 526ff179-62a6-4763-ab25-797617c4ed57] Powered off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 974.486502] env[61957]: DEBUG nova.compute.manager [None req-76247214-76e0-4102-8741-73979bd5e95a tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] [instance: 526ff179-62a6-4763-ab25-797617c4ed57] Checking state {{(pid=61957) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 974.487344] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90a60697-7004-472a-8064-ffc699762c67 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.563647] env[61957]: DEBUG oslo_vmware.api [None req-ce0965c3-bc7d-4a77-9d32-52f4d059d837 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Task: {'id': task-1277854, 'name': ReconfigVM_Task, 'duration_secs': 1.089065} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.563820] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-ce0965c3-bc7d-4a77-9d32-52f4d059d837 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] [instance: 964c8c9d-0721-4ad3-a146-fce1fd1d6513] Reconfigured VM instance instance-00000058 to attach disk [datastore1] 964c8c9d-0721-4ad3-a146-fce1fd1d6513/964c8c9d-0721-4ad3-a146-fce1fd1d6513.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 974.564980] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b5d3ec52-9f73-45ee-9696-dc625978a7a1 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.573893] env[61957]: DEBUG oslo_vmware.api [None req-ce0965c3-bc7d-4a77-9d32-52f4d059d837 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Waiting for the task: (returnval){ [ 974.573893] env[61957]: value = "task-1277857" [ 974.573893] env[61957]: _type = "Task" [ 974.573893] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.585751] env[61957]: DEBUG oslo_vmware.api [None req-ce0965c3-bc7d-4a77-9d32-52f4d059d837 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Task: {'id': task-1277857, 'name': Rename_Task} progress is 5%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.596593] env[61957]: DEBUG oslo_vmware.api [None req-9e60b82d-4fee-4f83-a3da-b48f03d27cc5 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': task-1277853, 'name': CopyVirtualDisk_Task} progress is 66%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.678543] env[61957]: DEBUG oslo_concurrency.lockutils [None req-06b74b35-57d5-4a4b-8055-77adbdba299a tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 974.819112] env[61957]: DEBUG oslo_vmware.api [None req-cedd08d9-20a0-42aa-9f1f-33859d216b7f tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1277855, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.820162] env[61957]: DEBUG nova.network.neutron [None req-5145e111-afa1-43c9-8275-d20aceeb3d2b tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 6f45c7c8-de07-4d21-bccc-e4afe3c8ad3d] Successfully updated port: e0a600d6-e6d6-4598-8f1d-647ba04dd492 {{(pid=61957) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 974.891720] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61957) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 974.891953] env[61957]: DEBUG oslo_concurrency.lockutils [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.860s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 974.892271] env[61957]: DEBUG oslo_concurrency.lockutils [None req-7a8537d3-1c8b-4082-adcd-02831c20c149 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 17.177s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 974.892461] env[61957]: DEBUG oslo_concurrency.lockutils [None req-7a8537d3-1c8b-4082-adcd-02831c20c149 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 974.894778] env[61957]: DEBUG oslo_concurrency.lockutils [None req-19f7f8b2-5996-490c-9f47-41f02815c525 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 15.948s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 974.915062] env[61957]: INFO nova.scheduler.client.report [None req-7a8537d3-1c8b-4082-adcd-02831c20c149 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Deleted allocations for instance e40fa112-4648-428f-a403-b3e3b8319ea8 [ 975.002066] env[61957]: DEBUG oslo_concurrency.lockutils [None req-76247214-76e0-4102-8741-73979bd5e95a tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Lock "526ff179-62a6-4763-ab25-797617c4ed57" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.557s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 975.085814] env[61957]: DEBUG oslo_vmware.api [None req-ce0965c3-bc7d-4a77-9d32-52f4d059d837 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Task: {'id': task-1277857, 'name': Rename_Task} progress is 14%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.096151] env[61957]: DEBUG oslo_vmware.api [None req-9e60b82d-4fee-4f83-a3da-b48f03d27cc5 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': task-1277853, 'name': CopyVirtualDisk_Task} progress is 85%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.319548] env[61957]: DEBUG oslo_vmware.api [None req-cedd08d9-20a0-42aa-9f1f-33859d216b7f tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1277855, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.323307] env[61957]: DEBUG oslo_concurrency.lockutils [None req-5145e111-afa1-43c9-8275-d20aceeb3d2b tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Acquiring lock "refresh_cache-6f45c7c8-de07-4d21-bccc-e4afe3c8ad3d" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 975.323465] env[61957]: DEBUG oslo_concurrency.lockutils [None req-5145e111-afa1-43c9-8275-d20aceeb3d2b tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Acquired lock "refresh_cache-6f45c7c8-de07-4d21-bccc-e4afe3c8ad3d" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 975.323662] env[61957]: DEBUG nova.network.neutron [None req-5145e111-afa1-43c9-8275-d20aceeb3d2b tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 6f45c7c8-de07-4d21-bccc-e4afe3c8ad3d] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 975.376944] env[61957]: DEBUG oslo_concurrency.lockutils [None req-8b32736b-e3a6-485c-a95e-afec23b58d86 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Acquiring lock "736eabcf-70f8-4e1e-9830-270d9971fc9a" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 975.426215] env[61957]: DEBUG oslo_concurrency.lockutils [None req-7a8537d3-1c8b-4082-adcd-02831c20c149 tempest-ListServerFiltersTestJSON-1751139153 tempest-ListServerFiltersTestJSON-1751139153-project-member] Lock "e40fa112-4648-428f-a403-b3e3b8319ea8" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 20.662s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 975.593378] env[61957]: DEBUG oslo_vmware.api [None req-ce0965c3-bc7d-4a77-9d32-52f4d059d837 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Task: {'id': task-1277857, 'name': Rename_Task, 'duration_secs': 1.000205} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.594040] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-ce0965c3-bc7d-4a77-9d32-52f4d059d837 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] [instance: 964c8c9d-0721-4ad3-a146-fce1fd1d6513] Powering on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 975.596811] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ed043e41-6ac8-4375-875d-ae2fa8259c60 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.602025] env[61957]: DEBUG oslo_vmware.api [None req-9e60b82d-4fee-4f83-a3da-b48f03d27cc5 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': task-1277853, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.486754} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.602671] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-9e60b82d-4fee-4f83-a3da-b48f03d27cc5 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/fa797977-e3da-4f16-8445-e20c97e9e8ae/fa797977-e3da-4f16-8445-e20c97e9e8ae.vmdk to [datastore1] d6c5c70d-86de-4dea-8b9e-76f321947a35/d6c5c70d-86de-4dea-8b9e-76f321947a35.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 975.603520] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b0b1075-f18b-4050-ac83-51d228f37cdc {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.607691] env[61957]: DEBUG oslo_vmware.api [None req-ce0965c3-bc7d-4a77-9d32-52f4d059d837 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Waiting for the task: (returnval){ [ 975.607691] env[61957]: value = "task-1277858" [ 975.607691] env[61957]: _type = "Task" [ 975.607691] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.631915] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-9e60b82d-4fee-4f83-a3da-b48f03d27cc5 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: d6c5c70d-86de-4dea-8b9e-76f321947a35] Reconfiguring VM instance instance-00000057 to attach disk [datastore1] d6c5c70d-86de-4dea-8b9e-76f321947a35/d6c5c70d-86de-4dea-8b9e-76f321947a35.vmdk or device None with type streamOptimized {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 975.633775] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-756f6f3d-57c4-4b4c-beae-aeb41c27c3e3 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.649411] env[61957]: DEBUG nova.compute.manager [req-be3b5372-2dbd-4d5f-a1f1-bd3830e55e5b req-dfd30694-cd82-4795-92f5-4c9f5bbc0913 service nova] [instance: 6f45c7c8-de07-4d21-bccc-e4afe3c8ad3d] Received event network-vif-plugged-e0a600d6-e6d6-4598-8f1d-647ba04dd492 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 975.649633] env[61957]: DEBUG oslo_concurrency.lockutils [req-be3b5372-2dbd-4d5f-a1f1-bd3830e55e5b req-dfd30694-cd82-4795-92f5-4c9f5bbc0913 service nova] Acquiring lock "6f45c7c8-de07-4d21-bccc-e4afe3c8ad3d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 975.649851] env[61957]: DEBUG oslo_concurrency.lockutils [req-be3b5372-2dbd-4d5f-a1f1-bd3830e55e5b req-dfd30694-cd82-4795-92f5-4c9f5bbc0913 service nova] Lock "6f45c7c8-de07-4d21-bccc-e4afe3c8ad3d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 975.650064] env[61957]: DEBUG oslo_concurrency.lockutils [req-be3b5372-2dbd-4d5f-a1f1-bd3830e55e5b req-dfd30694-cd82-4795-92f5-4c9f5bbc0913 service nova] Lock "6f45c7c8-de07-4d21-bccc-e4afe3c8ad3d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 975.650238] env[61957]: DEBUG nova.compute.manager [req-be3b5372-2dbd-4d5f-a1f1-bd3830e55e5b req-dfd30694-cd82-4795-92f5-4c9f5bbc0913 service nova] [instance: 6f45c7c8-de07-4d21-bccc-e4afe3c8ad3d] No waiting events found dispatching network-vif-plugged-e0a600d6-e6d6-4598-8f1d-647ba04dd492 {{(pid=61957) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 975.650401] env[61957]: WARNING nova.compute.manager [req-be3b5372-2dbd-4d5f-a1f1-bd3830e55e5b req-dfd30694-cd82-4795-92f5-4c9f5bbc0913 service nova] [instance: 6f45c7c8-de07-4d21-bccc-e4afe3c8ad3d] Received unexpected event network-vif-plugged-e0a600d6-e6d6-4598-8f1d-647ba04dd492 for instance with vm_state building and task_state spawning. [ 975.650559] env[61957]: DEBUG nova.compute.manager [req-be3b5372-2dbd-4d5f-a1f1-bd3830e55e5b req-dfd30694-cd82-4795-92f5-4c9f5bbc0913 service nova] [instance: 6f45c7c8-de07-4d21-bccc-e4afe3c8ad3d] Received event network-changed-e0a600d6-e6d6-4598-8f1d-647ba04dd492 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 975.650712] env[61957]: DEBUG nova.compute.manager [req-be3b5372-2dbd-4d5f-a1f1-bd3830e55e5b req-dfd30694-cd82-4795-92f5-4c9f5bbc0913 service nova] [instance: 6f45c7c8-de07-4d21-bccc-e4afe3c8ad3d] Refreshing instance network info cache due to event network-changed-e0a600d6-e6d6-4598-8f1d-647ba04dd492. {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 975.650894] env[61957]: DEBUG oslo_concurrency.lockutils [req-be3b5372-2dbd-4d5f-a1f1-bd3830e55e5b req-dfd30694-cd82-4795-92f5-4c9f5bbc0913 service nova] Acquiring lock "refresh_cache-6f45c7c8-de07-4d21-bccc-e4afe3c8ad3d" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 975.656049] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7b51834-3f3c-4098-9f64-885cccc69787 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.658493] env[61957]: DEBUG oslo_vmware.api [None req-ce0965c3-bc7d-4a77-9d32-52f4d059d837 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Task: {'id': task-1277858, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.665583] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acde41de-209a-4446-9926-b2da5e766e63 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.668697] env[61957]: DEBUG oslo_vmware.api [None req-9e60b82d-4fee-4f83-a3da-b48f03d27cc5 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Waiting for the task: (returnval){ [ 975.668697] env[61957]: value = "task-1277859" [ 975.668697] env[61957]: _type = "Task" [ 975.668697] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.696366] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff5ad54a-dea5-44b3-b458-30c13a9e4e8d {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.702047] env[61957]: DEBUG oslo_vmware.api [None req-9e60b82d-4fee-4f83-a3da-b48f03d27cc5 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': task-1277859, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.706918] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be516096-ce44-4d1c-8e68-d7efdbc1c5f4 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.722524] env[61957]: DEBUG nova.compute.provider_tree [None req-19f7f8b2-5996-490c-9f47-41f02815c525 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 975.820731] env[61957]: DEBUG oslo_vmware.api [None req-cedd08d9-20a0-42aa-9f1f-33859d216b7f tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1277855, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.338037} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.821028] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-cedd08d9-20a0-42aa-9f1f-33859d216b7f tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore1] 98929ed3-d420-45cd-9cde-9738b2ac8251/98929ed3-d420-45cd-9cde-9738b2ac8251.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 975.821248] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-cedd08d9-20a0-42aa-9f1f-33859d216b7f tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 98929ed3-d420-45cd-9cde-9738b2ac8251] Extending root virtual disk to 1048576 {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 975.821547] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-82fc6d2b-9c26-4fed-b8bf-d691bb37fc1e {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.828955] env[61957]: DEBUG oslo_vmware.api [None req-cedd08d9-20a0-42aa-9f1f-33859d216b7f tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Waiting for the task: (returnval){ [ 975.828955] env[61957]: value = "task-1277860" [ 975.828955] env[61957]: _type = "Task" [ 975.828955] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.836988] env[61957]: DEBUG oslo_vmware.api [None req-cedd08d9-20a0-42aa-9f1f-33859d216b7f tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1277860, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.861968] env[61957]: DEBUG nova.network.neutron [None req-5145e111-afa1-43c9-8275-d20aceeb3d2b tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 6f45c7c8-de07-4d21-bccc-e4afe3c8ad3d] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 975.999626] env[61957]: DEBUG nova.network.neutron [None req-5145e111-afa1-43c9-8275-d20aceeb3d2b tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 6f45c7c8-de07-4d21-bccc-e4afe3c8ad3d] Updating instance_info_cache with network_info: [{"id": "e0a600d6-e6d6-4598-8f1d-647ba04dd492", "address": "fa:16:3e:14:b2:6d", "network": {"id": "4f0e57be-8906-48b5-865f-59ff12250a8b", "bridge": "br-int", "label": "tempest-ServersTestJSON-1916182687-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "65fd71bcd1bf41238a9cc3a5d6dd4924", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "939c05b6-8f31-4f3a-95ac-6297e0bd243e", "external-id": "nsx-vlan-transportzone-825", "segmentation_id": 825, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape0a600d6-e6", "ovs_interfaceid": "e0a600d6-e6d6-4598-8f1d-647ba04dd492", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 976.061210] env[61957]: DEBUG nova.objects.instance [None req-d7902eee-d4c8-4b18-9cf3-f7ae3578856d tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Lazy-loading 'flavor' on Instance uuid 526ff179-62a6-4763-ab25-797617c4ed57 {{(pid=61957) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 976.119465] env[61957]: DEBUG oslo_vmware.api [None req-ce0965c3-bc7d-4a77-9d32-52f4d059d837 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Task: {'id': task-1277858, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.178403] env[61957]: DEBUG oslo_vmware.api [None req-9e60b82d-4fee-4f83-a3da-b48f03d27cc5 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': task-1277859, 'name': ReconfigVM_Task, 'duration_secs': 0.408294} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.178707] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-9e60b82d-4fee-4f83-a3da-b48f03d27cc5 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: d6c5c70d-86de-4dea-8b9e-76f321947a35] Reconfigured VM instance instance-00000057 to attach disk [datastore1] d6c5c70d-86de-4dea-8b9e-76f321947a35/d6c5c70d-86de-4dea-8b9e-76f321947a35.vmdk or device None with type streamOptimized {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 976.179398] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3061f40b-431a-4ce5-9cbb-cc683efe8b16 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.185886] env[61957]: DEBUG oslo_vmware.api [None req-9e60b82d-4fee-4f83-a3da-b48f03d27cc5 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Waiting for the task: (returnval){ [ 976.185886] env[61957]: value = "task-1277861" [ 976.185886] env[61957]: _type = "Task" [ 976.185886] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.193859] env[61957]: DEBUG oslo_vmware.api [None req-9e60b82d-4fee-4f83-a3da-b48f03d27cc5 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': task-1277861, 'name': Rename_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.225445] env[61957]: DEBUG nova.scheduler.client.report [None req-19f7f8b2-5996-490c-9f47-41f02815c525 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 976.339368] env[61957]: DEBUG oslo_vmware.api [None req-cedd08d9-20a0-42aa-9f1f-33859d216b7f tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1277860, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070201} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.339981] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-cedd08d9-20a0-42aa-9f1f-33859d216b7f tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 98929ed3-d420-45cd-9cde-9738b2ac8251] Extended root virtual disk {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 976.340479] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f4363bd-0463-4533-bbfc-2780d64ca964 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.364304] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-cedd08d9-20a0-42aa-9f1f-33859d216b7f tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 98929ed3-d420-45cd-9cde-9738b2ac8251] Reconfiguring VM instance instance-00000059 to attach disk [datastore1] 98929ed3-d420-45cd-9cde-9738b2ac8251/98929ed3-d420-45cd-9cde-9738b2ac8251.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 976.364636] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0ec96f63-0296-4445-8ad3-09507a8f867f {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.384567] env[61957]: DEBUG oslo_vmware.api [None req-cedd08d9-20a0-42aa-9f1f-33859d216b7f tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Waiting for the task: (returnval){ [ 976.384567] env[61957]: value = "task-1277862" [ 976.384567] env[61957]: _type = "Task" [ 976.384567] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.397650] env[61957]: DEBUG oslo_vmware.api [None req-cedd08d9-20a0-42aa-9f1f-33859d216b7f tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1277862, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.503496] env[61957]: DEBUG oslo_concurrency.lockutils [None req-5145e111-afa1-43c9-8275-d20aceeb3d2b tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Releasing lock "refresh_cache-6f45c7c8-de07-4d21-bccc-e4afe3c8ad3d" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 976.503857] env[61957]: DEBUG nova.compute.manager [None req-5145e111-afa1-43c9-8275-d20aceeb3d2b tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 6f45c7c8-de07-4d21-bccc-e4afe3c8ad3d] Instance network_info: |[{"id": "e0a600d6-e6d6-4598-8f1d-647ba04dd492", "address": "fa:16:3e:14:b2:6d", "network": {"id": "4f0e57be-8906-48b5-865f-59ff12250a8b", "bridge": "br-int", "label": "tempest-ServersTestJSON-1916182687-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "65fd71bcd1bf41238a9cc3a5d6dd4924", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "939c05b6-8f31-4f3a-95ac-6297e0bd243e", "external-id": "nsx-vlan-transportzone-825", "segmentation_id": 825, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape0a600d6-e6", "ovs_interfaceid": "e0a600d6-e6d6-4598-8f1d-647ba04dd492", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 976.504205] env[61957]: DEBUG oslo_concurrency.lockutils [req-be3b5372-2dbd-4d5f-a1f1-bd3830e55e5b req-dfd30694-cd82-4795-92f5-4c9f5bbc0913 service nova] Acquired lock "refresh_cache-6f45c7c8-de07-4d21-bccc-e4afe3c8ad3d" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 976.504400] env[61957]: DEBUG nova.network.neutron [req-be3b5372-2dbd-4d5f-a1f1-bd3830e55e5b req-dfd30694-cd82-4795-92f5-4c9f5bbc0913 service nova] [instance: 6f45c7c8-de07-4d21-bccc-e4afe3c8ad3d] Refreshing network info cache for port e0a600d6-e6d6-4598-8f1d-647ba04dd492 {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 976.505709] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-5145e111-afa1-43c9-8275-d20aceeb3d2b tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 6f45c7c8-de07-4d21-bccc-e4afe3c8ad3d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:14:b2:6d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '939c05b6-8f31-4f3a-95ac-6297e0bd243e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e0a600d6-e6d6-4598-8f1d-647ba04dd492', 'vif_model': 'vmxnet3'}] {{(pid=61957) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 976.513274] env[61957]: DEBUG oslo.service.loopingcall [None req-5145e111-afa1-43c9-8275-d20aceeb3d2b tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 976.514101] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6f45c7c8-de07-4d21-bccc-e4afe3c8ad3d] Creating VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 976.514420] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-bd8427c7-660b-406f-bdf9-44312e196fa5 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.536787] env[61957]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 976.536787] env[61957]: value = "task-1277863" [ 976.536787] env[61957]: _type = "Task" [ 976.536787] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.545400] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277863, 'name': CreateVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.566768] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d7902eee-d4c8-4b18-9cf3-f7ae3578856d tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Acquiring lock "refresh_cache-526ff179-62a6-4763-ab25-797617c4ed57" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 976.566943] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d7902eee-d4c8-4b18-9cf3-f7ae3578856d tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Acquired lock "refresh_cache-526ff179-62a6-4763-ab25-797617c4ed57" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 976.567150] env[61957]: DEBUG nova.network.neutron [None req-d7902eee-d4c8-4b18-9cf3-f7ae3578856d tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] [instance: 526ff179-62a6-4763-ab25-797617c4ed57] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 976.567337] env[61957]: DEBUG nova.objects.instance [None req-d7902eee-d4c8-4b18-9cf3-f7ae3578856d tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Lazy-loading 'info_cache' on Instance uuid 526ff179-62a6-4763-ab25-797617c4ed57 {{(pid=61957) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 976.623123] env[61957]: DEBUG oslo_vmware.api [None req-ce0965c3-bc7d-4a77-9d32-52f4d059d837 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Task: {'id': task-1277858, 'name': PowerOnVM_Task, 'duration_secs': 0.529505} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.623438] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-ce0965c3-bc7d-4a77-9d32-52f4d059d837 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] [instance: 964c8c9d-0721-4ad3-a146-fce1fd1d6513] Powered on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 976.623656] env[61957]: INFO nova.compute.manager [None req-ce0965c3-bc7d-4a77-9d32-52f4d059d837 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] [instance: 964c8c9d-0721-4ad3-a146-fce1fd1d6513] Took 10.65 seconds to spawn the instance on the hypervisor. [ 976.623859] env[61957]: DEBUG nova.compute.manager [None req-ce0965c3-bc7d-4a77-9d32-52f4d059d837 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] [instance: 964c8c9d-0721-4ad3-a146-fce1fd1d6513] Checking state {{(pid=61957) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 976.624764] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-534ff147-1c9c-481d-ad11-5cc05e54fdab {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.701080] env[61957]: DEBUG oslo_vmware.api [None req-9e60b82d-4fee-4f83-a3da-b48f03d27cc5 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': task-1277861, 'name': Rename_Task, 'duration_secs': 0.213432} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.701556] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-9e60b82d-4fee-4f83-a3da-b48f03d27cc5 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: d6c5c70d-86de-4dea-8b9e-76f321947a35] Powering on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 976.701939] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c27697a0-cb96-4d6a-bdec-f43b4939c598 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.710908] env[61957]: DEBUG oslo_vmware.api [None req-9e60b82d-4fee-4f83-a3da-b48f03d27cc5 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Waiting for the task: (returnval){ [ 976.710908] env[61957]: value = "task-1277864" [ 976.710908] env[61957]: _type = "Task" [ 976.710908] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.722875] env[61957]: DEBUG oslo_vmware.api [None req-9e60b82d-4fee-4f83-a3da-b48f03d27cc5 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': task-1277864, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.896129] env[61957]: DEBUG oslo_vmware.api [None req-cedd08d9-20a0-42aa-9f1f-33859d216b7f tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1277862, 'name': ReconfigVM_Task, 'duration_secs': 0.294136} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.896462] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-cedd08d9-20a0-42aa-9f1f-33859d216b7f tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 98929ed3-d420-45cd-9cde-9738b2ac8251] Reconfigured VM instance instance-00000059 to attach disk [datastore1] 98929ed3-d420-45cd-9cde-9738b2ac8251/98929ed3-d420-45cd-9cde-9738b2ac8251.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 976.897190] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7005ca06-44cf-4dd3-976d-bf4eb3f42b1d {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.903241] env[61957]: DEBUG oslo_vmware.api [None req-cedd08d9-20a0-42aa-9f1f-33859d216b7f tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Waiting for the task: (returnval){ [ 976.903241] env[61957]: value = "task-1277865" [ 976.903241] env[61957]: _type = "Task" [ 976.903241] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.911305] env[61957]: DEBUG oslo_vmware.api [None req-cedd08d9-20a0-42aa-9f1f-33859d216b7f tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1277865, 'name': Rename_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.047448] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277863, 'name': CreateVM_Task, 'duration_secs': 0.330936} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.048331] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6f45c7c8-de07-4d21-bccc-e4afe3c8ad3d] Created VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 977.049017] env[61957]: DEBUG oslo_concurrency.lockutils [None req-5145e111-afa1-43c9-8275-d20aceeb3d2b tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 977.049201] env[61957]: DEBUG oslo_concurrency.lockutils [None req-5145e111-afa1-43c9-8275-d20aceeb3d2b tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Acquired lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 977.049601] env[61957]: DEBUG oslo_concurrency.lockutils [None req-5145e111-afa1-43c9-8275-d20aceeb3d2b tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 977.049964] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-db86d6c2-95c8-439a-bc31-04c6c970cabd {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.055081] env[61957]: DEBUG oslo_vmware.api [None req-5145e111-afa1-43c9-8275-d20aceeb3d2b tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Waiting for the task: (returnval){ [ 977.055081] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52b9b5ab-a0a2-38c6-5285-6606c415d968" [ 977.055081] env[61957]: _type = "Task" [ 977.055081] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.062922] env[61957]: DEBUG oslo_vmware.api [None req-5145e111-afa1-43c9-8275-d20aceeb3d2b tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52b9b5ab-a0a2-38c6-5285-6606c415d968, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.070244] env[61957]: DEBUG nova.objects.base [None req-d7902eee-d4c8-4b18-9cf3-f7ae3578856d tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Object Instance<526ff179-62a6-4763-ab25-797617c4ed57> lazy-loaded attributes: flavor,info_cache {{(pid=61957) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 977.142017] env[61957]: INFO nova.compute.manager [None req-ce0965c3-bc7d-4a77-9d32-52f4d059d837 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] [instance: 964c8c9d-0721-4ad3-a146-fce1fd1d6513] Took 32.38 seconds to build instance. [ 977.222196] env[61957]: DEBUG oslo_vmware.api [None req-9e60b82d-4fee-4f83-a3da-b48f03d27cc5 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': task-1277864, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.238467] env[61957]: DEBUG oslo_concurrency.lockutils [None req-19f7f8b2-5996-490c-9f47-41f02815c525 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.344s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 977.241701] env[61957]: DEBUG oslo_concurrency.lockutils [None req-731394e9-d0a3-46c3-8b02-41ef72d369f5 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.074s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 977.243214] env[61957]: INFO nova.compute.claims [None req-731394e9-d0a3-46c3-8b02-41ef72d369f5 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] [instance: f65458b2-0db6-4d98-8c82-b27b6945a3fd] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 977.414256] env[61957]: DEBUG oslo_vmware.api [None req-cedd08d9-20a0-42aa-9f1f-33859d216b7f tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1277865, 'name': Rename_Task, 'duration_secs': 0.133939} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.414640] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-cedd08d9-20a0-42aa-9f1f-33859d216b7f tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 98929ed3-d420-45cd-9cde-9738b2ac8251] Powering on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 977.414830] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-436a3b6d-973d-4761-a513-536ceaa0d526 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.423292] env[61957]: DEBUG oslo_vmware.api [None req-cedd08d9-20a0-42aa-9f1f-33859d216b7f tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Waiting for the task: (returnval){ [ 977.423292] env[61957]: value = "task-1277866" [ 977.423292] env[61957]: _type = "Task" [ 977.423292] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.438259] env[61957]: DEBUG oslo_vmware.api [None req-cedd08d9-20a0-42aa-9f1f-33859d216b7f tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1277866, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.511461] env[61957]: DEBUG nova.network.neutron [req-be3b5372-2dbd-4d5f-a1f1-bd3830e55e5b req-dfd30694-cd82-4795-92f5-4c9f5bbc0913 service nova] [instance: 6f45c7c8-de07-4d21-bccc-e4afe3c8ad3d] Updated VIF entry in instance network info cache for port e0a600d6-e6d6-4598-8f1d-647ba04dd492. {{(pid=61957) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 977.511934] env[61957]: DEBUG nova.network.neutron [req-be3b5372-2dbd-4d5f-a1f1-bd3830e55e5b req-dfd30694-cd82-4795-92f5-4c9f5bbc0913 service nova] [instance: 6f45c7c8-de07-4d21-bccc-e4afe3c8ad3d] Updating instance_info_cache with network_info: [{"id": "e0a600d6-e6d6-4598-8f1d-647ba04dd492", "address": "fa:16:3e:14:b2:6d", "network": {"id": "4f0e57be-8906-48b5-865f-59ff12250a8b", "bridge": "br-int", "label": "tempest-ServersTestJSON-1916182687-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "65fd71bcd1bf41238a9cc3a5d6dd4924", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "939c05b6-8f31-4f3a-95ac-6297e0bd243e", "external-id": "nsx-vlan-transportzone-825", "segmentation_id": 825, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape0a600d6-e6", "ovs_interfaceid": "e0a600d6-e6d6-4598-8f1d-647ba04dd492", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 977.566939] env[61957]: DEBUG oslo_vmware.api [None req-5145e111-afa1-43c9-8275-d20aceeb3d2b tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52b9b5ab-a0a2-38c6-5285-6606c415d968, 'name': SearchDatastore_Task, 'duration_secs': 0.023937} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.567274] env[61957]: DEBUG oslo_concurrency.lockutils [None req-5145e111-afa1-43c9-8275-d20aceeb3d2b tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Releasing lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 977.567518] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-5145e111-afa1-43c9-8275-d20aceeb3d2b tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 6f45c7c8-de07-4d21-bccc-e4afe3c8ad3d] Processing image 11c76a2c-f705-470a-ba9d-4657858bab38 {{(pid=61957) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 977.567791] env[61957]: DEBUG oslo_concurrency.lockutils [None req-5145e111-afa1-43c9-8275-d20aceeb3d2b tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 977.568069] env[61957]: DEBUG oslo_concurrency.lockutils [None req-5145e111-afa1-43c9-8275-d20aceeb3d2b tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Acquired lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 977.568340] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-5145e111-afa1-43c9-8275-d20aceeb3d2b tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 977.568624] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ce43c28f-e061-41ed-b9bd-f2c85270dd2f {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.576678] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-5145e111-afa1-43c9-8275-d20aceeb3d2b tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 977.576863] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-5145e111-afa1-43c9-8275-d20aceeb3d2b tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61957) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 977.577598] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-76f39c48-8c2d-4239-90e2-49b956915cbc {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.582943] env[61957]: DEBUG oslo_vmware.api [None req-5145e111-afa1-43c9-8275-d20aceeb3d2b tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Waiting for the task: (returnval){ [ 977.582943] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52015ae0-c210-7478-80c6-60428b99f073" [ 977.582943] env[61957]: _type = "Task" [ 977.582943] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.591322] env[61957]: DEBUG oslo_vmware.api [None req-5145e111-afa1-43c9-8275-d20aceeb3d2b tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52015ae0-c210-7478-80c6-60428b99f073, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.644315] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ce0965c3-bc7d-4a77-9d32-52f4d059d837 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Lock "964c8c9d-0721-4ad3-a146-fce1fd1d6513" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 33.889s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 977.723417] env[61957]: DEBUG oslo_vmware.api [None req-9e60b82d-4fee-4f83-a3da-b48f03d27cc5 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': task-1277864, 'name': PowerOnVM_Task, 'duration_secs': 0.771871} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.723753] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-9e60b82d-4fee-4f83-a3da-b48f03d27cc5 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: d6c5c70d-86de-4dea-8b9e-76f321947a35] Powered on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 977.723959] env[61957]: INFO nova.compute.manager [None req-9e60b82d-4fee-4f83-a3da-b48f03d27cc5 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: d6c5c70d-86de-4dea-8b9e-76f321947a35] Took 14.11 seconds to spawn the instance on the hypervisor. [ 977.724314] env[61957]: DEBUG nova.compute.manager [None req-9e60b82d-4fee-4f83-a3da-b48f03d27cc5 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: d6c5c70d-86de-4dea-8b9e-76f321947a35] Checking state {{(pid=61957) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 977.725146] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45e24eae-4df1-49b0-86a7-d7078482bcef {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.738906] env[61957]: DEBUG nova.compute.manager [req-b2ce9ed2-a35c-4ef6-8ce3-d5a4ab832fe3 req-e6afb85a-b6ff-4a78-82ca-20b682fe84b1 service nova] [instance: 964c8c9d-0721-4ad3-a146-fce1fd1d6513] Received event network-changed-57483094-30a7-4f77-8c6e-25dbd1db61c7 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 977.739153] env[61957]: DEBUG nova.compute.manager [req-b2ce9ed2-a35c-4ef6-8ce3-d5a4ab832fe3 req-e6afb85a-b6ff-4a78-82ca-20b682fe84b1 service nova] [instance: 964c8c9d-0721-4ad3-a146-fce1fd1d6513] Refreshing instance network info cache due to event network-changed-57483094-30a7-4f77-8c6e-25dbd1db61c7. {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 977.739401] env[61957]: DEBUG oslo_concurrency.lockutils [req-b2ce9ed2-a35c-4ef6-8ce3-d5a4ab832fe3 req-e6afb85a-b6ff-4a78-82ca-20b682fe84b1 service nova] Acquiring lock "refresh_cache-964c8c9d-0721-4ad3-a146-fce1fd1d6513" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 977.739578] env[61957]: DEBUG oslo_concurrency.lockutils [req-b2ce9ed2-a35c-4ef6-8ce3-d5a4ab832fe3 req-e6afb85a-b6ff-4a78-82ca-20b682fe84b1 service nova] Acquired lock "refresh_cache-964c8c9d-0721-4ad3-a146-fce1fd1d6513" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 977.739756] env[61957]: DEBUG nova.network.neutron [req-b2ce9ed2-a35c-4ef6-8ce3-d5a4ab832fe3 req-e6afb85a-b6ff-4a78-82ca-20b682fe84b1 service nova] [instance: 964c8c9d-0721-4ad3-a146-fce1fd1d6513] Refreshing network info cache for port 57483094-30a7-4f77-8c6e-25dbd1db61c7 {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 977.818886] env[61957]: INFO nova.scheduler.client.report [None req-19f7f8b2-5996-490c-9f47-41f02815c525 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Deleted allocation for migration f4269166-72d4-48ca-b806-e1e29520eb19 [ 977.871713] env[61957]: DEBUG nova.network.neutron [None req-d7902eee-d4c8-4b18-9cf3-f7ae3578856d tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] [instance: 526ff179-62a6-4763-ab25-797617c4ed57] Updating instance_info_cache with network_info: [{"id": "25abeea6-dd92-49e0-b014-d35dd1c0f8ee", "address": "fa:16:3e:0a:44:84", "network": {"id": "8e60ab13-6dd3-4cda-9cdf-974ad8bebe1f", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1404238496-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.173", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ea955bd4d84548ff8e6a0c21542d6ee7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9f208df-1fb5-4403-9796-7fd19e4bfb85", "external-id": "cl2-zone-400", "segmentation_id": 400, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap25abeea6-dd", "ovs_interfaceid": "25abeea6-dd92-49e0-b014-d35dd1c0f8ee", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 977.932717] env[61957]: DEBUG oslo_vmware.api [None req-cedd08d9-20a0-42aa-9f1f-33859d216b7f tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1277866, 'name': PowerOnVM_Task, 'duration_secs': 0.450866} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.933053] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-cedd08d9-20a0-42aa-9f1f-33859d216b7f tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 98929ed3-d420-45cd-9cde-9738b2ac8251] Powered on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 977.933213] env[61957]: INFO nova.compute.manager [None req-cedd08d9-20a0-42aa-9f1f-33859d216b7f tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 98929ed3-d420-45cd-9cde-9738b2ac8251] Took 9.58 seconds to spawn the instance on the hypervisor. [ 977.933394] env[61957]: DEBUG nova.compute.manager [None req-cedd08d9-20a0-42aa-9f1f-33859d216b7f tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 98929ed3-d420-45cd-9cde-9738b2ac8251] Checking state {{(pid=61957) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 977.934180] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddd6f697-8541-4d00-ae31-374961d0e622 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.014525] env[61957]: DEBUG oslo_concurrency.lockutils [req-be3b5372-2dbd-4d5f-a1f1-bd3830e55e5b req-dfd30694-cd82-4795-92f5-4c9f5bbc0913 service nova] Releasing lock "refresh_cache-6f45c7c8-de07-4d21-bccc-e4afe3c8ad3d" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 978.094216] env[61957]: DEBUG oslo_vmware.api [None req-5145e111-afa1-43c9-8275-d20aceeb3d2b tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52015ae0-c210-7478-80c6-60428b99f073, 'name': SearchDatastore_Task, 'duration_secs': 0.009912} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.094735] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fba58a5c-83f9-42bd-aed6-d69f2ac10e8e {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.100080] env[61957]: DEBUG oslo_vmware.api [None req-5145e111-afa1-43c9-8275-d20aceeb3d2b tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Waiting for the task: (returnval){ [ 978.100080] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]5261d01c-fcdb-f341-320c-6c4e78ccc9bf" [ 978.100080] env[61957]: _type = "Task" [ 978.100080] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.107621] env[61957]: DEBUG oslo_vmware.api [None req-5145e111-afa1-43c9-8275-d20aceeb3d2b tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]5261d01c-fcdb-f341-320c-6c4e78ccc9bf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.252127] env[61957]: INFO nova.compute.manager [None req-9e60b82d-4fee-4f83-a3da-b48f03d27cc5 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: d6c5c70d-86de-4dea-8b9e-76f321947a35] Took 34.96 seconds to build instance. [ 978.329792] env[61957]: DEBUG oslo_concurrency.lockutils [None req-19f7f8b2-5996-490c-9f47-41f02815c525 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Lock "dbeeb200-70b9-4cb4-b5a4-182389d21918" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 22.428s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 978.376586] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d7902eee-d4c8-4b18-9cf3-f7ae3578856d tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Releasing lock "refresh_cache-526ff179-62a6-4763-ab25-797617c4ed57" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 978.450580] env[61957]: INFO nova.compute.manager [None req-cedd08d9-20a0-42aa-9f1f-33859d216b7f tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 98929ed3-d420-45cd-9cde-9738b2ac8251] Took 25.21 seconds to build instance. [ 978.510437] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20b6a59b-415d-4d3c-af4d-54ff2f861e55 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.518300] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da431af9-e57c-42e7-a7f4-7a021d2a314e {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.554768] env[61957]: DEBUG nova.network.neutron [req-b2ce9ed2-a35c-4ef6-8ce3-d5a4ab832fe3 req-e6afb85a-b6ff-4a78-82ca-20b682fe84b1 service nova] [instance: 964c8c9d-0721-4ad3-a146-fce1fd1d6513] Updated VIF entry in instance network info cache for port 57483094-30a7-4f77-8c6e-25dbd1db61c7. {{(pid=61957) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 978.555483] env[61957]: DEBUG nova.network.neutron [req-b2ce9ed2-a35c-4ef6-8ce3-d5a4ab832fe3 req-e6afb85a-b6ff-4a78-82ca-20b682fe84b1 service nova] [instance: 964c8c9d-0721-4ad3-a146-fce1fd1d6513] Updating instance_info_cache with network_info: [{"id": "57483094-30a7-4f77-8c6e-25dbd1db61c7", "address": "fa:16:3e:55:82:02", "network": {"id": "82fd2357-53b6-4903-963e-a2866736643e", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-612681178-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.174", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1673c854272f4b14a623a73b8f4b687b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6eb7e3e9-5cc2-40f1-a6eb-f70f06531667", "external-id": "nsx-vlan-transportzone-938", "segmentation_id": 938, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap57483094-30", "ovs_interfaceid": "57483094-30a7-4f77-8c6e-25dbd1db61c7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 978.558833] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41cf9b43-0e1f-4377-98a8-5a9114a62723 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.567893] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d1c11b4-0c28-4531-808c-4f71eb175019 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.585176] env[61957]: DEBUG nova.compute.provider_tree [None req-731394e9-d0a3-46c3-8b02-41ef72d369f5 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 978.610489] env[61957]: DEBUG oslo_vmware.api [None req-5145e111-afa1-43c9-8275-d20aceeb3d2b tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]5261d01c-fcdb-f341-320c-6c4e78ccc9bf, 'name': SearchDatastore_Task, 'duration_secs': 0.02425} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.610813] env[61957]: DEBUG oslo_concurrency.lockutils [None req-5145e111-afa1-43c9-8275-d20aceeb3d2b tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Releasing lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 978.611086] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-5145e111-afa1-43c9-8275-d20aceeb3d2b tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore1] 6f45c7c8-de07-4d21-bccc-e4afe3c8ad3d/6f45c7c8-de07-4d21-bccc-e4afe3c8ad3d.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 978.611537] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-27e4617f-6743-42e3-b290-4d951c4625a0 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.620673] env[61957]: DEBUG oslo_vmware.api [None req-5145e111-afa1-43c9-8275-d20aceeb3d2b tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Waiting for the task: (returnval){ [ 978.620673] env[61957]: value = "task-1277867" [ 978.620673] env[61957]: _type = "Task" [ 978.620673] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.628643] env[61957]: DEBUG oslo_vmware.api [None req-5145e111-afa1-43c9-8275-d20aceeb3d2b tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': task-1277867, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.756824] env[61957]: DEBUG oslo_concurrency.lockutils [None req-9e60b82d-4fee-4f83-a3da-b48f03d27cc5 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Lock "d6c5c70d-86de-4dea-8b9e-76f321947a35" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 36.481s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 978.881408] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-d7902eee-d4c8-4b18-9cf3-f7ae3578856d tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] [instance: 526ff179-62a6-4763-ab25-797617c4ed57] Powering on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 978.881803] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-cc682f5c-e5e3-449a-b4d4-9d4b40341944 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.889889] env[61957]: DEBUG oslo_vmware.api [None req-d7902eee-d4c8-4b18-9cf3-f7ae3578856d tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Waiting for the task: (returnval){ [ 978.889889] env[61957]: value = "task-1277868" [ 978.889889] env[61957]: _type = "Task" [ 978.889889] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.902053] env[61957]: DEBUG oslo_vmware.api [None req-d7902eee-d4c8-4b18-9cf3-f7ae3578856d tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Task: {'id': task-1277868, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.953573] env[61957]: DEBUG oslo_concurrency.lockutils [None req-cedd08d9-20a0-42aa-9f1f-33859d216b7f tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Lock "98929ed3-d420-45cd-9cde-9738b2ac8251" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 26.723s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 979.059462] env[61957]: DEBUG oslo_concurrency.lockutils [req-b2ce9ed2-a35c-4ef6-8ce3-d5a4ab832fe3 req-e6afb85a-b6ff-4a78-82ca-20b682fe84b1 service nova] Releasing lock "refresh_cache-964c8c9d-0721-4ad3-a146-fce1fd1d6513" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 979.089228] env[61957]: DEBUG nova.scheduler.client.report [None req-731394e9-d0a3-46c3-8b02-41ef72d369f5 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 979.131853] env[61957]: DEBUG oslo_vmware.api [None req-5145e111-afa1-43c9-8275-d20aceeb3d2b tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': task-1277867, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.400683] env[61957]: DEBUG oslo_vmware.api [None req-d7902eee-d4c8-4b18-9cf3-f7ae3578856d tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Task: {'id': task-1277868, 'name': PowerOnVM_Task, 'duration_secs': 0.495825} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.400924] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-d7902eee-d4c8-4b18-9cf3-f7ae3578856d tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] [instance: 526ff179-62a6-4763-ab25-797617c4ed57] Powered on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 979.401203] env[61957]: DEBUG nova.compute.manager [None req-d7902eee-d4c8-4b18-9cf3-f7ae3578856d tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] [instance: 526ff179-62a6-4763-ab25-797617c4ed57] Checking state {{(pid=61957) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 979.402015] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f840315-9d61-42d6-a173-ba03702bb6f0 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.596168] env[61957]: DEBUG oslo_concurrency.lockutils [None req-731394e9-d0a3-46c3-8b02-41ef72d369f5 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.354s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 979.597726] env[61957]: DEBUG nova.compute.manager [None req-731394e9-d0a3-46c3-8b02-41ef72d369f5 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] [instance: f65458b2-0db6-4d98-8c82-b27b6945a3fd] Start building networks asynchronously for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 979.601735] env[61957]: DEBUG oslo_concurrency.lockutils [None req-3141e8c0-2e34-4c8b-a8ac-0f12d351422c tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 7.152s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 979.601838] env[61957]: DEBUG oslo_concurrency.lockutils [None req-3141e8c0-2e34-4c8b-a8ac-0f12d351422c tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 979.604216] env[61957]: DEBUG oslo_concurrency.lockutils [None req-06b74b35-57d5-4a4b-8055-77adbdba299a tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.926s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 979.604451] env[61957]: DEBUG nova.objects.instance [None req-06b74b35-57d5-4a4b-8055-77adbdba299a tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Lazy-loading 'resources' on Instance uuid 00ecc8c2-398e-4a29-b19f-a2013985b481 {{(pid=61957) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 979.637388] env[61957]: DEBUG oslo_vmware.api [None req-5145e111-afa1-43c9-8275-d20aceeb3d2b tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': task-1277867, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.564019} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.638587] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-5145e111-afa1-43c9-8275-d20aceeb3d2b tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore1] 6f45c7c8-de07-4d21-bccc-e4afe3c8ad3d/6f45c7c8-de07-4d21-bccc-e4afe3c8ad3d.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 979.638587] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-5145e111-afa1-43c9-8275-d20aceeb3d2b tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 6f45c7c8-de07-4d21-bccc-e4afe3c8ad3d] Extending root virtual disk to 1048576 {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 979.638587] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-cad4bc3d-8593-410d-9a20-9258032e24cd {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.648115] env[61957]: DEBUG oslo_vmware.api [None req-5145e111-afa1-43c9-8275-d20aceeb3d2b tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Waiting for the task: (returnval){ [ 979.648115] env[61957]: value = "task-1277869" [ 979.648115] env[61957]: _type = "Task" [ 979.648115] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.661811] env[61957]: DEBUG oslo_vmware.api [None req-5145e111-afa1-43c9-8275-d20aceeb3d2b tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': task-1277869, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.107354] env[61957]: DEBUG nova.compute.utils [None req-731394e9-d0a3-46c3-8b02-41ef72d369f5 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Using /dev/sd instead of None {{(pid=61957) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 980.118462] env[61957]: DEBUG nova.compute.manager [None req-731394e9-d0a3-46c3-8b02-41ef72d369f5 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] [instance: f65458b2-0db6-4d98-8c82-b27b6945a3fd] Allocating IP information in the background. {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 980.118647] env[61957]: DEBUG nova.network.neutron [None req-731394e9-d0a3-46c3-8b02-41ef72d369f5 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] [instance: f65458b2-0db6-4d98-8c82-b27b6945a3fd] allocate_for_instance() {{(pid=61957) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 980.124910] env[61957]: DEBUG oslo_concurrency.lockutils [None req-3141e8c0-2e34-4c8b-a8ac-0f12d351422c tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Lock "736eabcf-70f8-4e1e-9830-270d9971fc9a" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 29.168s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 980.125861] env[61957]: DEBUG oslo_concurrency.lockutils [None req-8b32736b-e3a6-485c-a95e-afec23b58d86 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Lock "736eabcf-70f8-4e1e-9830-270d9971fc9a" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 4.749s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 980.126069] env[61957]: INFO nova.compute.manager [None req-8b32736b-e3a6-485c-a95e-afec23b58d86 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] [instance: 736eabcf-70f8-4e1e-9830-270d9971fc9a] Unshelving [ 980.148997] env[61957]: DEBUG oslo_concurrency.lockutils [None req-2f034be2-1f7c-41d0-91ad-8c549dbb7f8a tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Acquiring lock "d6c5c70d-86de-4dea-8b9e-76f321947a35" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 980.149386] env[61957]: DEBUG oslo_concurrency.lockutils [None req-2f034be2-1f7c-41d0-91ad-8c549dbb7f8a tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Lock "d6c5c70d-86de-4dea-8b9e-76f321947a35" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 980.149742] env[61957]: DEBUG oslo_concurrency.lockutils [None req-2f034be2-1f7c-41d0-91ad-8c549dbb7f8a tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Acquiring lock "d6c5c70d-86de-4dea-8b9e-76f321947a35-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 980.150055] env[61957]: DEBUG oslo_concurrency.lockutils [None req-2f034be2-1f7c-41d0-91ad-8c549dbb7f8a tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Lock "d6c5c70d-86de-4dea-8b9e-76f321947a35-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 980.150267] env[61957]: DEBUG oslo_concurrency.lockutils [None req-2f034be2-1f7c-41d0-91ad-8c549dbb7f8a tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Lock "d6c5c70d-86de-4dea-8b9e-76f321947a35-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 980.157139] env[61957]: INFO nova.compute.manager [None req-2f034be2-1f7c-41d0-91ad-8c549dbb7f8a tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: d6c5c70d-86de-4dea-8b9e-76f321947a35] Terminating instance [ 980.159225] env[61957]: DEBUG nova.compute.manager [None req-2f034be2-1f7c-41d0-91ad-8c549dbb7f8a tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: d6c5c70d-86de-4dea-8b9e-76f321947a35] Start destroying the instance on the hypervisor. {{(pid=61957) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 980.159547] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-2f034be2-1f7c-41d0-91ad-8c549dbb7f8a tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: d6c5c70d-86de-4dea-8b9e-76f321947a35] Destroying instance {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 980.162273] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0e5768d-5d78-4c9c-9b4b-4c52d354e5c3 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.173324] env[61957]: DEBUG oslo_vmware.api [None req-5145e111-afa1-43c9-8275-d20aceeb3d2b tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': task-1277869, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069005} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.174178] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-5145e111-afa1-43c9-8275-d20aceeb3d2b tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 6f45c7c8-de07-4d21-bccc-e4afe3c8ad3d] Extended root virtual disk {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 980.175699] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e2e31c4-9096-427d-9658-68f3c7e7d812 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.180522] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-2f034be2-1f7c-41d0-91ad-8c549dbb7f8a tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: d6c5c70d-86de-4dea-8b9e-76f321947a35] Powering off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 980.182043] env[61957]: DEBUG nova.policy [None req-731394e9-d0a3-46c3-8b02-41ef72d369f5 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd6d0d8fdb5f94a7f99c24ec9db21673c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7758bdd10505455b8cecb8129d70ebd5', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61957) authorize /opt/stack/nova/nova/policy.py:203}} [ 980.184187] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2c4ea459-3983-48c7-82b3-52a40907d100 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.204978] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-5145e111-afa1-43c9-8275-d20aceeb3d2b tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 6f45c7c8-de07-4d21-bccc-e4afe3c8ad3d] Reconfiguring VM instance instance-0000005a to attach disk [datastore1] 6f45c7c8-de07-4d21-bccc-e4afe3c8ad3d/6f45c7c8-de07-4d21-bccc-e4afe3c8ad3d.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 980.209590] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f38738d0-f8e4-470b-9407-8ba4ab8721fc {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.226722] env[61957]: DEBUG oslo_vmware.api [None req-2f034be2-1f7c-41d0-91ad-8c549dbb7f8a tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Waiting for the task: (returnval){ [ 980.226722] env[61957]: value = "task-1277870" [ 980.226722] env[61957]: _type = "Task" [ 980.226722] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.233947] env[61957]: DEBUG oslo_vmware.api [None req-5145e111-afa1-43c9-8275-d20aceeb3d2b tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Waiting for the task: (returnval){ [ 980.233947] env[61957]: value = "task-1277871" [ 980.233947] env[61957]: _type = "Task" [ 980.233947] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.237343] env[61957]: DEBUG oslo_vmware.api [None req-2f034be2-1f7c-41d0-91ad-8c549dbb7f8a tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': task-1277870, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.250222] env[61957]: DEBUG oslo_vmware.api [None req-5145e111-afa1-43c9-8275-d20aceeb3d2b tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': task-1277871, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.406110] env[61957]: DEBUG oslo_concurrency.lockutils [None req-b10a76eb-ac59-4162-9083-648b4df991d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Acquiring lock "98929ed3-d420-45cd-9cde-9738b2ac8251" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 980.406110] env[61957]: DEBUG oslo_concurrency.lockutils [None req-b10a76eb-ac59-4162-9083-648b4df991d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Lock "98929ed3-d420-45cd-9cde-9738b2ac8251" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 980.406110] env[61957]: INFO nova.compute.manager [None req-b10a76eb-ac59-4162-9083-648b4df991d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 98929ed3-d420-45cd-9cde-9738b2ac8251] Shelving [ 980.407384] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44fd4a5b-851e-4e40-8e3e-63a0b31c0700 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.420262] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-146f92fc-ee69-4060-a5ed-6116bb5632a0 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.454254] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-613e4e4b-651c-451c-b053-9d058c9478f7 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.462271] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ebd8f05-20f4-4bd6-b83a-759afd693864 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.482254] env[61957]: DEBUG nova.compute.provider_tree [None req-06b74b35-57d5-4a4b-8055-77adbdba299a tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 980.491984] env[61957]: DEBUG nova.network.neutron [None req-731394e9-d0a3-46c3-8b02-41ef72d369f5 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] [instance: f65458b2-0db6-4d98-8c82-b27b6945a3fd] Successfully created port: bb715e7d-14c5-418b-b3ee-1ac91aba709d {{(pid=61957) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 980.620612] env[61957]: DEBUG nova.compute.manager [None req-731394e9-d0a3-46c3-8b02-41ef72d369f5 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] [instance: f65458b2-0db6-4d98-8c82-b27b6945a3fd] Start building block device mappings for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 980.738180] env[61957]: DEBUG oslo_vmware.api [None req-2f034be2-1f7c-41d0-91ad-8c549dbb7f8a tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': task-1277870, 'name': PowerOffVM_Task, 'duration_secs': 0.262141} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.741358] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-2f034be2-1f7c-41d0-91ad-8c549dbb7f8a tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: d6c5c70d-86de-4dea-8b9e-76f321947a35] Powered off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 980.741675] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-2f034be2-1f7c-41d0-91ad-8c549dbb7f8a tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: d6c5c70d-86de-4dea-8b9e-76f321947a35] Unregistering the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 980.742376] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-917a0287-3606-4141-b0ec-a3b916d468e5 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.749909] env[61957]: DEBUG oslo_vmware.api [None req-5145e111-afa1-43c9-8275-d20aceeb3d2b tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': task-1277871, 'name': ReconfigVM_Task, 'duration_secs': 0.311238} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.749909] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-5145e111-afa1-43c9-8275-d20aceeb3d2b tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 6f45c7c8-de07-4d21-bccc-e4afe3c8ad3d] Reconfigured VM instance instance-0000005a to attach disk [datastore1] 6f45c7c8-de07-4d21-bccc-e4afe3c8ad3d/6f45c7c8-de07-4d21-bccc-e4afe3c8ad3d.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 980.749909] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d4a6144d-83ea-48b1-a56a-3e97311cd9da {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.757230] env[61957]: DEBUG oslo_vmware.api [None req-5145e111-afa1-43c9-8275-d20aceeb3d2b tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Waiting for the task: (returnval){ [ 980.757230] env[61957]: value = "task-1277873" [ 980.757230] env[61957]: _type = "Task" [ 980.757230] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.765143] env[61957]: DEBUG oslo_vmware.api [None req-5145e111-afa1-43c9-8275-d20aceeb3d2b tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': task-1277873, 'name': Rename_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.820392] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-2f034be2-1f7c-41d0-91ad-8c549dbb7f8a tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: d6c5c70d-86de-4dea-8b9e-76f321947a35] Unregistered the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 980.820654] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-2f034be2-1f7c-41d0-91ad-8c549dbb7f8a tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: d6c5c70d-86de-4dea-8b9e-76f321947a35] Deleting contents of the VM from datastore datastore1 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 980.820866] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-2f034be2-1f7c-41d0-91ad-8c549dbb7f8a tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Deleting the datastore file [datastore1] d6c5c70d-86de-4dea-8b9e-76f321947a35 {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 980.821187] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-eb884ae7-55d9-4f8f-827d-0ae3ddb69821 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.830025] env[61957]: DEBUG oslo_vmware.api [None req-2f034be2-1f7c-41d0-91ad-8c549dbb7f8a tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Waiting for the task: (returnval){ [ 980.830025] env[61957]: value = "task-1277874" [ 980.830025] env[61957]: _type = "Task" [ 980.830025] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.843016] env[61957]: DEBUG oslo_vmware.api [None req-2f034be2-1f7c-41d0-91ad-8c549dbb7f8a tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': task-1277874, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.896251] env[61957]: DEBUG oslo_concurrency.lockutils [None req-6d13ccf5-9ccd-48d0-a35e-c6df97be9c15 tempest-ServerAddressesTestJSON-2054963291 tempest-ServerAddressesTestJSON-2054963291-project-member] Acquiring lock "fa4c150e-7a85-41a4-8890-9f2deb3b34b1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 980.896487] env[61957]: DEBUG oslo_concurrency.lockutils [None req-6d13ccf5-9ccd-48d0-a35e-c6df97be9c15 tempest-ServerAddressesTestJSON-2054963291 tempest-ServerAddressesTestJSON-2054963291-project-member] Lock "fa4c150e-7a85-41a4-8890-9f2deb3b34b1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 980.916353] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-b10a76eb-ac59-4162-9083-648b4df991d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 98929ed3-d420-45cd-9cde-9738b2ac8251] Powering off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 980.916631] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-52c16b3d-0e0d-4595-a392-5533a1b74519 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.924191] env[61957]: DEBUG oslo_vmware.api [None req-b10a76eb-ac59-4162-9083-648b4df991d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Waiting for the task: (returnval){ [ 980.924191] env[61957]: value = "task-1277875" [ 980.924191] env[61957]: _type = "Task" [ 980.924191] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.933834] env[61957]: DEBUG oslo_vmware.api [None req-b10a76eb-ac59-4162-9083-648b4df991d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1277875, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.982751] env[61957]: DEBUG nova.scheduler.client.report [None req-06b74b35-57d5-4a4b-8055-77adbdba299a tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 981.006926] env[61957]: DEBUG oslo_concurrency.lockutils [None req-c255bb00-cba6-445a-adb0-12124fe511af tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Acquiring lock "dbeeb200-70b9-4cb4-b5a4-182389d21918" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 981.007253] env[61957]: DEBUG oslo_concurrency.lockutils [None req-c255bb00-cba6-445a-adb0-12124fe511af tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Lock "dbeeb200-70b9-4cb4-b5a4-182389d21918" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 981.007776] env[61957]: DEBUG oslo_concurrency.lockutils [None req-c255bb00-cba6-445a-adb0-12124fe511af tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Acquiring lock "dbeeb200-70b9-4cb4-b5a4-182389d21918-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 981.008629] env[61957]: DEBUG oslo_concurrency.lockutils [None req-c255bb00-cba6-445a-adb0-12124fe511af tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Lock "dbeeb200-70b9-4cb4-b5a4-182389d21918-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 981.008836] env[61957]: DEBUG oslo_concurrency.lockutils [None req-c255bb00-cba6-445a-adb0-12124fe511af tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Lock "dbeeb200-70b9-4cb4-b5a4-182389d21918-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 981.011858] env[61957]: INFO nova.compute.manager [None req-c255bb00-cba6-445a-adb0-12124fe511af tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: dbeeb200-70b9-4cb4-b5a4-182389d21918] Terminating instance [ 981.014529] env[61957]: DEBUG nova.compute.manager [None req-c255bb00-cba6-445a-adb0-12124fe511af tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: dbeeb200-70b9-4cb4-b5a4-182389d21918] Start destroying the instance on the hypervisor. {{(pid=61957) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 981.014780] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-c255bb00-cba6-445a-adb0-12124fe511af tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: dbeeb200-70b9-4cb4-b5a4-182389d21918] Destroying instance {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 981.015957] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-975a1e3d-2172-47ef-9e55-5183a8963346 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.024246] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-c255bb00-cba6-445a-adb0-12124fe511af tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: dbeeb200-70b9-4cb4-b5a4-182389d21918] Powering off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 981.024504] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6a3336e2-6d83-459d-bb2f-df2433d5ab78 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.030639] env[61957]: DEBUG oslo_vmware.api [None req-c255bb00-cba6-445a-adb0-12124fe511af tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Waiting for the task: (returnval){ [ 981.030639] env[61957]: value = "task-1277876" [ 981.030639] env[61957]: _type = "Task" [ 981.030639] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.038976] env[61957]: DEBUG oslo_vmware.api [None req-c255bb00-cba6-445a-adb0-12124fe511af tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': task-1277876, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.869878] env[61957]: DEBUG oslo_vmware.api [None req-5145e111-afa1-43c9-8275-d20aceeb3d2b tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': task-1277873, 'name': Rename_Task} progress is 99%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.885314] env[61957]: DEBUG nova.compute.manager [None req-6d13ccf5-9ccd-48d0-a35e-c6df97be9c15 tempest-ServerAddressesTestJSON-2054963291 tempest-ServerAddressesTestJSON-2054963291-project-member] [instance: fa4c150e-7a85-41a4-8890-9f2deb3b34b1] Starting instance... {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 981.898196] env[61957]: DEBUG oslo_concurrency.lockutils [None req-06b74b35-57d5-4a4b-8055-77adbdba299a tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.292s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 981.902628] env[61957]: DEBUG oslo_concurrency.lockutils [None req-8b32736b-e3a6-485c-a95e-afec23b58d86 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 981.903228] env[61957]: DEBUG oslo_concurrency.lockutils [None req-8b32736b-e3a6-485c-a95e-afec23b58d86 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 981.904373] env[61957]: DEBUG nova.objects.instance [None req-8b32736b-e3a6-485c-a95e-afec23b58d86 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Lazy-loading 'pci_requests' on Instance uuid 736eabcf-70f8-4e1e-9830-270d9971fc9a {{(pid=61957) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 981.926085] env[61957]: DEBUG nova.objects.instance [None req-8b32736b-e3a6-485c-a95e-afec23b58d86 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Lazy-loading 'numa_topology' on Instance uuid 736eabcf-70f8-4e1e-9830-270d9971fc9a {{(pid=61957) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 981.933791] env[61957]: DEBUG oslo_vmware.api [None req-2f034be2-1f7c-41d0-91ad-8c549dbb7f8a tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': task-1277874, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.198674} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.944442] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-2f034be2-1f7c-41d0-91ad-8c549dbb7f8a tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Deleted the datastore file {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 981.945393] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-2f034be2-1f7c-41d0-91ad-8c549dbb7f8a tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: d6c5c70d-86de-4dea-8b9e-76f321947a35] Deleted contents of the VM from datastore datastore1 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 981.945393] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-2f034be2-1f7c-41d0-91ad-8c549dbb7f8a tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: d6c5c70d-86de-4dea-8b9e-76f321947a35] Instance destroyed {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 981.945393] env[61957]: INFO nova.compute.manager [None req-2f034be2-1f7c-41d0-91ad-8c549dbb7f8a tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: d6c5c70d-86de-4dea-8b9e-76f321947a35] Took 1.79 seconds to destroy the instance on the hypervisor. [ 981.945880] env[61957]: DEBUG oslo.service.loopingcall [None req-2f034be2-1f7c-41d0-91ad-8c549dbb7f8a tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 981.948456] env[61957]: DEBUG oslo_vmware.api [None req-b10a76eb-ac59-4162-9083-648b4df991d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1277875, 'name': PowerOffVM_Task, 'duration_secs': 0.249782} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.948456] env[61957]: DEBUG oslo_vmware.api [None req-c255bb00-cba6-445a-adb0-12124fe511af tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': task-1277876, 'name': PowerOffVM_Task, 'duration_secs': 0.161798} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.948456] env[61957]: DEBUG oslo_vmware.api [None req-5145e111-afa1-43c9-8275-d20aceeb3d2b tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': task-1277873, 'name': Rename_Task, 'duration_secs': 1.153929} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.950513] env[61957]: INFO nova.scheduler.client.report [None req-06b74b35-57d5-4a4b-8055-77adbdba299a tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Deleted allocations for instance 00ecc8c2-398e-4a29-b19f-a2013985b481 [ 981.951601] env[61957]: DEBUG nova.compute.manager [-] [instance: d6c5c70d-86de-4dea-8b9e-76f321947a35] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 981.951713] env[61957]: DEBUG nova.network.neutron [-] [instance: d6c5c70d-86de-4dea-8b9e-76f321947a35] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 981.954150] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-b10a76eb-ac59-4162-9083-648b4df991d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 98929ed3-d420-45cd-9cde-9738b2ac8251] Powered off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 981.954514] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-c255bb00-cba6-445a-adb0-12124fe511af tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: dbeeb200-70b9-4cb4-b5a4-182389d21918] Powered off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 981.954683] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-c255bb00-cba6-445a-adb0-12124fe511af tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: dbeeb200-70b9-4cb4-b5a4-182389d21918] Unregistering the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 981.955010] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-5145e111-afa1-43c9-8275-d20aceeb3d2b tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 6f45c7c8-de07-4d21-bccc-e4afe3c8ad3d] Powering on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 981.960702] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fb24c04-bb33-4282-b6bc-ab6a42d4f7d5 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.963756] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-03dec98f-0e3a-47e6-90c7-66f8c9d937d8 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.965308] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-43c87d60-23a0-43b2-8124-95e35448e1b5 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.973692] env[61957]: DEBUG oslo_vmware.api [None req-5145e111-afa1-43c9-8275-d20aceeb3d2b tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Waiting for the task: (returnval){ [ 981.973692] env[61957]: value = "task-1277877" [ 981.973692] env[61957]: _type = "Task" [ 981.973692] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.993603] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-684d81a0-36ba-4584-984e-2d748524c3c4 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.003103] env[61957]: DEBUG oslo_vmware.api [None req-5145e111-afa1-43c9-8275-d20aceeb3d2b tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': task-1277877, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.042132] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-c255bb00-cba6-445a-adb0-12124fe511af tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: dbeeb200-70b9-4cb4-b5a4-182389d21918] Unregistered the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 982.042326] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-c255bb00-cba6-445a-adb0-12124fe511af tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: dbeeb200-70b9-4cb4-b5a4-182389d21918] Deleting contents of the VM from datastore datastore1 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 982.042784] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-c255bb00-cba6-445a-adb0-12124fe511af tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Deleting the datastore file [datastore1] dbeeb200-70b9-4cb4-b5a4-182389d21918 {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 982.042784] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1eb5de37-2a4d-4d48-af6b-21d2c3c2568e {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.051194] env[61957]: DEBUG oslo_vmware.api [None req-c255bb00-cba6-445a-adb0-12124fe511af tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Waiting for the task: (returnval){ [ 982.051194] env[61957]: value = "task-1277879" [ 982.051194] env[61957]: _type = "Task" [ 982.051194] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.065098] env[61957]: DEBUG oslo_vmware.api [None req-c255bb00-cba6-445a-adb0-12124fe511af tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': task-1277879, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.380353] env[61957]: DEBUG nova.compute.manager [req-7b576ccc-cea9-4eb4-b2d5-414bd8cb987d req-f47b492f-3c4f-4405-979b-b77735b8e215 service nova] [instance: d6c5c70d-86de-4dea-8b9e-76f321947a35] Received event network-vif-deleted-642fc1a5-237a-4bde-b28f-69b3f0f46a8a {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 982.380587] env[61957]: INFO nova.compute.manager [req-7b576ccc-cea9-4eb4-b2d5-414bd8cb987d req-f47b492f-3c4f-4405-979b-b77735b8e215 service nova] [instance: d6c5c70d-86de-4dea-8b9e-76f321947a35] Neutron deleted interface 642fc1a5-237a-4bde-b28f-69b3f0f46a8a; detaching it from the instance and deleting it from the info cache [ 982.380770] env[61957]: DEBUG nova.network.neutron [req-7b576ccc-cea9-4eb4-b2d5-414bd8cb987d req-f47b492f-3c4f-4405-979b-b77735b8e215 service nova] [instance: d6c5c70d-86de-4dea-8b9e-76f321947a35] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 982.401554] env[61957]: DEBUG nova.compute.manager [None req-731394e9-d0a3-46c3-8b02-41ef72d369f5 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] [instance: f65458b2-0db6-4d98-8c82-b27b6945a3fd] Start spawning the instance on the hypervisor. {{(pid=61957) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 982.431079] env[61957]: INFO nova.compute.claims [None req-8b32736b-e3a6-485c-a95e-afec23b58d86 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] [instance: 736eabcf-70f8-4e1e-9830-270d9971fc9a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 982.434939] env[61957]: DEBUG oslo_concurrency.lockutils [None req-6d13ccf5-9ccd-48d0-a35e-c6df97be9c15 tempest-ServerAddressesTestJSON-2054963291 tempest-ServerAddressesTestJSON-2054963291-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 982.446697] env[61957]: DEBUG nova.virt.hardware [None req-731394e9-d0a3-46c3-8b02-41ef72d369f5 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T17:22:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T17:21:50Z,direct_url=,disk_format='vmdk',id=11c76a2c-f705-470a-ba9d-4657858bab38,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='81d3096d0f094373913fc3dc8f5c3c6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T17:21:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 982.446981] env[61957]: DEBUG nova.virt.hardware [None req-731394e9-d0a3-46c3-8b02-41ef72d369f5 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Flavor limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 982.447185] env[61957]: DEBUG nova.virt.hardware [None req-731394e9-d0a3-46c3-8b02-41ef72d369f5 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Image limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 982.447382] env[61957]: DEBUG nova.virt.hardware [None req-731394e9-d0a3-46c3-8b02-41ef72d369f5 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Flavor pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 982.451094] env[61957]: DEBUG nova.virt.hardware [None req-731394e9-d0a3-46c3-8b02-41ef72d369f5 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Image pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 982.451325] env[61957]: DEBUG nova.virt.hardware [None req-731394e9-d0a3-46c3-8b02-41ef72d369f5 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 982.451583] env[61957]: DEBUG nova.virt.hardware [None req-731394e9-d0a3-46c3-8b02-41ef72d369f5 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 982.451773] env[61957]: DEBUG nova.virt.hardware [None req-731394e9-d0a3-46c3-8b02-41ef72d369f5 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 982.451961] env[61957]: DEBUG nova.virt.hardware [None req-731394e9-d0a3-46c3-8b02-41ef72d369f5 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Got 1 possible topologies {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 982.452152] env[61957]: DEBUG nova.virt.hardware [None req-731394e9-d0a3-46c3-8b02-41ef72d369f5 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 982.452379] env[61957]: DEBUG nova.virt.hardware [None req-731394e9-d0a3-46c3-8b02-41ef72d369f5 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 982.453662] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efd4207b-2b68-4aa3-8936-19d6ee52f8e8 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.467918] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bf8a954-cefa-404c-83f4-ae5f44be9b37 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.476588] env[61957]: DEBUG oslo_concurrency.lockutils [None req-06b74b35-57d5-4a4b-8055-77adbdba299a tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Lock "00ecc8c2-398e-4a29-b19f-a2013985b481" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 12.078s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 982.498141] env[61957]: DEBUG oslo_vmware.api [None req-5145e111-afa1-43c9-8275-d20aceeb3d2b tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': task-1277877, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.513876] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-b10a76eb-ac59-4162-9083-648b4df991d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 98929ed3-d420-45cd-9cde-9738b2ac8251] Creating Snapshot of the VM instance {{(pid=61957) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 982.514064] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-bc717405-a31d-4c88-ba01-eb92a66744f8 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.523117] env[61957]: DEBUG oslo_vmware.api [None req-b10a76eb-ac59-4162-9083-648b4df991d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Waiting for the task: (returnval){ [ 982.523117] env[61957]: value = "task-1277880" [ 982.523117] env[61957]: _type = "Task" [ 982.523117] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.531607] env[61957]: DEBUG oslo_concurrency.lockutils [None req-bc1354f0-0ce9-464c-89e1-e1e216546ebe tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Acquiring lock "6509707e-afc6-4c4a-88f9-a471865d510c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 982.531960] env[61957]: DEBUG oslo_concurrency.lockutils [None req-bc1354f0-0ce9-464c-89e1-e1e216546ebe tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Lock "6509707e-afc6-4c4a-88f9-a471865d510c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 982.540082] env[61957]: DEBUG oslo_vmware.api [None req-b10a76eb-ac59-4162-9083-648b4df991d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1277880, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.563975] env[61957]: DEBUG oslo_vmware.api [None req-c255bb00-cba6-445a-adb0-12124fe511af tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': task-1277879, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.29819} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 982.563975] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-c255bb00-cba6-445a-adb0-12124fe511af tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Deleted the datastore file {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 982.563975] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-c255bb00-cba6-445a-adb0-12124fe511af tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: dbeeb200-70b9-4cb4-b5a4-182389d21918] Deleted contents of the VM from datastore datastore1 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 982.563975] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-c255bb00-cba6-445a-adb0-12124fe511af tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: dbeeb200-70b9-4cb4-b5a4-182389d21918] Instance destroyed {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 982.563975] env[61957]: INFO nova.compute.manager [None req-c255bb00-cba6-445a-adb0-12124fe511af tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: dbeeb200-70b9-4cb4-b5a4-182389d21918] Took 1.55 seconds to destroy the instance on the hypervisor. [ 982.563975] env[61957]: DEBUG oslo.service.loopingcall [None req-c255bb00-cba6-445a-adb0-12124fe511af tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 982.564800] env[61957]: DEBUG nova.compute.manager [-] [instance: dbeeb200-70b9-4cb4-b5a4-182389d21918] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 982.564800] env[61957]: DEBUG nova.network.neutron [-] [instance: dbeeb200-70b9-4cb4-b5a4-182389d21918] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 982.566783] env[61957]: DEBUG nova.network.neutron [None req-731394e9-d0a3-46c3-8b02-41ef72d369f5 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] [instance: f65458b2-0db6-4d98-8c82-b27b6945a3fd] Successfully updated port: bb715e7d-14c5-418b-b3ee-1ac91aba709d {{(pid=61957) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 982.768980] env[61957]: DEBUG nova.network.neutron [-] [instance: d6c5c70d-86de-4dea-8b9e-76f321947a35] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 982.887998] env[61957]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-68d4a6d1-a70b-41f6-862f-55b4f13eb16a {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.898970] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-468fd5c2-637a-404b-91d9-677acff73f65 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.941324] env[61957]: DEBUG nova.compute.manager [req-7b576ccc-cea9-4eb4-b2d5-414bd8cb987d req-f47b492f-3c4f-4405-979b-b77735b8e215 service nova] [instance: d6c5c70d-86de-4dea-8b9e-76f321947a35] Detach interface failed, port_id=642fc1a5-237a-4bde-b28f-69b3f0f46a8a, reason: Instance d6c5c70d-86de-4dea-8b9e-76f321947a35 could not be found. {{(pid=61957) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 983.004652] env[61957]: DEBUG oslo_vmware.api [None req-5145e111-afa1-43c9-8275-d20aceeb3d2b tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': task-1277877, 'name': PowerOnVM_Task, 'duration_secs': 0.602354} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.004652] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-5145e111-afa1-43c9-8275-d20aceeb3d2b tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 6f45c7c8-de07-4d21-bccc-e4afe3c8ad3d] Powered on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 983.004652] env[61957]: INFO nova.compute.manager [None req-5145e111-afa1-43c9-8275-d20aceeb3d2b tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 6f45c7c8-de07-4d21-bccc-e4afe3c8ad3d] Took 8.92 seconds to spawn the instance on the hypervisor. [ 983.004652] env[61957]: DEBUG nova.compute.manager [None req-5145e111-afa1-43c9-8275-d20aceeb3d2b tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 6f45c7c8-de07-4d21-bccc-e4afe3c8ad3d] Checking state {{(pid=61957) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 983.004652] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc312e5b-aa91-45fc-9826-f96565621993 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.034173] env[61957]: DEBUG oslo_vmware.api [None req-b10a76eb-ac59-4162-9083-648b4df991d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1277880, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.040874] env[61957]: DEBUG nova.compute.manager [None req-bc1354f0-0ce9-464c-89e1-e1e216546ebe tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: 6509707e-afc6-4c4a-88f9-a471865d510c] Starting instance... {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 983.071171] env[61957]: DEBUG oslo_concurrency.lockutils [None req-731394e9-d0a3-46c3-8b02-41ef72d369f5 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Acquiring lock "refresh_cache-f65458b2-0db6-4d98-8c82-b27b6945a3fd" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 983.071171] env[61957]: DEBUG oslo_concurrency.lockutils [None req-731394e9-d0a3-46c3-8b02-41ef72d369f5 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Acquired lock "refresh_cache-f65458b2-0db6-4d98-8c82-b27b6945a3fd" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 983.071171] env[61957]: DEBUG nova.network.neutron [None req-731394e9-d0a3-46c3-8b02-41ef72d369f5 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] [instance: f65458b2-0db6-4d98-8c82-b27b6945a3fd] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 983.170124] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bdeecd63-e20f-4590-a01b-8b02f80fe7ed {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.178338] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ae0e7f6-0d1e-4666-9885-eecd4683560f {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.209778] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d79be237-87d9-4470-9622-01c403ac1e67 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.217665] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0e0e3b7-d614-4174-95e7-b8ac0f4b678b {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.231144] env[61957]: DEBUG nova.compute.provider_tree [None req-8b32736b-e3a6-485c-a95e-afec23b58d86 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 983.271548] env[61957]: INFO nova.compute.manager [-] [instance: d6c5c70d-86de-4dea-8b9e-76f321947a35] Took 1.32 seconds to deallocate network for instance. [ 983.535327] env[61957]: DEBUG nova.network.neutron [-] [instance: dbeeb200-70b9-4cb4-b5a4-182389d21918] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 983.536629] env[61957]: INFO nova.compute.manager [None req-5145e111-afa1-43c9-8275-d20aceeb3d2b tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 6f45c7c8-de07-4d21-bccc-e4afe3c8ad3d] Took 29.14 seconds to build instance. [ 983.541401] env[61957]: DEBUG oslo_vmware.api [None req-b10a76eb-ac59-4162-9083-648b4df991d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1277880, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.565833] env[61957]: DEBUG oslo_concurrency.lockutils [None req-bc1354f0-0ce9-464c-89e1-e1e216546ebe tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 983.603306] env[61957]: DEBUG nova.network.neutron [None req-731394e9-d0a3-46c3-8b02-41ef72d369f5 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] [instance: f65458b2-0db6-4d98-8c82-b27b6945a3fd] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 983.728354] env[61957]: DEBUG nova.network.neutron [None req-731394e9-d0a3-46c3-8b02-41ef72d369f5 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] [instance: f65458b2-0db6-4d98-8c82-b27b6945a3fd] Updating instance_info_cache with network_info: [{"id": "bb715e7d-14c5-418b-b3ee-1ac91aba709d", "address": "fa:16:3e:fd:f1:f4", "network": {"id": "d55c72c6-107d-4965-b4e1-d18383813953", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-193304765-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.2", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "7758bdd10505455b8cecb8129d70ebd5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c24464bb-bb6b-43a2-bdcd-8086ad1a307f", "external-id": "nsx-vlan-transportzone-781", "segmentation_id": 781, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbb715e7d-14", "ovs_interfaceid": "bb715e7d-14c5-418b-b3ee-1ac91aba709d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 983.733938] env[61957]: DEBUG nova.scheduler.client.report [None req-8b32736b-e3a6-485c-a95e-afec23b58d86 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 983.780182] env[61957]: DEBUG oslo_concurrency.lockutils [None req-2f034be2-1f7c-41d0-91ad-8c549dbb7f8a tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 984.039337] env[61957]: DEBUG oslo_vmware.api [None req-b10a76eb-ac59-4162-9083-648b4df991d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1277880, 'name': CreateSnapshot_Task, 'duration_secs': 1.421998} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.039661] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-b10a76eb-ac59-4162-9083-648b4df991d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 98929ed3-d420-45cd-9cde-9738b2ac8251] Created Snapshot of the VM instance {{(pid=61957) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 984.040839] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d789013-b7dc-4ba9-b25f-4bc29e2033d2 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.045148] env[61957]: DEBUG oslo_concurrency.lockutils [None req-5145e111-afa1-43c9-8275-d20aceeb3d2b tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Lock "6f45c7c8-de07-4d21-bccc-e4afe3c8ad3d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 30.660s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 984.045148] env[61957]: INFO nova.compute.manager [-] [instance: dbeeb200-70b9-4cb4-b5a4-182389d21918] Took 1.48 seconds to deallocate network for instance. [ 984.048438] env[61957]: DEBUG oslo_concurrency.lockutils [None req-74e1032e-bbb9-4915-a886-4f8885996564 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Acquiring lock "6f45c7c8-de07-4d21-bccc-e4afe3c8ad3d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 984.048724] env[61957]: DEBUG oslo_concurrency.lockutils [None req-74e1032e-bbb9-4915-a886-4f8885996564 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Lock "6f45c7c8-de07-4d21-bccc-e4afe3c8ad3d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 984.048954] env[61957]: DEBUG oslo_concurrency.lockutils [None req-74e1032e-bbb9-4915-a886-4f8885996564 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Acquiring lock "6f45c7c8-de07-4d21-bccc-e4afe3c8ad3d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 984.049159] env[61957]: DEBUG oslo_concurrency.lockutils [None req-74e1032e-bbb9-4915-a886-4f8885996564 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Lock "6f45c7c8-de07-4d21-bccc-e4afe3c8ad3d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 984.049332] env[61957]: DEBUG oslo_concurrency.lockutils [None req-74e1032e-bbb9-4915-a886-4f8885996564 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Lock "6f45c7c8-de07-4d21-bccc-e4afe3c8ad3d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 984.054157] env[61957]: INFO nova.compute.manager [None req-74e1032e-bbb9-4915-a886-4f8885996564 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 6f45c7c8-de07-4d21-bccc-e4afe3c8ad3d] Terminating instance [ 984.056705] env[61957]: DEBUG nova.compute.manager [None req-74e1032e-bbb9-4915-a886-4f8885996564 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 6f45c7c8-de07-4d21-bccc-e4afe3c8ad3d] Start destroying the instance on the hypervisor. {{(pid=61957) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 984.056905] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-74e1032e-bbb9-4915-a886-4f8885996564 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 6f45c7c8-de07-4d21-bccc-e4afe3c8ad3d] Destroying instance {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 984.059943] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29af4145-241e-4422-a2e5-824317cb6a15 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.069026] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-74e1032e-bbb9-4915-a886-4f8885996564 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 6f45c7c8-de07-4d21-bccc-e4afe3c8ad3d] Powering off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 984.069303] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a73abef1-4cec-407e-9453-97c86a58244d {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.077258] env[61957]: DEBUG oslo_vmware.api [None req-74e1032e-bbb9-4915-a886-4f8885996564 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Waiting for the task: (returnval){ [ 984.077258] env[61957]: value = "task-1277881" [ 984.077258] env[61957]: _type = "Task" [ 984.077258] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.088280] env[61957]: DEBUG oslo_vmware.api [None req-74e1032e-bbb9-4915-a886-4f8885996564 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': task-1277881, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.230872] env[61957]: DEBUG oslo_concurrency.lockutils [None req-731394e9-d0a3-46c3-8b02-41ef72d369f5 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Releasing lock "refresh_cache-f65458b2-0db6-4d98-8c82-b27b6945a3fd" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 984.231289] env[61957]: DEBUG nova.compute.manager [None req-731394e9-d0a3-46c3-8b02-41ef72d369f5 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] [instance: f65458b2-0db6-4d98-8c82-b27b6945a3fd] Instance network_info: |[{"id": "bb715e7d-14c5-418b-b3ee-1ac91aba709d", "address": "fa:16:3e:fd:f1:f4", "network": {"id": "d55c72c6-107d-4965-b4e1-d18383813953", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-193304765-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.2", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "7758bdd10505455b8cecb8129d70ebd5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c24464bb-bb6b-43a2-bdcd-8086ad1a307f", "external-id": "nsx-vlan-transportzone-781", "segmentation_id": 781, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbb715e7d-14", "ovs_interfaceid": "bb715e7d-14c5-418b-b3ee-1ac91aba709d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 984.231807] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-731394e9-d0a3-46c3-8b02-41ef72d369f5 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] [instance: f65458b2-0db6-4d98-8c82-b27b6945a3fd] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:fd:f1:f4', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c24464bb-bb6b-43a2-bdcd-8086ad1a307f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'bb715e7d-14c5-418b-b3ee-1ac91aba709d', 'vif_model': 'vmxnet3'}] {{(pid=61957) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 984.240366] env[61957]: DEBUG oslo.service.loopingcall [None req-731394e9-d0a3-46c3-8b02-41ef72d369f5 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 984.241252] env[61957]: DEBUG oslo_concurrency.lockutils [None req-8b32736b-e3a6-485c-a95e-afec23b58d86 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.338s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 984.243569] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f65458b2-0db6-4d98-8c82-b27b6945a3fd] Creating VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 984.243992] env[61957]: DEBUG oslo_concurrency.lockutils [None req-6d13ccf5-9ccd-48d0-a35e-c6df97be9c15 tempest-ServerAddressesTestJSON-2054963291 tempest-ServerAddressesTestJSON-2054963291-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.809s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 984.245651] env[61957]: INFO nova.compute.claims [None req-6d13ccf5-9ccd-48d0-a35e-c6df97be9c15 tempest-ServerAddressesTestJSON-2054963291 tempest-ServerAddressesTestJSON-2054963291-project-member] [instance: fa4c150e-7a85-41a4-8890-9f2deb3b34b1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 984.248186] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-08810167-4018-41ad-b4a7-c6c64f8c69d6 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.272368] env[61957]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 984.272368] env[61957]: value = "task-1277882" [ 984.272368] env[61957]: _type = "Task" [ 984.272368] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.281132] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277882, 'name': CreateVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.300057] env[61957]: INFO nova.network.neutron [None req-8b32736b-e3a6-485c-a95e-afec23b58d86 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] [instance: 736eabcf-70f8-4e1e-9830-270d9971fc9a] Updating port 95763a65-6383-409f-a9d6-1f37aaf44d34 with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 984.449694] env[61957]: DEBUG nova.compute.manager [req-0cab5fe0-de1a-4a99-95c8-41d3812bfa3b req-ecac115c-69c9-4ab9-b3a2-f14e60328b80 service nova] [instance: f65458b2-0db6-4d98-8c82-b27b6945a3fd] Received event network-vif-plugged-bb715e7d-14c5-418b-b3ee-1ac91aba709d {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 984.449984] env[61957]: DEBUG oslo_concurrency.lockutils [req-0cab5fe0-de1a-4a99-95c8-41d3812bfa3b req-ecac115c-69c9-4ab9-b3a2-f14e60328b80 service nova] Acquiring lock "f65458b2-0db6-4d98-8c82-b27b6945a3fd-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 984.450263] env[61957]: DEBUG oslo_concurrency.lockutils [req-0cab5fe0-de1a-4a99-95c8-41d3812bfa3b req-ecac115c-69c9-4ab9-b3a2-f14e60328b80 service nova] Lock "f65458b2-0db6-4d98-8c82-b27b6945a3fd-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 984.450465] env[61957]: DEBUG oslo_concurrency.lockutils [req-0cab5fe0-de1a-4a99-95c8-41d3812bfa3b req-ecac115c-69c9-4ab9-b3a2-f14e60328b80 service nova] Lock "f65458b2-0db6-4d98-8c82-b27b6945a3fd-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 984.450666] env[61957]: DEBUG nova.compute.manager [req-0cab5fe0-de1a-4a99-95c8-41d3812bfa3b req-ecac115c-69c9-4ab9-b3a2-f14e60328b80 service nova] [instance: f65458b2-0db6-4d98-8c82-b27b6945a3fd] No waiting events found dispatching network-vif-plugged-bb715e7d-14c5-418b-b3ee-1ac91aba709d {{(pid=61957) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 984.450875] env[61957]: WARNING nova.compute.manager [req-0cab5fe0-de1a-4a99-95c8-41d3812bfa3b req-ecac115c-69c9-4ab9-b3a2-f14e60328b80 service nova] [instance: f65458b2-0db6-4d98-8c82-b27b6945a3fd] Received unexpected event network-vif-plugged-bb715e7d-14c5-418b-b3ee-1ac91aba709d for instance with vm_state building and task_state spawning. [ 984.451090] env[61957]: DEBUG nova.compute.manager [req-0cab5fe0-de1a-4a99-95c8-41d3812bfa3b req-ecac115c-69c9-4ab9-b3a2-f14e60328b80 service nova] [instance: f65458b2-0db6-4d98-8c82-b27b6945a3fd] Received event network-changed-bb715e7d-14c5-418b-b3ee-1ac91aba709d {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 984.451287] env[61957]: DEBUG nova.compute.manager [req-0cab5fe0-de1a-4a99-95c8-41d3812bfa3b req-ecac115c-69c9-4ab9-b3a2-f14e60328b80 service nova] [instance: f65458b2-0db6-4d98-8c82-b27b6945a3fd] Refreshing instance network info cache due to event network-changed-bb715e7d-14c5-418b-b3ee-1ac91aba709d. {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 984.451506] env[61957]: DEBUG oslo_concurrency.lockutils [req-0cab5fe0-de1a-4a99-95c8-41d3812bfa3b req-ecac115c-69c9-4ab9-b3a2-f14e60328b80 service nova] Acquiring lock "refresh_cache-f65458b2-0db6-4d98-8c82-b27b6945a3fd" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 984.451716] env[61957]: DEBUG oslo_concurrency.lockutils [req-0cab5fe0-de1a-4a99-95c8-41d3812bfa3b req-ecac115c-69c9-4ab9-b3a2-f14e60328b80 service nova] Acquired lock "refresh_cache-f65458b2-0db6-4d98-8c82-b27b6945a3fd" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 984.452038] env[61957]: DEBUG nova.network.neutron [req-0cab5fe0-de1a-4a99-95c8-41d3812bfa3b req-ecac115c-69c9-4ab9-b3a2-f14e60328b80 service nova] [instance: f65458b2-0db6-4d98-8c82-b27b6945a3fd] Refreshing network info cache for port bb715e7d-14c5-418b-b3ee-1ac91aba709d {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 984.570200] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-b10a76eb-ac59-4162-9083-648b4df991d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 98929ed3-d420-45cd-9cde-9738b2ac8251] Creating linked-clone VM from snapshot {{(pid=61957) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 984.572682] env[61957]: DEBUG oslo_concurrency.lockutils [None req-c255bb00-cba6-445a-adb0-12124fe511af tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 984.572682] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-0fa57222-1d6c-4b1c-912b-cbd2afd6626a {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.586031] env[61957]: DEBUG oslo_vmware.api [None req-b10a76eb-ac59-4162-9083-648b4df991d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Waiting for the task: (returnval){ [ 984.586031] env[61957]: value = "task-1277883" [ 984.586031] env[61957]: _type = "Task" [ 984.586031] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.591042] env[61957]: DEBUG oslo_vmware.api [None req-74e1032e-bbb9-4915-a886-4f8885996564 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': task-1277881, 'name': PowerOffVM_Task, 'duration_secs': 0.174556} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.594186] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-74e1032e-bbb9-4915-a886-4f8885996564 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 6f45c7c8-de07-4d21-bccc-e4afe3c8ad3d] Powered off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 984.594400] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-74e1032e-bbb9-4915-a886-4f8885996564 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 6f45c7c8-de07-4d21-bccc-e4afe3c8ad3d] Unregistering the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 984.594676] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b5c65384-242d-4dad-8eb3-3ec28bec8d4d {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.604590] env[61957]: DEBUG oslo_vmware.api [None req-b10a76eb-ac59-4162-9083-648b4df991d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1277883, 'name': CloneVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.668969] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-74e1032e-bbb9-4915-a886-4f8885996564 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 6f45c7c8-de07-4d21-bccc-e4afe3c8ad3d] Unregistered the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 984.668969] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-74e1032e-bbb9-4915-a886-4f8885996564 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 6f45c7c8-de07-4d21-bccc-e4afe3c8ad3d] Deleting contents of the VM from datastore datastore1 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 984.668969] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-74e1032e-bbb9-4915-a886-4f8885996564 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Deleting the datastore file [datastore1] 6f45c7c8-de07-4d21-bccc-e4afe3c8ad3d {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 984.668969] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-22c39640-216f-42e9-b60e-5ecebed5d973 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.676543] env[61957]: DEBUG oslo_vmware.api [None req-74e1032e-bbb9-4915-a886-4f8885996564 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Waiting for the task: (returnval){ [ 984.676543] env[61957]: value = "task-1277885" [ 984.676543] env[61957]: _type = "Task" [ 984.676543] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.685761] env[61957]: DEBUG oslo_vmware.api [None req-74e1032e-bbb9-4915-a886-4f8885996564 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': task-1277885, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.783127] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277882, 'name': CreateVM_Task, 'duration_secs': 0.350315} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.783296] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f65458b2-0db6-4d98-8c82-b27b6945a3fd] Created VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 984.784110] env[61957]: DEBUG oslo_concurrency.lockutils [None req-731394e9-d0a3-46c3-8b02-41ef72d369f5 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 984.784231] env[61957]: DEBUG oslo_concurrency.lockutils [None req-731394e9-d0a3-46c3-8b02-41ef72d369f5 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Acquired lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 984.784548] env[61957]: DEBUG oslo_concurrency.lockutils [None req-731394e9-d0a3-46c3-8b02-41ef72d369f5 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 984.784848] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d05c6adc-ac62-4935-854c-0190e4f72931 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.789889] env[61957]: DEBUG oslo_vmware.api [None req-731394e9-d0a3-46c3-8b02-41ef72d369f5 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Waiting for the task: (returnval){ [ 984.789889] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52b9bba4-0f52-826e-636d-d4880c536553" [ 984.789889] env[61957]: _type = "Task" [ 984.789889] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.798397] env[61957]: DEBUG oslo_vmware.api [None req-731394e9-d0a3-46c3-8b02-41ef72d369f5 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52b9bba4-0f52-826e-636d-d4880c536553, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.099801] env[61957]: DEBUG oslo_vmware.api [None req-b10a76eb-ac59-4162-9083-648b4df991d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1277883, 'name': CloneVM_Task} progress is 94%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.147096] env[61957]: DEBUG nova.network.neutron [req-0cab5fe0-de1a-4a99-95c8-41d3812bfa3b req-ecac115c-69c9-4ab9-b3a2-f14e60328b80 service nova] [instance: f65458b2-0db6-4d98-8c82-b27b6945a3fd] Updated VIF entry in instance network info cache for port bb715e7d-14c5-418b-b3ee-1ac91aba709d. {{(pid=61957) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 985.147096] env[61957]: DEBUG nova.network.neutron [req-0cab5fe0-de1a-4a99-95c8-41d3812bfa3b req-ecac115c-69c9-4ab9-b3a2-f14e60328b80 service nova] [instance: f65458b2-0db6-4d98-8c82-b27b6945a3fd] Updating instance_info_cache with network_info: [{"id": "bb715e7d-14c5-418b-b3ee-1ac91aba709d", "address": "fa:16:3e:fd:f1:f4", "network": {"id": "d55c72c6-107d-4965-b4e1-d18383813953", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-193304765-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.2", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "7758bdd10505455b8cecb8129d70ebd5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c24464bb-bb6b-43a2-bdcd-8086ad1a307f", "external-id": "nsx-vlan-transportzone-781", "segmentation_id": 781, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbb715e7d-14", "ovs_interfaceid": "bb715e7d-14c5-418b-b3ee-1ac91aba709d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 985.187806] env[61957]: DEBUG oslo_vmware.api [None req-74e1032e-bbb9-4915-a886-4f8885996564 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': task-1277885, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.152688} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.188131] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-74e1032e-bbb9-4915-a886-4f8885996564 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Deleted the datastore file {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 985.188299] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-74e1032e-bbb9-4915-a886-4f8885996564 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 6f45c7c8-de07-4d21-bccc-e4afe3c8ad3d] Deleted contents of the VM from datastore datastore1 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 985.188470] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-74e1032e-bbb9-4915-a886-4f8885996564 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 6f45c7c8-de07-4d21-bccc-e4afe3c8ad3d] Instance destroyed {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 985.188646] env[61957]: INFO nova.compute.manager [None req-74e1032e-bbb9-4915-a886-4f8885996564 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 6f45c7c8-de07-4d21-bccc-e4afe3c8ad3d] Took 1.13 seconds to destroy the instance on the hypervisor. [ 985.188889] env[61957]: DEBUG oslo.service.loopingcall [None req-74e1032e-bbb9-4915-a886-4f8885996564 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 985.189364] env[61957]: DEBUG nova.compute.manager [-] [instance: 6f45c7c8-de07-4d21-bccc-e4afe3c8ad3d] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 985.189466] env[61957]: DEBUG nova.network.neutron [-] [instance: 6f45c7c8-de07-4d21-bccc-e4afe3c8ad3d] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 985.304251] env[61957]: DEBUG oslo_vmware.api [None req-731394e9-d0a3-46c3-8b02-41ef72d369f5 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52b9bba4-0f52-826e-636d-d4880c536553, 'name': SearchDatastore_Task, 'duration_secs': 0.010195} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.304661] env[61957]: DEBUG oslo_concurrency.lockutils [None req-731394e9-d0a3-46c3-8b02-41ef72d369f5 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Releasing lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 985.304990] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-731394e9-d0a3-46c3-8b02-41ef72d369f5 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] [instance: f65458b2-0db6-4d98-8c82-b27b6945a3fd] Processing image 11c76a2c-f705-470a-ba9d-4657858bab38 {{(pid=61957) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 985.305408] env[61957]: DEBUG oslo_concurrency.lockutils [None req-731394e9-d0a3-46c3-8b02-41ef72d369f5 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 985.305562] env[61957]: DEBUG oslo_concurrency.lockutils [None req-731394e9-d0a3-46c3-8b02-41ef72d369f5 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Acquired lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 985.305827] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-731394e9-d0a3-46c3-8b02-41ef72d369f5 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 985.306216] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-fa9a4839-e527-4812-96d2-2a2fcf7f1722 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.322134] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-731394e9-d0a3-46c3-8b02-41ef72d369f5 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 985.322464] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-731394e9-d0a3-46c3-8b02-41ef72d369f5 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61957) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 985.326814] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-025432e9-6f9b-4368-bff7-0fef537e5fe2 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.335240] env[61957]: DEBUG oslo_vmware.api [None req-731394e9-d0a3-46c3-8b02-41ef72d369f5 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Waiting for the task: (returnval){ [ 985.335240] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52d50aa8-5cd9-5f4a-8dc4-d4493e948d93" [ 985.335240] env[61957]: _type = "Task" [ 985.335240] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.345220] env[61957]: DEBUG oslo_vmware.api [None req-731394e9-d0a3-46c3-8b02-41ef72d369f5 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52d50aa8-5cd9-5f4a-8dc4-d4493e948d93, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.494430] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b70a1d0-bb34-462b-b5a8-0435254bfcdf {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.502899] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bebbd5eb-e3de-4d2e-83d1-63a163a33e5f {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.534996] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-022b7641-784c-407a-a610-eb7c7a62943d {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.543231] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db8d1df7-78e7-47bd-a1b7-cb8ea7aab631 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.557235] env[61957]: DEBUG nova.compute.provider_tree [None req-6d13ccf5-9ccd-48d0-a35e-c6df97be9c15 tempest-ServerAddressesTestJSON-2054963291 tempest-ServerAddressesTestJSON-2054963291-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 985.601802] env[61957]: DEBUG oslo_vmware.api [None req-b10a76eb-ac59-4162-9083-648b4df991d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1277883, 'name': CloneVM_Task} progress is 94%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.649222] env[61957]: DEBUG oslo_concurrency.lockutils [req-0cab5fe0-de1a-4a99-95c8-41d3812bfa3b req-ecac115c-69c9-4ab9-b3a2-f14e60328b80 service nova] Releasing lock "refresh_cache-f65458b2-0db6-4d98-8c82-b27b6945a3fd" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 985.649519] env[61957]: DEBUG nova.compute.manager [req-0cab5fe0-de1a-4a99-95c8-41d3812bfa3b req-ecac115c-69c9-4ab9-b3a2-f14e60328b80 service nova] [instance: dbeeb200-70b9-4cb4-b5a4-182389d21918] Received event network-vif-deleted-86553395-da58-4c44-b9f4-c67db304d3fe {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 985.850023] env[61957]: DEBUG oslo_vmware.api [None req-731394e9-d0a3-46c3-8b02-41ef72d369f5 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52d50aa8-5cd9-5f4a-8dc4-d4493e948d93, 'name': SearchDatastore_Task, 'duration_secs': 0.00962} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.850023] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e4379e1e-c564-4969-ad15-25c6b44632ba {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.855502] env[61957]: DEBUG oslo_vmware.api [None req-731394e9-d0a3-46c3-8b02-41ef72d369f5 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Waiting for the task: (returnval){ [ 985.855502] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]5282cf9c-6236-1984-7223-1bdbe4dd88e6" [ 985.855502] env[61957]: _type = "Task" [ 985.855502] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.860632] env[61957]: DEBUG oslo_concurrency.lockutils [None req-8b32736b-e3a6-485c-a95e-afec23b58d86 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Acquiring lock "refresh_cache-736eabcf-70f8-4e1e-9830-270d9971fc9a" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 985.860929] env[61957]: DEBUG oslo_concurrency.lockutils [None req-8b32736b-e3a6-485c-a95e-afec23b58d86 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Acquired lock "refresh_cache-736eabcf-70f8-4e1e-9830-270d9971fc9a" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 985.861136] env[61957]: DEBUG nova.network.neutron [None req-8b32736b-e3a6-485c-a95e-afec23b58d86 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] [instance: 736eabcf-70f8-4e1e-9830-270d9971fc9a] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 985.866186] env[61957]: DEBUG oslo_vmware.api [None req-731394e9-d0a3-46c3-8b02-41ef72d369f5 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]5282cf9c-6236-1984-7223-1bdbe4dd88e6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.970220] env[61957]: DEBUG nova.network.neutron [-] [instance: 6f45c7c8-de07-4d21-bccc-e4afe3c8ad3d] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 986.063406] env[61957]: DEBUG nova.scheduler.client.report [None req-6d13ccf5-9ccd-48d0-a35e-c6df97be9c15 tempest-ServerAddressesTestJSON-2054963291 tempest-ServerAddressesTestJSON-2054963291-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 986.102071] env[61957]: DEBUG oslo_vmware.api [None req-b10a76eb-ac59-4162-9083-648b4df991d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1277883, 'name': CloneVM_Task} progress is 95%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.369947] env[61957]: DEBUG oslo_vmware.api [None req-731394e9-d0a3-46c3-8b02-41ef72d369f5 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]5282cf9c-6236-1984-7223-1bdbe4dd88e6, 'name': SearchDatastore_Task, 'duration_secs': 0.011629} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.370337] env[61957]: DEBUG oslo_concurrency.lockutils [None req-731394e9-d0a3-46c3-8b02-41ef72d369f5 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Releasing lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 986.370647] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-731394e9-d0a3-46c3-8b02-41ef72d369f5 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore1] f65458b2-0db6-4d98-8c82-b27b6945a3fd/f65458b2-0db6-4d98-8c82-b27b6945a3fd.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 986.371268] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d947b82a-357a-4d53-b2a2-c6779b809167 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.379256] env[61957]: DEBUG oslo_vmware.api [None req-731394e9-d0a3-46c3-8b02-41ef72d369f5 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Waiting for the task: (returnval){ [ 986.379256] env[61957]: value = "task-1277886" [ 986.379256] env[61957]: _type = "Task" [ 986.379256] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.390060] env[61957]: DEBUG oslo_vmware.api [None req-731394e9-d0a3-46c3-8b02-41ef72d369f5 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Task: {'id': task-1277886, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.473285] env[61957]: INFO nova.compute.manager [-] [instance: 6f45c7c8-de07-4d21-bccc-e4afe3c8ad3d] Took 1.28 seconds to deallocate network for instance. [ 986.484582] env[61957]: DEBUG nova.compute.manager [req-94058206-e54b-4216-9144-249674e2d939 req-052a2dcd-259c-466e-94d3-84bf68fe7614 service nova] [instance: 6f45c7c8-de07-4d21-bccc-e4afe3c8ad3d] Received event network-vif-deleted-e0a600d6-e6d6-4598-8f1d-647ba04dd492 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 986.484813] env[61957]: DEBUG nova.compute.manager [req-94058206-e54b-4216-9144-249674e2d939 req-052a2dcd-259c-466e-94d3-84bf68fe7614 service nova] [instance: 736eabcf-70f8-4e1e-9830-270d9971fc9a] Received event network-vif-plugged-95763a65-6383-409f-a9d6-1f37aaf44d34 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 986.485064] env[61957]: DEBUG oslo_concurrency.lockutils [req-94058206-e54b-4216-9144-249674e2d939 req-052a2dcd-259c-466e-94d3-84bf68fe7614 service nova] Acquiring lock "736eabcf-70f8-4e1e-9830-270d9971fc9a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 986.485232] env[61957]: DEBUG oslo_concurrency.lockutils [req-94058206-e54b-4216-9144-249674e2d939 req-052a2dcd-259c-466e-94d3-84bf68fe7614 service nova] Lock "736eabcf-70f8-4e1e-9830-270d9971fc9a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 986.485823] env[61957]: DEBUG oslo_concurrency.lockutils [req-94058206-e54b-4216-9144-249674e2d939 req-052a2dcd-259c-466e-94d3-84bf68fe7614 service nova] Lock "736eabcf-70f8-4e1e-9830-270d9971fc9a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 986.485823] env[61957]: DEBUG nova.compute.manager [req-94058206-e54b-4216-9144-249674e2d939 req-052a2dcd-259c-466e-94d3-84bf68fe7614 service nova] [instance: 736eabcf-70f8-4e1e-9830-270d9971fc9a] No waiting events found dispatching network-vif-plugged-95763a65-6383-409f-a9d6-1f37aaf44d34 {{(pid=61957) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 986.485823] env[61957]: WARNING nova.compute.manager [req-94058206-e54b-4216-9144-249674e2d939 req-052a2dcd-259c-466e-94d3-84bf68fe7614 service nova] [instance: 736eabcf-70f8-4e1e-9830-270d9971fc9a] Received unexpected event network-vif-plugged-95763a65-6383-409f-a9d6-1f37aaf44d34 for instance with vm_state shelved_offloaded and task_state spawning. [ 986.485981] env[61957]: DEBUG nova.compute.manager [req-94058206-e54b-4216-9144-249674e2d939 req-052a2dcd-259c-466e-94d3-84bf68fe7614 service nova] [instance: 736eabcf-70f8-4e1e-9830-270d9971fc9a] Received event network-changed-95763a65-6383-409f-a9d6-1f37aaf44d34 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 986.486063] env[61957]: DEBUG nova.compute.manager [req-94058206-e54b-4216-9144-249674e2d939 req-052a2dcd-259c-466e-94d3-84bf68fe7614 service nova] [instance: 736eabcf-70f8-4e1e-9830-270d9971fc9a] Refreshing instance network info cache due to event network-changed-95763a65-6383-409f-a9d6-1f37aaf44d34. {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 986.486249] env[61957]: DEBUG oslo_concurrency.lockutils [req-94058206-e54b-4216-9144-249674e2d939 req-052a2dcd-259c-466e-94d3-84bf68fe7614 service nova] Acquiring lock "refresh_cache-736eabcf-70f8-4e1e-9830-270d9971fc9a" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 986.568937] env[61957]: DEBUG oslo_concurrency.lockutils [None req-6d13ccf5-9ccd-48d0-a35e-c6df97be9c15 tempest-ServerAddressesTestJSON-2054963291 tempest-ServerAddressesTestJSON-2054963291-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.325s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 986.570016] env[61957]: DEBUG nova.compute.manager [None req-6d13ccf5-9ccd-48d0-a35e-c6df97be9c15 tempest-ServerAddressesTestJSON-2054963291 tempest-ServerAddressesTestJSON-2054963291-project-member] [instance: fa4c150e-7a85-41a4-8890-9f2deb3b34b1] Start building networks asynchronously for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 986.572167] env[61957]: DEBUG oslo_concurrency.lockutils [None req-bc1354f0-0ce9-464c-89e1-e1e216546ebe tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.006s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 986.573656] env[61957]: INFO nova.compute.claims [None req-bc1354f0-0ce9-464c-89e1-e1e216546ebe tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: 6509707e-afc6-4c4a-88f9-a471865d510c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 986.606654] env[61957]: DEBUG oslo_vmware.api [None req-b10a76eb-ac59-4162-9083-648b4df991d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1277883, 'name': CloneVM_Task, 'duration_secs': 1.820597} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.607155] env[61957]: INFO nova.virt.vmwareapi.vmops [None req-b10a76eb-ac59-4162-9083-648b4df991d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 98929ed3-d420-45cd-9cde-9738b2ac8251] Created linked-clone VM from snapshot [ 986.608372] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e23a703-cc9a-4b3d-bf16-8217c1553609 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.622016] env[61957]: DEBUG nova.virt.vmwareapi.images [None req-b10a76eb-ac59-4162-9083-648b4df991d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 98929ed3-d420-45cd-9cde-9738b2ac8251] Uploading image 74e02d56-c151-4bb1-b998-cd7ba6872fac {{(pid=61957) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 986.625363] env[61957]: DEBUG nova.network.neutron [None req-8b32736b-e3a6-485c-a95e-afec23b58d86 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] [instance: 736eabcf-70f8-4e1e-9830-270d9971fc9a] Updating instance_info_cache with network_info: [{"id": "95763a65-6383-409f-a9d6-1f37aaf44d34", "address": "fa:16:3e:64:07:83", "network": {"id": "b303ad01-ef9b-4119-9722-3be27ffb30bd", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-96064402-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.145", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "55742b1d4fbb401ea61f990ecaa2def6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a8b99a46-3e7f-4ef1-9e45-58e6cd17f210", "external-id": "nsx-vlan-transportzone-704", "segmentation_id": 704, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap95763a65-63", "ovs_interfaceid": "95763a65-6383-409f-a9d6-1f37aaf44d34", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 986.658495] env[61957]: DEBUG oslo_vmware.rw_handles [None req-b10a76eb-ac59-4162-9083-648b4df991d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 986.658495] env[61957]: value = "vm-274582" [ 986.658495] env[61957]: _type = "VirtualMachine" [ 986.658495] env[61957]: }. {{(pid=61957) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 986.659086] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-d60ff898-5e43-4eaa-9381-cbbd737f4947 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.668753] env[61957]: DEBUG oslo_vmware.rw_handles [None req-b10a76eb-ac59-4162-9083-648b4df991d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Lease: (returnval){ [ 986.668753] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]525242e3-17f3-69d2-e303-2936a536c5fd" [ 986.668753] env[61957]: _type = "HttpNfcLease" [ 986.668753] env[61957]: } obtained for exporting VM: (result){ [ 986.668753] env[61957]: value = "vm-274582" [ 986.668753] env[61957]: _type = "VirtualMachine" [ 986.668753] env[61957]: }. {{(pid=61957) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 986.669198] env[61957]: DEBUG oslo_vmware.api [None req-b10a76eb-ac59-4162-9083-648b4df991d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Waiting for the lease: (returnval){ [ 986.669198] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]525242e3-17f3-69d2-e303-2936a536c5fd" [ 986.669198] env[61957]: _type = "HttpNfcLease" [ 986.669198] env[61957]: } to be ready. {{(pid=61957) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 986.678665] env[61957]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 986.678665] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]525242e3-17f3-69d2-e303-2936a536c5fd" [ 986.678665] env[61957]: _type = "HttpNfcLease" [ 986.678665] env[61957]: } is initializing. {{(pid=61957) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 986.889512] env[61957]: DEBUG oslo_vmware.api [None req-731394e9-d0a3-46c3-8b02-41ef72d369f5 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Task: {'id': task-1277886, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.48748} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.889809] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-731394e9-d0a3-46c3-8b02-41ef72d369f5 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore1] f65458b2-0db6-4d98-8c82-b27b6945a3fd/f65458b2-0db6-4d98-8c82-b27b6945a3fd.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 986.890046] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-731394e9-d0a3-46c3-8b02-41ef72d369f5 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] [instance: f65458b2-0db6-4d98-8c82-b27b6945a3fd] Extending root virtual disk to 1048576 {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 986.890400] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-44ae6961-adba-407b-9399-5d1f8eac31cd {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.898735] env[61957]: DEBUG oslo_vmware.api [None req-731394e9-d0a3-46c3-8b02-41ef72d369f5 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Waiting for the task: (returnval){ [ 986.898735] env[61957]: value = "task-1277888" [ 986.898735] env[61957]: _type = "Task" [ 986.898735] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.909164] env[61957]: DEBUG oslo_vmware.api [None req-731394e9-d0a3-46c3-8b02-41ef72d369f5 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Task: {'id': task-1277888, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.980706] env[61957]: DEBUG oslo_concurrency.lockutils [None req-74e1032e-bbb9-4915-a886-4f8885996564 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 987.078100] env[61957]: DEBUG nova.compute.utils [None req-6d13ccf5-9ccd-48d0-a35e-c6df97be9c15 tempest-ServerAddressesTestJSON-2054963291 tempest-ServerAddressesTestJSON-2054963291-project-member] Using /dev/sd instead of None {{(pid=61957) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 987.081763] env[61957]: DEBUG nova.compute.manager [None req-6d13ccf5-9ccd-48d0-a35e-c6df97be9c15 tempest-ServerAddressesTestJSON-2054963291 tempest-ServerAddressesTestJSON-2054963291-project-member] [instance: fa4c150e-7a85-41a4-8890-9f2deb3b34b1] Allocating IP information in the background. {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 987.081763] env[61957]: DEBUG nova.network.neutron [None req-6d13ccf5-9ccd-48d0-a35e-c6df97be9c15 tempest-ServerAddressesTestJSON-2054963291 tempest-ServerAddressesTestJSON-2054963291-project-member] [instance: fa4c150e-7a85-41a4-8890-9f2deb3b34b1] allocate_for_instance() {{(pid=61957) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 987.121122] env[61957]: DEBUG nova.policy [None req-6d13ccf5-9ccd-48d0-a35e-c6df97be9c15 tempest-ServerAddressesTestJSON-2054963291 tempest-ServerAddressesTestJSON-2054963291-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b1e2054e12f546af93b8c8b9509a8e1a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd5dffaf886624e1381979f2061556d2a', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61957) authorize /opt/stack/nova/nova/policy.py:203}} [ 987.127873] env[61957]: DEBUG oslo_concurrency.lockutils [None req-8b32736b-e3a6-485c-a95e-afec23b58d86 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Releasing lock "refresh_cache-736eabcf-70f8-4e1e-9830-270d9971fc9a" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 987.131027] env[61957]: DEBUG oslo_concurrency.lockutils [req-94058206-e54b-4216-9144-249674e2d939 req-052a2dcd-259c-466e-94d3-84bf68fe7614 service nova] Acquired lock "refresh_cache-736eabcf-70f8-4e1e-9830-270d9971fc9a" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 987.131027] env[61957]: DEBUG nova.network.neutron [req-94058206-e54b-4216-9144-249674e2d939 req-052a2dcd-259c-466e-94d3-84bf68fe7614 service nova] [instance: 736eabcf-70f8-4e1e-9830-270d9971fc9a] Refreshing network info cache for port 95763a65-6383-409f-a9d6-1f37aaf44d34 {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 987.150611] env[61957]: DEBUG nova.virt.hardware [None req-8b32736b-e3a6-485c-a95e-afec23b58d86 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T17:22:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='b6c9ccc51799fbdc3c7978b94ad879d7',container_format='bare',created_at=2024-10-12T17:30:16Z,direct_url=,disk_format='vmdk',id=0018d8e9-1d6e-4ca4-a224-31fc9e9f9d2e,min_disk=1,min_ram=0,name='tempest-AttachVolumeShelveTestJSON-server-661360609-shelved',owner='55742b1d4fbb401ea61f990ecaa2def6',properties=ImageMetaProps,protected=,size=31668736,status='active',tags=,updated_at=2024-10-12T17:30:32Z,virtual_size=,visibility=), allow threads: False {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 987.150873] env[61957]: DEBUG nova.virt.hardware [None req-8b32736b-e3a6-485c-a95e-afec23b58d86 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Flavor limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 987.151052] env[61957]: DEBUG nova.virt.hardware [None req-8b32736b-e3a6-485c-a95e-afec23b58d86 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Image limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 987.151248] env[61957]: DEBUG nova.virt.hardware [None req-8b32736b-e3a6-485c-a95e-afec23b58d86 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Flavor pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 987.151400] env[61957]: DEBUG nova.virt.hardware [None req-8b32736b-e3a6-485c-a95e-afec23b58d86 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Image pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 987.151548] env[61957]: DEBUG nova.virt.hardware [None req-8b32736b-e3a6-485c-a95e-afec23b58d86 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 987.151831] env[61957]: DEBUG nova.virt.hardware [None req-8b32736b-e3a6-485c-a95e-afec23b58d86 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 987.152007] env[61957]: DEBUG nova.virt.hardware [None req-8b32736b-e3a6-485c-a95e-afec23b58d86 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 987.152191] env[61957]: DEBUG nova.virt.hardware [None req-8b32736b-e3a6-485c-a95e-afec23b58d86 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Got 1 possible topologies {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 987.152361] env[61957]: DEBUG nova.virt.hardware [None req-8b32736b-e3a6-485c-a95e-afec23b58d86 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 987.152534] env[61957]: DEBUG nova.virt.hardware [None req-8b32736b-e3a6-485c-a95e-afec23b58d86 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 987.153434] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4acbf450-d77c-49d4-b6b8-d53c7bc6321d {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.162440] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee38de15-b5bd-400f-b03d-5eb7aa0ad6da {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.177032] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-8b32736b-e3a6-485c-a95e-afec23b58d86 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] [instance: 736eabcf-70f8-4e1e-9830-270d9971fc9a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:64:07:83', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a8b99a46-3e7f-4ef1-9e45-58e6cd17f210', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '95763a65-6383-409f-a9d6-1f37aaf44d34', 'vif_model': 'vmxnet3'}] {{(pid=61957) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 987.184201] env[61957]: DEBUG oslo.service.loopingcall [None req-8b32736b-e3a6-485c-a95e-afec23b58d86 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 987.190201] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 736eabcf-70f8-4e1e-9830-270d9971fc9a] Creating VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 987.190490] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0160a847-d2ea-45b6-be11-9218fa26cfdf {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.212600] env[61957]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 987.212600] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]525242e3-17f3-69d2-e303-2936a536c5fd" [ 987.212600] env[61957]: _type = "HttpNfcLease" [ 987.212600] env[61957]: } is ready. {{(pid=61957) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 987.212600] env[61957]: DEBUG oslo_vmware.rw_handles [None req-b10a76eb-ac59-4162-9083-648b4df991d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 987.212600] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]525242e3-17f3-69d2-e303-2936a536c5fd" [ 987.212600] env[61957]: _type = "HttpNfcLease" [ 987.212600] env[61957]: }. {{(pid=61957) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 987.213120] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b97d5b44-aed2-47ac-bc45-189fe5ec0317 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.216436] env[61957]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 987.216436] env[61957]: value = "task-1277889" [ 987.216436] env[61957]: _type = "Task" [ 987.216436] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.223203] env[61957]: DEBUG oslo_vmware.rw_handles [None req-b10a76eb-ac59-4162-9083-648b4df991d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52ae571d-e901-2583-cb41-ba12bc62eac2/disk-0.vmdk from lease info. {{(pid=61957) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 987.223418] env[61957]: DEBUG oslo_vmware.rw_handles [None req-b10a76eb-ac59-4162-9083-648b4df991d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52ae571d-e901-2583-cb41-ba12bc62eac2/disk-0.vmdk for reading. {{(pid=61957) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 987.285945] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277889, 'name': CreateVM_Task} progress is 5%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.357049] env[61957]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-57f2cd8d-6a8f-4035-9699-5956043318b9 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.406009] env[61957]: DEBUG nova.network.neutron [None req-6d13ccf5-9ccd-48d0-a35e-c6df97be9c15 tempest-ServerAddressesTestJSON-2054963291 tempest-ServerAddressesTestJSON-2054963291-project-member] [instance: fa4c150e-7a85-41a4-8890-9f2deb3b34b1] Successfully created port: 4b2ccba5-48f1-44f1-8746-3792d3695e58 {{(pid=61957) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 987.411470] env[61957]: DEBUG oslo_vmware.api [None req-731394e9-d0a3-46c3-8b02-41ef72d369f5 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Task: {'id': task-1277888, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.082693} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.411814] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-731394e9-d0a3-46c3-8b02-41ef72d369f5 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] [instance: f65458b2-0db6-4d98-8c82-b27b6945a3fd] Extended root virtual disk {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 987.412678] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5456fe99-8fd4-4e8c-a57e-4ee53dbb930c {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.436671] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-731394e9-d0a3-46c3-8b02-41ef72d369f5 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] [instance: f65458b2-0db6-4d98-8c82-b27b6945a3fd] Reconfiguring VM instance instance-0000005b to attach disk [datastore1] f65458b2-0db6-4d98-8c82-b27b6945a3fd/f65458b2-0db6-4d98-8c82-b27b6945a3fd.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 987.436949] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3491feb6-5af0-4b6b-8488-a23be2cc1020 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.459225] env[61957]: DEBUG oslo_vmware.api [None req-731394e9-d0a3-46c3-8b02-41ef72d369f5 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Waiting for the task: (returnval){ [ 987.459225] env[61957]: value = "task-1277890" [ 987.459225] env[61957]: _type = "Task" [ 987.459225] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.468393] env[61957]: DEBUG oslo_vmware.api [None req-731394e9-d0a3-46c3-8b02-41ef72d369f5 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Task: {'id': task-1277890, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.582935] env[61957]: DEBUG nova.compute.manager [None req-6d13ccf5-9ccd-48d0-a35e-c6df97be9c15 tempest-ServerAddressesTestJSON-2054963291 tempest-ServerAddressesTestJSON-2054963291-project-member] [instance: fa4c150e-7a85-41a4-8890-9f2deb3b34b1] Start building block device mappings for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 987.730561] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277889, 'name': CreateVM_Task, 'duration_secs': 0.423579} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.730561] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 736eabcf-70f8-4e1e-9830-270d9971fc9a] Created VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 987.731338] env[61957]: DEBUG oslo_concurrency.lockutils [None req-8b32736b-e3a6-485c-a95e-afec23b58d86 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0018d8e9-1d6e-4ca4-a224-31fc9e9f9d2e" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 987.731642] env[61957]: DEBUG oslo_concurrency.lockutils [None req-8b32736b-e3a6-485c-a95e-afec23b58d86 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0018d8e9-1d6e-4ca4-a224-31fc9e9f9d2e" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 987.732137] env[61957]: DEBUG oslo_concurrency.lockutils [None req-8b32736b-e3a6-485c-a95e-afec23b58d86 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0018d8e9-1d6e-4ca4-a224-31fc9e9f9d2e" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 987.732427] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3c7c18bd-4f1f-4515-ab8b-f468c815b587 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.741853] env[61957]: DEBUG oslo_vmware.api [None req-8b32736b-e3a6-485c-a95e-afec23b58d86 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Waiting for the task: (returnval){ [ 987.741853] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]525672df-3337-ae93-df03-b1e08e0714af" [ 987.741853] env[61957]: _type = "Task" [ 987.741853] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.759928] env[61957]: DEBUG oslo_vmware.api [None req-8b32736b-e3a6-485c-a95e-afec23b58d86 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]525672df-3337-ae93-df03-b1e08e0714af, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.882312] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d0bf04f-2f8d-4f53-b6ab-368543b80ec9 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.890207] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48a85dc7-c78e-4723-b4a6-6f48df1d3ef9 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.927893] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-064f6bf6-4cef-4802-9040-b09381fc30ea {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.942824] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-046b9573-527c-46cc-88fa-4eb172cf86e2 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.959525] env[61957]: DEBUG nova.compute.provider_tree [None req-bc1354f0-0ce9-464c-89e1-e1e216546ebe tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 987.974024] env[61957]: DEBUG oslo_vmware.api [None req-731394e9-d0a3-46c3-8b02-41ef72d369f5 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Task: {'id': task-1277890, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.977786] env[61957]: DEBUG nova.network.neutron [req-94058206-e54b-4216-9144-249674e2d939 req-052a2dcd-259c-466e-94d3-84bf68fe7614 service nova] [instance: 736eabcf-70f8-4e1e-9830-270d9971fc9a] Updated VIF entry in instance network info cache for port 95763a65-6383-409f-a9d6-1f37aaf44d34. {{(pid=61957) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 987.978219] env[61957]: DEBUG nova.network.neutron [req-94058206-e54b-4216-9144-249674e2d939 req-052a2dcd-259c-466e-94d3-84bf68fe7614 service nova] [instance: 736eabcf-70f8-4e1e-9830-270d9971fc9a] Updating instance_info_cache with network_info: [{"id": "95763a65-6383-409f-a9d6-1f37aaf44d34", "address": "fa:16:3e:64:07:83", "network": {"id": "b303ad01-ef9b-4119-9722-3be27ffb30bd", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-96064402-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.145", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "55742b1d4fbb401ea61f990ecaa2def6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a8b99a46-3e7f-4ef1-9e45-58e6cd17f210", "external-id": "nsx-vlan-transportzone-704", "segmentation_id": 704, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap95763a65-63", "ovs_interfaceid": "95763a65-6383-409f-a9d6-1f37aaf44d34", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 988.256019] env[61957]: DEBUG oslo_concurrency.lockutils [None req-8b32736b-e3a6-485c-a95e-afec23b58d86 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0018d8e9-1d6e-4ca4-a224-31fc9e9f9d2e" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 988.256576] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-8b32736b-e3a6-485c-a95e-afec23b58d86 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] [instance: 736eabcf-70f8-4e1e-9830-270d9971fc9a] Processing image 0018d8e9-1d6e-4ca4-a224-31fc9e9f9d2e {{(pid=61957) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 988.256949] env[61957]: DEBUG oslo_concurrency.lockutils [None req-8b32736b-e3a6-485c-a95e-afec23b58d86 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0018d8e9-1d6e-4ca4-a224-31fc9e9f9d2e/0018d8e9-1d6e-4ca4-a224-31fc9e9f9d2e.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 988.257206] env[61957]: DEBUG oslo_concurrency.lockutils [None req-8b32736b-e3a6-485c-a95e-afec23b58d86 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0018d8e9-1d6e-4ca4-a224-31fc9e9f9d2e/0018d8e9-1d6e-4ca4-a224-31fc9e9f9d2e.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 988.257481] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-8b32736b-e3a6-485c-a95e-afec23b58d86 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 988.257847] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-aac46e16-824a-4cc8-99bb-d0de014461a2 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.268522] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-8b32736b-e3a6-485c-a95e-afec23b58d86 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 988.268737] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-8b32736b-e3a6-485c-a95e-afec23b58d86 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61957) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 988.269557] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e67c97e6-97a7-45a1-b174-16e00c030411 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.276207] env[61957]: DEBUG oslo_vmware.api [None req-8b32736b-e3a6-485c-a95e-afec23b58d86 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Waiting for the task: (returnval){ [ 988.276207] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52fc9148-4919-3823-cf46-b363e18f1cb6" [ 988.276207] env[61957]: _type = "Task" [ 988.276207] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.289357] env[61957]: DEBUG oslo_vmware.api [None req-8b32736b-e3a6-485c-a95e-afec23b58d86 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52fc9148-4919-3823-cf46-b363e18f1cb6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.465734] env[61957]: DEBUG nova.scheduler.client.report [None req-bc1354f0-0ce9-464c-89e1-e1e216546ebe tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 988.475237] env[61957]: DEBUG oslo_vmware.api [None req-731394e9-d0a3-46c3-8b02-41ef72d369f5 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Task: {'id': task-1277890, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.481077] env[61957]: DEBUG oslo_concurrency.lockutils [req-94058206-e54b-4216-9144-249674e2d939 req-052a2dcd-259c-466e-94d3-84bf68fe7614 service nova] Releasing lock "refresh_cache-736eabcf-70f8-4e1e-9830-270d9971fc9a" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 988.597543] env[61957]: DEBUG nova.compute.manager [None req-6d13ccf5-9ccd-48d0-a35e-c6df97be9c15 tempest-ServerAddressesTestJSON-2054963291 tempest-ServerAddressesTestJSON-2054963291-project-member] [instance: fa4c150e-7a85-41a4-8890-9f2deb3b34b1] Start spawning the instance on the hypervisor. {{(pid=61957) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 988.628055] env[61957]: DEBUG nova.virt.hardware [None req-6d13ccf5-9ccd-48d0-a35e-c6df97be9c15 tempest-ServerAddressesTestJSON-2054963291 tempest-ServerAddressesTestJSON-2054963291-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T17:22:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T17:21:50Z,direct_url=,disk_format='vmdk',id=11c76a2c-f705-470a-ba9d-4657858bab38,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='81d3096d0f094373913fc3dc8f5c3c6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T17:21:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 988.628507] env[61957]: DEBUG nova.virt.hardware [None req-6d13ccf5-9ccd-48d0-a35e-c6df97be9c15 tempest-ServerAddressesTestJSON-2054963291 tempest-ServerAddressesTestJSON-2054963291-project-member] Flavor limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 988.628973] env[61957]: DEBUG nova.virt.hardware [None req-6d13ccf5-9ccd-48d0-a35e-c6df97be9c15 tempest-ServerAddressesTestJSON-2054963291 tempest-ServerAddressesTestJSON-2054963291-project-member] Image limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 988.629530] env[61957]: DEBUG nova.virt.hardware [None req-6d13ccf5-9ccd-48d0-a35e-c6df97be9c15 tempest-ServerAddressesTestJSON-2054963291 tempest-ServerAddressesTestJSON-2054963291-project-member] Flavor pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 988.629875] env[61957]: DEBUG nova.virt.hardware [None req-6d13ccf5-9ccd-48d0-a35e-c6df97be9c15 tempest-ServerAddressesTestJSON-2054963291 tempest-ServerAddressesTestJSON-2054963291-project-member] Image pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 988.630175] env[61957]: DEBUG nova.virt.hardware [None req-6d13ccf5-9ccd-48d0-a35e-c6df97be9c15 tempest-ServerAddressesTestJSON-2054963291 tempest-ServerAddressesTestJSON-2054963291-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 988.630547] env[61957]: DEBUG nova.virt.hardware [None req-6d13ccf5-9ccd-48d0-a35e-c6df97be9c15 tempest-ServerAddressesTestJSON-2054963291 tempest-ServerAddressesTestJSON-2054963291-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 988.630844] env[61957]: DEBUG nova.virt.hardware [None req-6d13ccf5-9ccd-48d0-a35e-c6df97be9c15 tempest-ServerAddressesTestJSON-2054963291 tempest-ServerAddressesTestJSON-2054963291-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 988.633761] env[61957]: DEBUG nova.virt.hardware [None req-6d13ccf5-9ccd-48d0-a35e-c6df97be9c15 tempest-ServerAddressesTestJSON-2054963291 tempest-ServerAddressesTestJSON-2054963291-project-member] Got 1 possible topologies {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 988.633761] env[61957]: DEBUG nova.virt.hardware [None req-6d13ccf5-9ccd-48d0-a35e-c6df97be9c15 tempest-ServerAddressesTestJSON-2054963291 tempest-ServerAddressesTestJSON-2054963291-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 988.633761] env[61957]: DEBUG nova.virt.hardware [None req-6d13ccf5-9ccd-48d0-a35e-c6df97be9c15 tempest-ServerAddressesTestJSON-2054963291 tempest-ServerAddressesTestJSON-2054963291-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 988.633761] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66f91c2b-62bd-4ad8-a745-07ab4d1c6af7 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.644613] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ac0998d-6db5-4f3a-b7b8-69e06760f8a3 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.788196] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-8b32736b-e3a6-485c-a95e-afec23b58d86 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] [instance: 736eabcf-70f8-4e1e-9830-270d9971fc9a] Preparing fetch location {{(pid=61957) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 988.788549] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-8b32736b-e3a6-485c-a95e-afec23b58d86 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] [instance: 736eabcf-70f8-4e1e-9830-270d9971fc9a] Fetch image to [datastore2] OSTACK_IMG_e717072e-0b5a-4ae5-acc6-5d7a40856a3f/OSTACK_IMG_e717072e-0b5a-4ae5-acc6-5d7a40856a3f.vmdk {{(pid=61957) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 988.789335] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-8b32736b-e3a6-485c-a95e-afec23b58d86 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] [instance: 736eabcf-70f8-4e1e-9830-270d9971fc9a] Downloading stream optimized image 0018d8e9-1d6e-4ca4-a224-31fc9e9f9d2e to [datastore2] OSTACK_IMG_e717072e-0b5a-4ae5-acc6-5d7a40856a3f/OSTACK_IMG_e717072e-0b5a-4ae5-acc6-5d7a40856a3f.vmdk on the data store datastore2 as vApp {{(pid=61957) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 988.789593] env[61957]: DEBUG nova.virt.vmwareapi.images [None req-8b32736b-e3a6-485c-a95e-afec23b58d86 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] [instance: 736eabcf-70f8-4e1e-9830-270d9971fc9a] Downloading image file data 0018d8e9-1d6e-4ca4-a224-31fc9e9f9d2e to the ESX as VM named 'OSTACK_IMG_e717072e-0b5a-4ae5-acc6-5d7a40856a3f' {{(pid=61957) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 988.861822] env[61957]: DEBUG oslo_vmware.rw_handles [None req-8b32736b-e3a6-485c-a95e-afec23b58d86 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 988.861822] env[61957]: value = "resgroup-9" [ 988.861822] env[61957]: _type = "ResourcePool" [ 988.861822] env[61957]: }. {{(pid=61957) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 988.862185] env[61957]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-b12c2dc9-c756-4c3b-ae2e-2efafebd21e0 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.884860] env[61957]: DEBUG oslo_vmware.rw_handles [None req-8b32736b-e3a6-485c-a95e-afec23b58d86 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Lease: (returnval){ [ 988.884860] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]528e4d36-5935-b0d9-7210-2a44dbf2d43e" [ 988.884860] env[61957]: _type = "HttpNfcLease" [ 988.884860] env[61957]: } obtained for vApp import into resource pool (val){ [ 988.884860] env[61957]: value = "resgroup-9" [ 988.884860] env[61957]: _type = "ResourcePool" [ 988.884860] env[61957]: }. {{(pid=61957) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 988.885162] env[61957]: DEBUG oslo_vmware.api [None req-8b32736b-e3a6-485c-a95e-afec23b58d86 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Waiting for the lease: (returnval){ [ 988.885162] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]528e4d36-5935-b0d9-7210-2a44dbf2d43e" [ 988.885162] env[61957]: _type = "HttpNfcLease" [ 988.885162] env[61957]: } to be ready. {{(pid=61957) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 988.892396] env[61957]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 988.892396] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]528e4d36-5935-b0d9-7210-2a44dbf2d43e" [ 988.892396] env[61957]: _type = "HttpNfcLease" [ 988.892396] env[61957]: } is initializing. {{(pid=61957) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 988.971354] env[61957]: DEBUG oslo_concurrency.lockutils [None req-bc1354f0-0ce9-464c-89e1-e1e216546ebe tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.399s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 988.972317] env[61957]: DEBUG nova.compute.manager [None req-bc1354f0-0ce9-464c-89e1-e1e216546ebe tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: 6509707e-afc6-4c4a-88f9-a471865d510c] Start building networks asynchronously for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 988.983162] env[61957]: DEBUG oslo_concurrency.lockutils [None req-2f034be2-1f7c-41d0-91ad-8c549dbb7f8a tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.203s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 988.983751] env[61957]: DEBUG nova.objects.instance [None req-2f034be2-1f7c-41d0-91ad-8c549dbb7f8a tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Lazy-loading 'resources' on Instance uuid d6c5c70d-86de-4dea-8b9e-76f321947a35 {{(pid=61957) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 988.986149] env[61957]: DEBUG oslo_vmware.api [None req-731394e9-d0a3-46c3-8b02-41ef72d369f5 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Task: {'id': task-1277890, 'name': ReconfigVM_Task, 'duration_secs': 1.29606} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.986548] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-731394e9-d0a3-46c3-8b02-41ef72d369f5 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] [instance: f65458b2-0db6-4d98-8c82-b27b6945a3fd] Reconfigured VM instance instance-0000005b to attach disk [datastore1] f65458b2-0db6-4d98-8c82-b27b6945a3fd/f65458b2-0db6-4d98-8c82-b27b6945a3fd.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 988.988118] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c3367b3b-dc28-472f-bdab-b018438c0a7c {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.993205] env[61957]: DEBUG nova.compute.manager [req-9e0f9ff7-32c6-43d1-b9bb-89ffad31acd0 req-25f6e7ec-dcc2-483b-987f-4f32f4126c40 service nova] [instance: fa4c150e-7a85-41a4-8890-9f2deb3b34b1] Received event network-vif-plugged-4b2ccba5-48f1-44f1-8746-3792d3695e58 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 988.993483] env[61957]: DEBUG oslo_concurrency.lockutils [req-9e0f9ff7-32c6-43d1-b9bb-89ffad31acd0 req-25f6e7ec-dcc2-483b-987f-4f32f4126c40 service nova] Acquiring lock "fa4c150e-7a85-41a4-8890-9f2deb3b34b1-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 988.993798] env[61957]: DEBUG oslo_concurrency.lockutils [req-9e0f9ff7-32c6-43d1-b9bb-89ffad31acd0 req-25f6e7ec-dcc2-483b-987f-4f32f4126c40 service nova] Lock "fa4c150e-7a85-41a4-8890-9f2deb3b34b1-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 988.994052] env[61957]: DEBUG oslo_concurrency.lockutils [req-9e0f9ff7-32c6-43d1-b9bb-89ffad31acd0 req-25f6e7ec-dcc2-483b-987f-4f32f4126c40 service nova] Lock "fa4c150e-7a85-41a4-8890-9f2deb3b34b1-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 988.994316] env[61957]: DEBUG nova.compute.manager [req-9e0f9ff7-32c6-43d1-b9bb-89ffad31acd0 req-25f6e7ec-dcc2-483b-987f-4f32f4126c40 service nova] [instance: fa4c150e-7a85-41a4-8890-9f2deb3b34b1] No waiting events found dispatching network-vif-plugged-4b2ccba5-48f1-44f1-8746-3792d3695e58 {{(pid=61957) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 988.994555] env[61957]: WARNING nova.compute.manager [req-9e0f9ff7-32c6-43d1-b9bb-89ffad31acd0 req-25f6e7ec-dcc2-483b-987f-4f32f4126c40 service nova] [instance: fa4c150e-7a85-41a4-8890-9f2deb3b34b1] Received unexpected event network-vif-plugged-4b2ccba5-48f1-44f1-8746-3792d3695e58 for instance with vm_state building and task_state spawning. [ 989.003018] env[61957]: DEBUG oslo_vmware.api [None req-731394e9-d0a3-46c3-8b02-41ef72d369f5 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Waiting for the task: (returnval){ [ 989.003018] env[61957]: value = "task-1277892" [ 989.003018] env[61957]: _type = "Task" [ 989.003018] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.019023] env[61957]: DEBUG oslo_vmware.api [None req-731394e9-d0a3-46c3-8b02-41ef72d369f5 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Task: {'id': task-1277892, 'name': Rename_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.110121] env[61957]: DEBUG nova.network.neutron [None req-6d13ccf5-9ccd-48d0-a35e-c6df97be9c15 tempest-ServerAddressesTestJSON-2054963291 tempest-ServerAddressesTestJSON-2054963291-project-member] [instance: fa4c150e-7a85-41a4-8890-9f2deb3b34b1] Successfully updated port: 4b2ccba5-48f1-44f1-8746-3792d3695e58 {{(pid=61957) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 989.394842] env[61957]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 989.394842] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]528e4d36-5935-b0d9-7210-2a44dbf2d43e" [ 989.394842] env[61957]: _type = "HttpNfcLease" [ 989.394842] env[61957]: } is initializing. {{(pid=61957) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 989.481578] env[61957]: DEBUG nova.compute.utils [None req-bc1354f0-0ce9-464c-89e1-e1e216546ebe tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Using /dev/sd instead of None {{(pid=61957) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 989.483339] env[61957]: DEBUG nova.compute.manager [None req-bc1354f0-0ce9-464c-89e1-e1e216546ebe tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: 6509707e-afc6-4c4a-88f9-a471865d510c] Allocating IP information in the background. {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 989.483524] env[61957]: DEBUG nova.network.neutron [None req-bc1354f0-0ce9-464c-89e1-e1e216546ebe tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: 6509707e-afc6-4c4a-88f9-a471865d510c] allocate_for_instance() {{(pid=61957) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 989.515578] env[61957]: DEBUG oslo_vmware.api [None req-731394e9-d0a3-46c3-8b02-41ef72d369f5 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Task: {'id': task-1277892, 'name': Rename_Task, 'duration_secs': 0.168661} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.518524] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-731394e9-d0a3-46c3-8b02-41ef72d369f5 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] [instance: f65458b2-0db6-4d98-8c82-b27b6945a3fd] Powering on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 989.519363] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-60f2dc02-9514-4329-8692-224670ced440 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.527477] env[61957]: DEBUG oslo_vmware.api [None req-731394e9-d0a3-46c3-8b02-41ef72d369f5 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Waiting for the task: (returnval){ [ 989.527477] env[61957]: value = "task-1277893" [ 989.527477] env[61957]: _type = "Task" [ 989.527477] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.539439] env[61957]: DEBUG nova.policy [None req-bc1354f0-0ce9-464c-89e1-e1e216546ebe tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3136bf0d1db3411fb1ff1ebe9c6045f6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1dbdc699741a48af93c52068d88a5357', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61957) authorize /opt/stack/nova/nova/policy.py:203}} [ 989.540892] env[61957]: DEBUG oslo_vmware.api [None req-731394e9-d0a3-46c3-8b02-41ef72d369f5 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Task: {'id': task-1277893, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.613165] env[61957]: DEBUG oslo_concurrency.lockutils [None req-6d13ccf5-9ccd-48d0-a35e-c6df97be9c15 tempest-ServerAddressesTestJSON-2054963291 tempest-ServerAddressesTestJSON-2054963291-project-member] Acquiring lock "refresh_cache-fa4c150e-7a85-41a4-8890-9f2deb3b34b1" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 989.613327] env[61957]: DEBUG oslo_concurrency.lockutils [None req-6d13ccf5-9ccd-48d0-a35e-c6df97be9c15 tempest-ServerAddressesTestJSON-2054963291 tempest-ServerAddressesTestJSON-2054963291-project-member] Acquired lock "refresh_cache-fa4c150e-7a85-41a4-8890-9f2deb3b34b1" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 989.613483] env[61957]: DEBUG nova.network.neutron [None req-6d13ccf5-9ccd-48d0-a35e-c6df97be9c15 tempest-ServerAddressesTestJSON-2054963291 tempest-ServerAddressesTestJSON-2054963291-project-member] [instance: fa4c150e-7a85-41a4-8890-9f2deb3b34b1] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 989.706649] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34c6bf7a-5974-4e97-a26f-61aadb1e765e {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.715179] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f68de44-3a37-4037-b18b-c05377d158a9 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.748574] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41ab43f4-0319-4358-8904-defe9607d4df {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.757452] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bfd7281-6308-4480-be10-a3ea4305968e {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.783634] env[61957]: DEBUG nova.compute.provider_tree [None req-2f034be2-1f7c-41d0-91ad-8c549dbb7f8a tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 989.849715] env[61957]: DEBUG nova.network.neutron [None req-bc1354f0-0ce9-464c-89e1-e1e216546ebe tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: 6509707e-afc6-4c4a-88f9-a471865d510c] Successfully created port: 7df2736b-10d0-4f9d-83aa-e2bbdf1ef45b {{(pid=61957) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 989.895034] env[61957]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 989.895034] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]528e4d36-5935-b0d9-7210-2a44dbf2d43e" [ 989.895034] env[61957]: _type = "HttpNfcLease" [ 989.895034] env[61957]: } is ready. {{(pid=61957) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 989.895034] env[61957]: DEBUG oslo_vmware.rw_handles [None req-8b32736b-e3a6-485c-a95e-afec23b58d86 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 989.895034] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]528e4d36-5935-b0d9-7210-2a44dbf2d43e" [ 989.895034] env[61957]: _type = "HttpNfcLease" [ 989.895034] env[61957]: }. {{(pid=61957) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 989.898812] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50c0bb6a-2809-4064-9aa1-2ed29cfc7337 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.904595] env[61957]: DEBUG oslo_vmware.rw_handles [None req-8b32736b-e3a6-485c-a95e-afec23b58d86 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5220add5-8f41-4b30-d79b-e7ac2271021e/disk-0.vmdk from lease info. {{(pid=61957) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 989.904595] env[61957]: DEBUG oslo_vmware.rw_handles [None req-8b32736b-e3a6-485c-a95e-afec23b58d86 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Creating HTTP connection to write to file with size = 31668736 and URL = https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5220add5-8f41-4b30-d79b-e7ac2271021e/disk-0.vmdk. {{(pid=61957) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 989.970252] env[61957]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-6a7e1af5-cef2-420f-a26f-3fe8a5b8ad54 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.986829] env[61957]: DEBUG nova.compute.manager [None req-bc1354f0-0ce9-464c-89e1-e1e216546ebe tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: 6509707e-afc6-4c4a-88f9-a471865d510c] Start building block device mappings for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 990.039262] env[61957]: DEBUG oslo_vmware.api [None req-731394e9-d0a3-46c3-8b02-41ef72d369f5 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Task: {'id': task-1277893, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.151894] env[61957]: DEBUG nova.network.neutron [None req-6d13ccf5-9ccd-48d0-a35e-c6df97be9c15 tempest-ServerAddressesTestJSON-2054963291 tempest-ServerAddressesTestJSON-2054963291-project-member] [instance: fa4c150e-7a85-41a4-8890-9f2deb3b34b1] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 990.275477] env[61957]: DEBUG nova.scheduler.client.report [None req-2f034be2-1f7c-41d0-91ad-8c549dbb7f8a tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 990.329768] env[61957]: DEBUG nova.network.neutron [None req-6d13ccf5-9ccd-48d0-a35e-c6df97be9c15 tempest-ServerAddressesTestJSON-2054963291 tempest-ServerAddressesTestJSON-2054963291-project-member] [instance: fa4c150e-7a85-41a4-8890-9f2deb3b34b1] Updating instance_info_cache with network_info: [{"id": "4b2ccba5-48f1-44f1-8746-3792d3695e58", "address": "fa:16:3e:83:a2:9e", "network": {"id": "e45c12da-f92f-4516-a616-e4478247693e", "bridge": "br-int", "label": "tempest-ServerAddressesTestJSON-96077499-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d5dffaf886624e1381979f2061556d2a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d0bdd11b-58af-4cc0-9d38-8322e1bb4e74", "external-id": "nsx-vlan-transportzone-398", "segmentation_id": 398, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4b2ccba5-48", "ovs_interfaceid": "4b2ccba5-48f1-44f1-8746-3792d3695e58", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 990.540751] env[61957]: DEBUG oslo_vmware.api [None req-731394e9-d0a3-46c3-8b02-41ef72d369f5 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Task: {'id': task-1277893, 'name': PowerOnVM_Task, 'duration_secs': 0.523111} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.541250] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-731394e9-d0a3-46c3-8b02-41ef72d369f5 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] [instance: f65458b2-0db6-4d98-8c82-b27b6945a3fd] Powered on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 990.541250] env[61957]: INFO nova.compute.manager [None req-731394e9-d0a3-46c3-8b02-41ef72d369f5 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] [instance: f65458b2-0db6-4d98-8c82-b27b6945a3fd] Took 8.14 seconds to spawn the instance on the hypervisor. [ 990.541456] env[61957]: DEBUG nova.compute.manager [None req-731394e9-d0a3-46c3-8b02-41ef72d369f5 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] [instance: f65458b2-0db6-4d98-8c82-b27b6945a3fd] Checking state {{(pid=61957) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 990.542325] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-773084aa-3bb0-40ee-94ed-d837d5f6b10f {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.786319] env[61957]: DEBUG oslo_concurrency.lockutils [None req-2f034be2-1f7c-41d0-91ad-8c549dbb7f8a tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.803s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 990.789846] env[61957]: DEBUG oslo_concurrency.lockutils [None req-c255bb00-cba6-445a-adb0-12124fe511af tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.218s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 990.789846] env[61957]: DEBUG oslo_concurrency.lockutils [None req-c255bb00-cba6-445a-adb0-12124fe511af tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 990.792718] env[61957]: DEBUG oslo_concurrency.lockutils [None req-74e1032e-bbb9-4915-a886-4f8885996564 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.811s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 990.792718] env[61957]: DEBUG nova.objects.instance [None req-74e1032e-bbb9-4915-a886-4f8885996564 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Lazy-loading 'resources' on Instance uuid 6f45c7c8-de07-4d21-bccc-e4afe3c8ad3d {{(pid=61957) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 990.813296] env[61957]: INFO nova.scheduler.client.report [None req-2f034be2-1f7c-41d0-91ad-8c549dbb7f8a tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Deleted allocations for instance d6c5c70d-86de-4dea-8b9e-76f321947a35 [ 990.815811] env[61957]: INFO nova.scheduler.client.report [None req-c255bb00-cba6-445a-adb0-12124fe511af tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Deleted allocations for instance dbeeb200-70b9-4cb4-b5a4-182389d21918 [ 990.839104] env[61957]: DEBUG oslo_concurrency.lockutils [None req-6d13ccf5-9ccd-48d0-a35e-c6df97be9c15 tempest-ServerAddressesTestJSON-2054963291 tempest-ServerAddressesTestJSON-2054963291-project-member] Releasing lock "refresh_cache-fa4c150e-7a85-41a4-8890-9f2deb3b34b1" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 990.839652] env[61957]: DEBUG nova.compute.manager [None req-6d13ccf5-9ccd-48d0-a35e-c6df97be9c15 tempest-ServerAddressesTestJSON-2054963291 tempest-ServerAddressesTestJSON-2054963291-project-member] [instance: fa4c150e-7a85-41a4-8890-9f2deb3b34b1] Instance network_info: |[{"id": "4b2ccba5-48f1-44f1-8746-3792d3695e58", "address": "fa:16:3e:83:a2:9e", "network": {"id": "e45c12da-f92f-4516-a616-e4478247693e", "bridge": "br-int", "label": "tempest-ServerAddressesTestJSON-96077499-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d5dffaf886624e1381979f2061556d2a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d0bdd11b-58af-4cc0-9d38-8322e1bb4e74", "external-id": "nsx-vlan-transportzone-398", "segmentation_id": 398, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4b2ccba5-48", "ovs_interfaceid": "4b2ccba5-48f1-44f1-8746-3792d3695e58", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 990.840606] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-6d13ccf5-9ccd-48d0-a35e-c6df97be9c15 tempest-ServerAddressesTestJSON-2054963291 tempest-ServerAddressesTestJSON-2054963291-project-member] [instance: fa4c150e-7a85-41a4-8890-9f2deb3b34b1] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:83:a2:9e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd0bdd11b-58af-4cc0-9d38-8322e1bb4e74', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4b2ccba5-48f1-44f1-8746-3792d3695e58', 'vif_model': 'vmxnet3'}] {{(pid=61957) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 990.848936] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d13ccf5-9ccd-48d0-a35e-c6df97be9c15 tempest-ServerAddressesTestJSON-2054963291 tempest-ServerAddressesTestJSON-2054963291-project-member] Creating folder: Project (d5dffaf886624e1381979f2061556d2a). Parent ref: group-v274445. {{(pid=61957) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 990.852115] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3f6a09cd-d9b0-4aa0-a534-51e65601366b {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.868112] env[61957]: INFO nova.virt.vmwareapi.vm_util [None req-6d13ccf5-9ccd-48d0-a35e-c6df97be9c15 tempest-ServerAddressesTestJSON-2054963291 tempest-ServerAddressesTestJSON-2054963291-project-member] Created folder: Project (d5dffaf886624e1381979f2061556d2a) in parent group-v274445. [ 990.868645] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d13ccf5-9ccd-48d0-a35e-c6df97be9c15 tempest-ServerAddressesTestJSON-2054963291 tempest-ServerAddressesTestJSON-2054963291-project-member] Creating folder: Instances. Parent ref: group-v274585. {{(pid=61957) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 990.871404] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-53b23288-b802-4bd5-b875-848d8b42cfab {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.885231] env[61957]: INFO nova.virt.vmwareapi.vm_util [None req-6d13ccf5-9ccd-48d0-a35e-c6df97be9c15 tempest-ServerAddressesTestJSON-2054963291 tempest-ServerAddressesTestJSON-2054963291-project-member] Created folder: Instances in parent group-v274585. [ 990.885861] env[61957]: DEBUG oslo.service.loopingcall [None req-6d13ccf5-9ccd-48d0-a35e-c6df97be9c15 tempest-ServerAddressesTestJSON-2054963291 tempest-ServerAddressesTestJSON-2054963291-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 990.886246] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fa4c150e-7a85-41a4-8890-9f2deb3b34b1] Creating VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 990.886607] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-01afb01a-b44c-4acd-b2d3-3959d6394542 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.908914] env[61957]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 990.908914] env[61957]: value = "task-1277896" [ 990.908914] env[61957]: _type = "Task" [ 990.908914] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.918791] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277896, 'name': CreateVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.999328] env[61957]: DEBUG nova.compute.manager [None req-bc1354f0-0ce9-464c-89e1-e1e216546ebe tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: 6509707e-afc6-4c4a-88f9-a471865d510c] Start spawning the instance on the hypervisor. {{(pid=61957) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 991.026922] env[61957]: DEBUG nova.compute.manager [req-3f469708-25dd-4783-84a5-996a112ec595 req-b430dae8-7755-478e-a693-3c310f9a9e10 service nova] [instance: fa4c150e-7a85-41a4-8890-9f2deb3b34b1] Received event network-changed-4b2ccba5-48f1-44f1-8746-3792d3695e58 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 991.027251] env[61957]: DEBUG nova.compute.manager [req-3f469708-25dd-4783-84a5-996a112ec595 req-b430dae8-7755-478e-a693-3c310f9a9e10 service nova] [instance: fa4c150e-7a85-41a4-8890-9f2deb3b34b1] Refreshing instance network info cache due to event network-changed-4b2ccba5-48f1-44f1-8746-3792d3695e58. {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 991.027619] env[61957]: DEBUG oslo_concurrency.lockutils [req-3f469708-25dd-4783-84a5-996a112ec595 req-b430dae8-7755-478e-a693-3c310f9a9e10 service nova] Acquiring lock "refresh_cache-fa4c150e-7a85-41a4-8890-9f2deb3b34b1" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 991.027767] env[61957]: DEBUG oslo_concurrency.lockutils [req-3f469708-25dd-4783-84a5-996a112ec595 req-b430dae8-7755-478e-a693-3c310f9a9e10 service nova] Acquired lock "refresh_cache-fa4c150e-7a85-41a4-8890-9f2deb3b34b1" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 991.027963] env[61957]: DEBUG nova.network.neutron [req-3f469708-25dd-4783-84a5-996a112ec595 req-b430dae8-7755-478e-a693-3c310f9a9e10 service nova] [instance: fa4c150e-7a85-41a4-8890-9f2deb3b34b1] Refreshing network info cache for port 4b2ccba5-48f1-44f1-8746-3792d3695e58 {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 991.040844] env[61957]: DEBUG nova.virt.hardware [None req-bc1354f0-0ce9-464c-89e1-e1e216546ebe tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T17:22:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T17:21:50Z,direct_url=,disk_format='vmdk',id=11c76a2c-f705-470a-ba9d-4657858bab38,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='81d3096d0f094373913fc3dc8f5c3c6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T17:21:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 991.041123] env[61957]: DEBUG nova.virt.hardware [None req-bc1354f0-0ce9-464c-89e1-e1e216546ebe tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Flavor limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 991.041328] env[61957]: DEBUG nova.virt.hardware [None req-bc1354f0-0ce9-464c-89e1-e1e216546ebe tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Image limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 991.041624] env[61957]: DEBUG nova.virt.hardware [None req-bc1354f0-0ce9-464c-89e1-e1e216546ebe tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Flavor pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 991.041825] env[61957]: DEBUG nova.virt.hardware [None req-bc1354f0-0ce9-464c-89e1-e1e216546ebe tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Image pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 991.041991] env[61957]: DEBUG nova.virt.hardware [None req-bc1354f0-0ce9-464c-89e1-e1e216546ebe tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 991.042328] env[61957]: DEBUG nova.virt.hardware [None req-bc1354f0-0ce9-464c-89e1-e1e216546ebe tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 991.042563] env[61957]: DEBUG nova.virt.hardware [None req-bc1354f0-0ce9-464c-89e1-e1e216546ebe tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 991.042759] env[61957]: DEBUG nova.virt.hardware [None req-bc1354f0-0ce9-464c-89e1-e1e216546ebe tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Got 1 possible topologies {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 991.042956] env[61957]: DEBUG nova.virt.hardware [None req-bc1354f0-0ce9-464c-89e1-e1e216546ebe tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 991.043152] env[61957]: DEBUG nova.virt.hardware [None req-bc1354f0-0ce9-464c-89e1-e1e216546ebe tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 991.044596] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-738c7aff-bb97-4226-940c-0f78af8396fc {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.059099] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41b8de23-993f-4069-87bf-f10f614244e4 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.074293] env[61957]: INFO nova.compute.manager [None req-731394e9-d0a3-46c3-8b02-41ef72d369f5 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] [instance: f65458b2-0db6-4d98-8c82-b27b6945a3fd] Took 25.92 seconds to build instance. [ 991.133449] env[61957]: DEBUG oslo_vmware.rw_handles [None req-8b32736b-e3a6-485c-a95e-afec23b58d86 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Completed reading data from the image iterator. {{(pid=61957) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 991.134861] env[61957]: DEBUG oslo_vmware.rw_handles [None req-8b32736b-e3a6-485c-a95e-afec23b58d86 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5220add5-8f41-4b30-d79b-e7ac2271021e/disk-0.vmdk. {{(pid=61957) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 991.134861] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b50a8247-c425-43af-ae21-d5da9d9f4ef7 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.144366] env[61957]: DEBUG oslo_vmware.rw_handles [None req-8b32736b-e3a6-485c-a95e-afec23b58d86 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5220add5-8f41-4b30-d79b-e7ac2271021e/disk-0.vmdk is in state: ready. {{(pid=61957) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 991.145168] env[61957]: DEBUG oslo_vmware.rw_handles [None req-8b32736b-e3a6-485c-a95e-afec23b58d86 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Releasing lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5220add5-8f41-4b30-d79b-e7ac2271021e/disk-0.vmdk. {{(pid=61957) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 991.145497] env[61957]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-a993edc0-bdc6-4d73-9d17-3a2bd3632da6 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.349641] env[61957]: DEBUG oslo_concurrency.lockutils [None req-2f034be2-1f7c-41d0-91ad-8c549dbb7f8a tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Lock "d6c5c70d-86de-4dea-8b9e-76f321947a35" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.197s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 991.350423] env[61957]: DEBUG oslo_concurrency.lockutils [None req-c255bb00-cba6-445a-adb0-12124fe511af tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Lock "dbeeb200-70b9-4cb4-b5a4-182389d21918" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.343s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 991.381085] env[61957]: DEBUG oslo_vmware.rw_handles [None req-8b32736b-e3a6-485c-a95e-afec23b58d86 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Closed VMDK write handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5220add5-8f41-4b30-d79b-e7ac2271021e/disk-0.vmdk. {{(pid=61957) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 991.381321] env[61957]: INFO nova.virt.vmwareapi.images [None req-8b32736b-e3a6-485c-a95e-afec23b58d86 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] [instance: 736eabcf-70f8-4e1e-9830-270d9971fc9a] Downloaded image file data 0018d8e9-1d6e-4ca4-a224-31fc9e9f9d2e [ 991.384021] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b401870-93d8-4428-9d95-3a7f43d69dc4 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.404211] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-988de167-33cc-4310-aec9-f698a4c9f3d2 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.420294] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277896, 'name': CreateVM_Task, 'duration_secs': 0.408823} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.421303] env[61957]: DEBUG nova.network.neutron [None req-bc1354f0-0ce9-464c-89e1-e1e216546ebe tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: 6509707e-afc6-4c4a-88f9-a471865d510c] Successfully updated port: 7df2736b-10d0-4f9d-83aa-e2bbdf1ef45b {{(pid=61957) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 991.422717] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fa4c150e-7a85-41a4-8890-9f2deb3b34b1] Created VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 991.423831] env[61957]: DEBUG oslo_concurrency.lockutils [None req-6d13ccf5-9ccd-48d0-a35e-c6df97be9c15 tempest-ServerAddressesTestJSON-2054963291 tempest-ServerAddressesTestJSON-2054963291-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 991.423831] env[61957]: DEBUG oslo_concurrency.lockutils [None req-6d13ccf5-9ccd-48d0-a35e-c6df97be9c15 tempest-ServerAddressesTestJSON-2054963291 tempest-ServerAddressesTestJSON-2054963291-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 991.424524] env[61957]: DEBUG oslo_concurrency.lockutils [None req-6d13ccf5-9ccd-48d0-a35e-c6df97be9c15 tempest-ServerAddressesTestJSON-2054963291 tempest-ServerAddressesTestJSON-2054963291-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 991.424524] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fc6f0989-2c8c-46ad-862c-9a523dddc994 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.431506] env[61957]: DEBUG oslo_vmware.api [None req-6d13ccf5-9ccd-48d0-a35e-c6df97be9c15 tempest-ServerAddressesTestJSON-2054963291 tempest-ServerAddressesTestJSON-2054963291-project-member] Waiting for the task: (returnval){ [ 991.431506] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52b3687d-e7d5-93af-de70-8e5b4d14c0ab" [ 991.431506] env[61957]: _type = "Task" [ 991.431506] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.436839] env[61957]: INFO nova.virt.vmwareapi.images [None req-8b32736b-e3a6-485c-a95e-afec23b58d86 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] [instance: 736eabcf-70f8-4e1e-9830-270d9971fc9a] The imported VM was unregistered [ 991.438928] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-8b32736b-e3a6-485c-a95e-afec23b58d86 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] [instance: 736eabcf-70f8-4e1e-9830-270d9971fc9a] Caching image {{(pid=61957) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 991.439321] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-8b32736b-e3a6-485c-a95e-afec23b58d86 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Creating directory with path [datastore2] devstack-image-cache_base/0018d8e9-1d6e-4ca4-a224-31fc9e9f9d2e {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 991.440637] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-adbeeea1-1eed-4a98-bdfa-f422c9717eb0 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.449589] env[61957]: DEBUG oslo_vmware.api [None req-6d13ccf5-9ccd-48d0-a35e-c6df97be9c15 tempest-ServerAddressesTestJSON-2054963291 tempest-ServerAddressesTestJSON-2054963291-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52b3687d-e7d5-93af-de70-8e5b4d14c0ab, 'name': SearchDatastore_Task, 'duration_secs': 0.011765} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.453159] env[61957]: DEBUG oslo_concurrency.lockutils [None req-6d13ccf5-9ccd-48d0-a35e-c6df97be9c15 tempest-ServerAddressesTestJSON-2054963291 tempest-ServerAddressesTestJSON-2054963291-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 991.453159] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-6d13ccf5-9ccd-48d0-a35e-c6df97be9c15 tempest-ServerAddressesTestJSON-2054963291 tempest-ServerAddressesTestJSON-2054963291-project-member] [instance: fa4c150e-7a85-41a4-8890-9f2deb3b34b1] Processing image 11c76a2c-f705-470a-ba9d-4657858bab38 {{(pid=61957) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 991.453273] env[61957]: DEBUG oslo_concurrency.lockutils [None req-6d13ccf5-9ccd-48d0-a35e-c6df97be9c15 tempest-ServerAddressesTestJSON-2054963291 tempest-ServerAddressesTestJSON-2054963291-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 991.453339] env[61957]: DEBUG oslo_concurrency.lockutils [None req-6d13ccf5-9ccd-48d0-a35e-c6df97be9c15 tempest-ServerAddressesTestJSON-2054963291 tempest-ServerAddressesTestJSON-2054963291-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 991.453521] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-6d13ccf5-9ccd-48d0-a35e-c6df97be9c15 tempest-ServerAddressesTestJSON-2054963291 tempest-ServerAddressesTestJSON-2054963291-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 991.454995] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-de7de108-0ce5-4562-92b5-e7aa0401bb87 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.462151] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-8b32736b-e3a6-485c-a95e-afec23b58d86 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Created directory with path [datastore2] devstack-image-cache_base/0018d8e9-1d6e-4ca4-a224-31fc9e9f9d2e {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 991.462359] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-8b32736b-e3a6-485c-a95e-afec23b58d86 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Moving virtual disk from [datastore2] OSTACK_IMG_e717072e-0b5a-4ae5-acc6-5d7a40856a3f/OSTACK_IMG_e717072e-0b5a-4ae5-acc6-5d7a40856a3f.vmdk to [datastore2] devstack-image-cache_base/0018d8e9-1d6e-4ca4-a224-31fc9e9f9d2e/0018d8e9-1d6e-4ca4-a224-31fc9e9f9d2e.vmdk. {{(pid=61957) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 991.463488] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-1f427816-8399-4fe0-9154-c02971ab993b {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.466256] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-6d13ccf5-9ccd-48d0-a35e-c6df97be9c15 tempest-ServerAddressesTestJSON-2054963291 tempest-ServerAddressesTestJSON-2054963291-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 991.466443] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-6d13ccf5-9ccd-48d0-a35e-c6df97be9c15 tempest-ServerAddressesTestJSON-2054963291 tempest-ServerAddressesTestJSON-2054963291-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61957) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 991.470013] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-681b53d4-2006-4853-8cac-872dec28cad7 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.476132] env[61957]: DEBUG oslo_vmware.api [None req-6d13ccf5-9ccd-48d0-a35e-c6df97be9c15 tempest-ServerAddressesTestJSON-2054963291 tempest-ServerAddressesTestJSON-2054963291-project-member] Waiting for the task: (returnval){ [ 991.476132] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52d7ad77-c4c0-51ff-a293-8645cb58f554" [ 991.476132] env[61957]: _type = "Task" [ 991.476132] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.481113] env[61957]: DEBUG oslo_vmware.api [None req-8b32736b-e3a6-485c-a95e-afec23b58d86 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Waiting for the task: (returnval){ [ 991.481113] env[61957]: value = "task-1277898" [ 991.481113] env[61957]: _type = "Task" [ 991.481113] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.490044] env[61957]: DEBUG oslo_vmware.api [None req-6d13ccf5-9ccd-48d0-a35e-c6df97be9c15 tempest-ServerAddressesTestJSON-2054963291 tempest-ServerAddressesTestJSON-2054963291-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52d7ad77-c4c0-51ff-a293-8645cb58f554, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.495802] env[61957]: DEBUG oslo_vmware.api [None req-8b32736b-e3a6-485c-a95e-afec23b58d86 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Task: {'id': task-1277898, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.553025] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cc3dc13-e4a7-4f2b-85bb-2873e4098359 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.561646] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd1de56e-ea01-42cd-99a2-418f948c62c4 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.592285] env[61957]: DEBUG oslo_concurrency.lockutils [None req-731394e9-d0a3-46c3-8b02-41ef72d369f5 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Lock "f65458b2-0db6-4d98-8c82-b27b6945a3fd" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 27.446s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 991.593665] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d60b1935-13bf-4b50-b371-d08f1781704b {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.601965] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f8ecd62-eca0-4190-8826-0164e70b88b7 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.617412] env[61957]: DEBUG nova.compute.provider_tree [None req-74e1032e-bbb9-4915-a886-4f8885996564 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 991.787765] env[61957]: DEBUG nova.network.neutron [req-3f469708-25dd-4783-84a5-996a112ec595 req-b430dae8-7755-478e-a693-3c310f9a9e10 service nova] [instance: fa4c150e-7a85-41a4-8890-9f2deb3b34b1] Updated VIF entry in instance network info cache for port 4b2ccba5-48f1-44f1-8746-3792d3695e58. {{(pid=61957) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 991.788318] env[61957]: DEBUG nova.network.neutron [req-3f469708-25dd-4783-84a5-996a112ec595 req-b430dae8-7755-478e-a693-3c310f9a9e10 service nova] [instance: fa4c150e-7a85-41a4-8890-9f2deb3b34b1] Updating instance_info_cache with network_info: [{"id": "4b2ccba5-48f1-44f1-8746-3792d3695e58", "address": "fa:16:3e:83:a2:9e", "network": {"id": "e45c12da-f92f-4516-a616-e4478247693e", "bridge": "br-int", "label": "tempest-ServerAddressesTestJSON-96077499-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d5dffaf886624e1381979f2061556d2a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d0bdd11b-58af-4cc0-9d38-8322e1bb4e74", "external-id": "nsx-vlan-transportzone-398", "segmentation_id": 398, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4b2ccba5-48", "ovs_interfaceid": "4b2ccba5-48f1-44f1-8746-3792d3695e58", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 991.925648] env[61957]: DEBUG oslo_concurrency.lockutils [None req-bc1354f0-0ce9-464c-89e1-e1e216546ebe tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Acquiring lock "refresh_cache-6509707e-afc6-4c4a-88f9-a471865d510c" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 991.925816] env[61957]: DEBUG oslo_concurrency.lockutils [None req-bc1354f0-0ce9-464c-89e1-e1e216546ebe tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Acquired lock "refresh_cache-6509707e-afc6-4c4a-88f9-a471865d510c" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 991.925977] env[61957]: DEBUG nova.network.neutron [None req-bc1354f0-0ce9-464c-89e1-e1e216546ebe tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: 6509707e-afc6-4c4a-88f9-a471865d510c] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 991.990152] env[61957]: DEBUG oslo_vmware.api [None req-6d13ccf5-9ccd-48d0-a35e-c6df97be9c15 tempest-ServerAddressesTestJSON-2054963291 tempest-ServerAddressesTestJSON-2054963291-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52d7ad77-c4c0-51ff-a293-8645cb58f554, 'name': SearchDatastore_Task, 'duration_secs': 0.010984} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.991557] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3f4230ca-dd0f-43f0-bb81-064bee30d06b {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.998704] env[61957]: DEBUG oslo_vmware.api [None req-8b32736b-e3a6-485c-a95e-afec23b58d86 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Task: {'id': task-1277898, 'name': MoveVirtualDisk_Task} progress is 24%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.000422] env[61957]: DEBUG oslo_vmware.api [None req-6d13ccf5-9ccd-48d0-a35e-c6df97be9c15 tempest-ServerAddressesTestJSON-2054963291 tempest-ServerAddressesTestJSON-2054963291-project-member] Waiting for the task: (returnval){ [ 992.000422] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52c0a9f4-dfb2-f90c-2b62-4910d0fd03fc" [ 992.000422] env[61957]: _type = "Task" [ 992.000422] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.011928] env[61957]: DEBUG oslo_vmware.api [None req-6d13ccf5-9ccd-48d0-a35e-c6df97be9c15 tempest-ServerAddressesTestJSON-2054963291 tempest-ServerAddressesTestJSON-2054963291-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52c0a9f4-dfb2-f90c-2b62-4910d0fd03fc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.122383] env[61957]: DEBUG nova.scheduler.client.report [None req-74e1032e-bbb9-4915-a886-4f8885996564 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 992.294050] env[61957]: DEBUG oslo_concurrency.lockutils [req-3f469708-25dd-4783-84a5-996a112ec595 req-b430dae8-7755-478e-a693-3c310f9a9e10 service nova] Releasing lock "refresh_cache-fa4c150e-7a85-41a4-8890-9f2deb3b34b1" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 992.331188] env[61957]: INFO nova.compute.manager [None req-a5bdcddf-64c6-4110-8e30-a016fe726a7a tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] [instance: f65458b2-0db6-4d98-8c82-b27b6945a3fd] Rescuing [ 992.331331] env[61957]: DEBUG oslo_concurrency.lockutils [None req-a5bdcddf-64c6-4110-8e30-a016fe726a7a tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Acquiring lock "refresh_cache-f65458b2-0db6-4d98-8c82-b27b6945a3fd" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 992.331501] env[61957]: DEBUG oslo_concurrency.lockutils [None req-a5bdcddf-64c6-4110-8e30-a016fe726a7a tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Acquired lock "refresh_cache-f65458b2-0db6-4d98-8c82-b27b6945a3fd" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 992.331682] env[61957]: DEBUG nova.network.neutron [None req-a5bdcddf-64c6-4110-8e30-a016fe726a7a tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] [instance: f65458b2-0db6-4d98-8c82-b27b6945a3fd] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 992.472773] env[61957]: DEBUG nova.network.neutron [None req-bc1354f0-0ce9-464c-89e1-e1e216546ebe tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: 6509707e-afc6-4c4a-88f9-a471865d510c] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 992.493252] env[61957]: DEBUG oslo_vmware.api [None req-8b32736b-e3a6-485c-a95e-afec23b58d86 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Task: {'id': task-1277898, 'name': MoveVirtualDisk_Task} progress is 43%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.512154] env[61957]: DEBUG oslo_vmware.api [None req-6d13ccf5-9ccd-48d0-a35e-c6df97be9c15 tempest-ServerAddressesTestJSON-2054963291 tempest-ServerAddressesTestJSON-2054963291-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52c0a9f4-dfb2-f90c-2b62-4910d0fd03fc, 'name': SearchDatastore_Task, 'duration_secs': 0.085354} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.512576] env[61957]: DEBUG oslo_concurrency.lockutils [None req-6d13ccf5-9ccd-48d0-a35e-c6df97be9c15 tempest-ServerAddressesTestJSON-2054963291 tempest-ServerAddressesTestJSON-2054963291-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 992.512988] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d13ccf5-9ccd-48d0-a35e-c6df97be9c15 tempest-ServerAddressesTestJSON-2054963291 tempest-ServerAddressesTestJSON-2054963291-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore2] fa4c150e-7a85-41a4-8890-9f2deb3b34b1/fa4c150e-7a85-41a4-8890-9f2deb3b34b1.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 992.513431] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-245c3c05-6a5d-47b8-b8d5-7e691cf5f946 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.525713] env[61957]: DEBUG oslo_vmware.api [None req-6d13ccf5-9ccd-48d0-a35e-c6df97be9c15 tempest-ServerAddressesTestJSON-2054963291 tempest-ServerAddressesTestJSON-2054963291-project-member] Waiting for the task: (returnval){ [ 992.525713] env[61957]: value = "task-1277899" [ 992.525713] env[61957]: _type = "Task" [ 992.525713] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.538162] env[61957]: DEBUG oslo_vmware.api [None req-6d13ccf5-9ccd-48d0-a35e-c6df97be9c15 tempest-ServerAddressesTestJSON-2054963291 tempest-ServerAddressesTestJSON-2054963291-project-member] Task: {'id': task-1277899, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.573660] env[61957]: DEBUG oslo_concurrency.lockutils [None req-4614d338-72f4-4519-9a7d-0a36e19fbd39 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Acquiring lock "d4674fbc-47b1-42d1-aaba-e86d46c51e8f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 992.574030] env[61957]: DEBUG oslo_concurrency.lockutils [None req-4614d338-72f4-4519-9a7d-0a36e19fbd39 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Lock "d4674fbc-47b1-42d1-aaba-e86d46c51e8f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 992.574234] env[61957]: DEBUG oslo_concurrency.lockutils [None req-4614d338-72f4-4519-9a7d-0a36e19fbd39 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Acquiring lock "d4674fbc-47b1-42d1-aaba-e86d46c51e8f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 992.574484] env[61957]: DEBUG oslo_concurrency.lockutils [None req-4614d338-72f4-4519-9a7d-0a36e19fbd39 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Lock "d4674fbc-47b1-42d1-aaba-e86d46c51e8f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 992.574646] env[61957]: DEBUG oslo_concurrency.lockutils [None req-4614d338-72f4-4519-9a7d-0a36e19fbd39 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Lock "d4674fbc-47b1-42d1-aaba-e86d46c51e8f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 992.576940] env[61957]: INFO nova.compute.manager [None req-4614d338-72f4-4519-9a7d-0a36e19fbd39 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: d4674fbc-47b1-42d1-aaba-e86d46c51e8f] Terminating instance [ 992.579464] env[61957]: DEBUG nova.compute.manager [None req-4614d338-72f4-4519-9a7d-0a36e19fbd39 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: d4674fbc-47b1-42d1-aaba-e86d46c51e8f] Start destroying the instance on the hypervisor. {{(pid=61957) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 992.580033] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-4614d338-72f4-4519-9a7d-0a36e19fbd39 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: d4674fbc-47b1-42d1-aaba-e86d46c51e8f] Destroying instance {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 992.580923] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd531fd7-0510-4c6a-a3a0-b9668a18c15d {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.591526] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-4614d338-72f4-4519-9a7d-0a36e19fbd39 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: d4674fbc-47b1-42d1-aaba-e86d46c51e8f] Powering off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 992.591878] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e2ceff09-2c37-46e5-9747-7ba9aba024a6 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.601099] env[61957]: DEBUG oslo_vmware.api [None req-4614d338-72f4-4519-9a7d-0a36e19fbd39 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Waiting for the task: (returnval){ [ 992.601099] env[61957]: value = "task-1277900" [ 992.601099] env[61957]: _type = "Task" [ 992.601099] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.616631] env[61957]: DEBUG oslo_vmware.api [None req-4614d338-72f4-4519-9a7d-0a36e19fbd39 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': task-1277900, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.628103] env[61957]: DEBUG oslo_concurrency.lockutils [None req-74e1032e-bbb9-4915-a886-4f8885996564 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.836s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 992.651958] env[61957]: DEBUG nova.network.neutron [None req-bc1354f0-0ce9-464c-89e1-e1e216546ebe tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: 6509707e-afc6-4c4a-88f9-a471865d510c] Updating instance_info_cache with network_info: [{"id": "7df2736b-10d0-4f9d-83aa-e2bbdf1ef45b", "address": "fa:16:3e:2e:da:a9", "network": {"id": "6fbccc17-6674-4cd5-8f1b-aa73570baac7", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-394544811-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1dbdc699741a48af93c52068d88a5357", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "20e3f794-c7a3-4696-9488-ecf34c570ef9", "external-id": "nsx-vlan-transportzone-509", "segmentation_id": 509, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7df2736b-10", "ovs_interfaceid": "7df2736b-10d0-4f9d-83aa-e2bbdf1ef45b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 992.658104] env[61957]: INFO nova.scheduler.client.report [None req-74e1032e-bbb9-4915-a886-4f8885996564 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Deleted allocations for instance 6f45c7c8-de07-4d21-bccc-e4afe3c8ad3d [ 992.996205] env[61957]: DEBUG oslo_vmware.api [None req-8b32736b-e3a6-485c-a95e-afec23b58d86 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Task: {'id': task-1277898, 'name': MoveVirtualDisk_Task} progress is 66%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.038049] env[61957]: DEBUG oslo_vmware.api [None req-6d13ccf5-9ccd-48d0-a35e-c6df97be9c15 tempest-ServerAddressesTestJSON-2054963291 tempest-ServerAddressesTestJSON-2054963291-project-member] Task: {'id': task-1277899, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.081508] env[61957]: DEBUG nova.network.neutron [None req-a5bdcddf-64c6-4110-8e30-a016fe726a7a tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] [instance: f65458b2-0db6-4d98-8c82-b27b6945a3fd] Updating instance_info_cache with network_info: [{"id": "bb715e7d-14c5-418b-b3ee-1ac91aba709d", "address": "fa:16:3e:fd:f1:f4", "network": {"id": "d55c72c6-107d-4965-b4e1-d18383813953", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-193304765-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.2", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "7758bdd10505455b8cecb8129d70ebd5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c24464bb-bb6b-43a2-bdcd-8086ad1a307f", "external-id": "nsx-vlan-transportzone-781", "segmentation_id": 781, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbb715e7d-14", "ovs_interfaceid": "bb715e7d-14c5-418b-b3ee-1ac91aba709d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 993.115673] env[61957]: DEBUG oslo_vmware.api [None req-4614d338-72f4-4519-9a7d-0a36e19fbd39 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': task-1277900, 'name': PowerOffVM_Task, 'duration_secs': 0.390661} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.115673] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-4614d338-72f4-4519-9a7d-0a36e19fbd39 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: d4674fbc-47b1-42d1-aaba-e86d46c51e8f] Powered off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 993.115673] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-4614d338-72f4-4519-9a7d-0a36e19fbd39 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: d4674fbc-47b1-42d1-aaba-e86d46c51e8f] Unregistering the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 993.115673] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-69209fd6-575f-4ea0-ab91-10d6936eabba {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.157072] env[61957]: DEBUG oslo_concurrency.lockutils [None req-bc1354f0-0ce9-464c-89e1-e1e216546ebe tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Releasing lock "refresh_cache-6509707e-afc6-4c4a-88f9-a471865d510c" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 993.157072] env[61957]: DEBUG nova.compute.manager [None req-bc1354f0-0ce9-464c-89e1-e1e216546ebe tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: 6509707e-afc6-4c4a-88f9-a471865d510c] Instance network_info: |[{"id": "7df2736b-10d0-4f9d-83aa-e2bbdf1ef45b", "address": "fa:16:3e:2e:da:a9", "network": {"id": "6fbccc17-6674-4cd5-8f1b-aa73570baac7", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-394544811-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1dbdc699741a48af93c52068d88a5357", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "20e3f794-c7a3-4696-9488-ecf34c570ef9", "external-id": "nsx-vlan-transportzone-509", "segmentation_id": 509, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7df2736b-10", "ovs_interfaceid": "7df2736b-10d0-4f9d-83aa-e2bbdf1ef45b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 993.157072] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-bc1354f0-0ce9-464c-89e1-e1e216546ebe tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: 6509707e-afc6-4c4a-88f9-a471865d510c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:2e:da:a9', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '20e3f794-c7a3-4696-9488-ecf34c570ef9', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7df2736b-10d0-4f9d-83aa-e2bbdf1ef45b', 'vif_model': 'vmxnet3'}] {{(pid=61957) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 993.166064] env[61957]: DEBUG oslo.service.loopingcall [None req-bc1354f0-0ce9-464c-89e1-e1e216546ebe tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 993.172048] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6509707e-afc6-4c4a-88f9-a471865d510c] Creating VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 993.172048] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b5e11648-f857-481c-9e1c-884ba51791f3 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.187534] env[61957]: DEBUG oslo_concurrency.lockutils [None req-74e1032e-bbb9-4915-a886-4f8885996564 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Lock "6f45c7c8-de07-4d21-bccc-e4afe3c8ad3d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.138s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 993.198985] env[61957]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 993.198985] env[61957]: value = "task-1277902" [ 993.198985] env[61957]: _type = "Task" [ 993.198985] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.211783] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277902, 'name': CreateVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.267264] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-4614d338-72f4-4519-9a7d-0a36e19fbd39 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: d4674fbc-47b1-42d1-aaba-e86d46c51e8f] Unregistered the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 993.267806] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-4614d338-72f4-4519-9a7d-0a36e19fbd39 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: d4674fbc-47b1-42d1-aaba-e86d46c51e8f] Deleting contents of the VM from datastore datastore2 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 993.267806] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-4614d338-72f4-4519-9a7d-0a36e19fbd39 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Deleting the datastore file [datastore2] d4674fbc-47b1-42d1-aaba-e86d46c51e8f {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 993.268234] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a2093799-9aff-44f4-9d23-6b09d8fdd30e {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.279514] env[61957]: DEBUG oslo_vmware.api [None req-4614d338-72f4-4519-9a7d-0a36e19fbd39 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Waiting for the task: (returnval){ [ 993.279514] env[61957]: value = "task-1277903" [ 993.279514] env[61957]: _type = "Task" [ 993.279514] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.292237] env[61957]: DEBUG oslo_vmware.api [None req-4614d338-72f4-4519-9a7d-0a36e19fbd39 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': task-1277903, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.330640] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ab927ca0-dc5c-48f5-8ef0-a84245f36801 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Acquiring lock "e2107d5c-8453-4d24-b364-2b5c1d125062" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 993.331486] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ab927ca0-dc5c-48f5-8ef0-a84245f36801 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Lock "e2107d5c-8453-4d24-b364-2b5c1d125062" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 993.343659] env[61957]: DEBUG nova.compute.manager [req-d51d4229-0fd5-46b1-a8e6-a3c02e676be0 req-3a0bbc78-cd07-4f4e-9a88-dd9cfc3cebb1 service nova] [instance: 6509707e-afc6-4c4a-88f9-a471865d510c] Received event network-vif-plugged-7df2736b-10d0-4f9d-83aa-e2bbdf1ef45b {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 993.343790] env[61957]: DEBUG oslo_concurrency.lockutils [req-d51d4229-0fd5-46b1-a8e6-a3c02e676be0 req-3a0bbc78-cd07-4f4e-9a88-dd9cfc3cebb1 service nova] Acquiring lock "6509707e-afc6-4c4a-88f9-a471865d510c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 993.344272] env[61957]: DEBUG oslo_concurrency.lockutils [req-d51d4229-0fd5-46b1-a8e6-a3c02e676be0 req-3a0bbc78-cd07-4f4e-9a88-dd9cfc3cebb1 service nova] Lock "6509707e-afc6-4c4a-88f9-a471865d510c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 993.344272] env[61957]: DEBUG oslo_concurrency.lockutils [req-d51d4229-0fd5-46b1-a8e6-a3c02e676be0 req-3a0bbc78-cd07-4f4e-9a88-dd9cfc3cebb1 service nova] Lock "6509707e-afc6-4c4a-88f9-a471865d510c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 993.344415] env[61957]: DEBUG nova.compute.manager [req-d51d4229-0fd5-46b1-a8e6-a3c02e676be0 req-3a0bbc78-cd07-4f4e-9a88-dd9cfc3cebb1 service nova] [instance: 6509707e-afc6-4c4a-88f9-a471865d510c] No waiting events found dispatching network-vif-plugged-7df2736b-10d0-4f9d-83aa-e2bbdf1ef45b {{(pid=61957) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 993.345147] env[61957]: WARNING nova.compute.manager [req-d51d4229-0fd5-46b1-a8e6-a3c02e676be0 req-3a0bbc78-cd07-4f4e-9a88-dd9cfc3cebb1 service nova] [instance: 6509707e-afc6-4c4a-88f9-a471865d510c] Received unexpected event network-vif-plugged-7df2736b-10d0-4f9d-83aa-e2bbdf1ef45b for instance with vm_state building and task_state spawning. [ 993.345147] env[61957]: DEBUG nova.compute.manager [req-d51d4229-0fd5-46b1-a8e6-a3c02e676be0 req-3a0bbc78-cd07-4f4e-9a88-dd9cfc3cebb1 service nova] [instance: 6509707e-afc6-4c4a-88f9-a471865d510c] Received event network-changed-7df2736b-10d0-4f9d-83aa-e2bbdf1ef45b {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 993.345147] env[61957]: DEBUG nova.compute.manager [req-d51d4229-0fd5-46b1-a8e6-a3c02e676be0 req-3a0bbc78-cd07-4f4e-9a88-dd9cfc3cebb1 service nova] [instance: 6509707e-afc6-4c4a-88f9-a471865d510c] Refreshing instance network info cache due to event network-changed-7df2736b-10d0-4f9d-83aa-e2bbdf1ef45b. {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 993.345147] env[61957]: DEBUG oslo_concurrency.lockutils [req-d51d4229-0fd5-46b1-a8e6-a3c02e676be0 req-3a0bbc78-cd07-4f4e-9a88-dd9cfc3cebb1 service nova] Acquiring lock "refresh_cache-6509707e-afc6-4c4a-88f9-a471865d510c" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 993.345348] env[61957]: DEBUG oslo_concurrency.lockutils [req-d51d4229-0fd5-46b1-a8e6-a3c02e676be0 req-3a0bbc78-cd07-4f4e-9a88-dd9cfc3cebb1 service nova] Acquired lock "refresh_cache-6509707e-afc6-4c4a-88f9-a471865d510c" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 993.345465] env[61957]: DEBUG nova.network.neutron [req-d51d4229-0fd5-46b1-a8e6-a3c02e676be0 req-3a0bbc78-cd07-4f4e-9a88-dd9cfc3cebb1 service nova] [instance: 6509707e-afc6-4c4a-88f9-a471865d510c] Refreshing network info cache for port 7df2736b-10d0-4f9d-83aa-e2bbdf1ef45b {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 993.499974] env[61957]: DEBUG oslo_vmware.api [None req-8b32736b-e3a6-485c-a95e-afec23b58d86 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Task: {'id': task-1277898, 'name': MoveVirtualDisk_Task} progress is 88%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.537939] env[61957]: DEBUG oslo_vmware.api [None req-6d13ccf5-9ccd-48d0-a35e-c6df97be9c15 tempest-ServerAddressesTestJSON-2054963291 tempest-ServerAddressesTestJSON-2054963291-project-member] Task: {'id': task-1277899, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.583908] env[61957]: DEBUG oslo_concurrency.lockutils [None req-a5bdcddf-64c6-4110-8e30-a016fe726a7a tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Releasing lock "refresh_cache-f65458b2-0db6-4d98-8c82-b27b6945a3fd" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 993.713043] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277902, 'name': CreateVM_Task} progress is 99%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.790908] env[61957]: DEBUG oslo_vmware.api [None req-4614d338-72f4-4519-9a7d-0a36e19fbd39 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': task-1277903, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.834897] env[61957]: DEBUG nova.compute.manager [None req-ab927ca0-dc5c-48f5-8ef0-a84245f36801 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: e2107d5c-8453-4d24-b364-2b5c1d125062] Starting instance... {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 993.987422] env[61957]: DEBUG oslo_concurrency.lockutils [None req-8cf153c7-e439-4990-9fa2-f6d040c68c3c tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Acquiring lock "84cc9489-2a11-4510-8ad6-13dc41814856" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 993.987422] env[61957]: DEBUG oslo_concurrency.lockutils [None req-8cf153c7-e439-4990-9fa2-f6d040c68c3c tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Lock "84cc9489-2a11-4510-8ad6-13dc41814856" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 994.004330] env[61957]: DEBUG oslo_vmware.api [None req-8b32736b-e3a6-485c-a95e-afec23b58d86 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Task: {'id': task-1277898, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.393974} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.005383] env[61957]: INFO nova.virt.vmwareapi.ds_util [None req-8b32736b-e3a6-485c-a95e-afec23b58d86 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Moved virtual disk from [datastore2] OSTACK_IMG_e717072e-0b5a-4ae5-acc6-5d7a40856a3f/OSTACK_IMG_e717072e-0b5a-4ae5-acc6-5d7a40856a3f.vmdk to [datastore2] devstack-image-cache_base/0018d8e9-1d6e-4ca4-a224-31fc9e9f9d2e/0018d8e9-1d6e-4ca4-a224-31fc9e9f9d2e.vmdk. [ 994.005685] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-8b32736b-e3a6-485c-a95e-afec23b58d86 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] [instance: 736eabcf-70f8-4e1e-9830-270d9971fc9a] Cleaning up location [datastore2] OSTACK_IMG_e717072e-0b5a-4ae5-acc6-5d7a40856a3f {{(pid=61957) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 994.005756] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-8b32736b-e3a6-485c-a95e-afec23b58d86 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Deleting the datastore file [datastore2] OSTACK_IMG_e717072e-0b5a-4ae5-acc6-5d7a40856a3f {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 994.006379] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-102e928b-45c0-429c-b623-abcdeb09fff5 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.016552] env[61957]: DEBUG oslo_vmware.api [None req-8b32736b-e3a6-485c-a95e-afec23b58d86 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Waiting for the task: (returnval){ [ 994.016552] env[61957]: value = "task-1277904" [ 994.016552] env[61957]: _type = "Task" [ 994.016552] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.028324] env[61957]: DEBUG oslo_vmware.api [None req-8b32736b-e3a6-485c-a95e-afec23b58d86 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Task: {'id': task-1277904, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.043164] env[61957]: DEBUG oslo_vmware.api [None req-6d13ccf5-9ccd-48d0-a35e-c6df97be9c15 tempest-ServerAddressesTestJSON-2054963291 tempest-ServerAddressesTestJSON-2054963291-project-member] Task: {'id': task-1277899, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.114570] env[61957]: DEBUG nova.network.neutron [req-d51d4229-0fd5-46b1-a8e6-a3c02e676be0 req-3a0bbc78-cd07-4f4e-9a88-dd9cfc3cebb1 service nova] [instance: 6509707e-afc6-4c4a-88f9-a471865d510c] Updated VIF entry in instance network info cache for port 7df2736b-10d0-4f9d-83aa-e2bbdf1ef45b. {{(pid=61957) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 994.115562] env[61957]: DEBUG nova.network.neutron [req-d51d4229-0fd5-46b1-a8e6-a3c02e676be0 req-3a0bbc78-cd07-4f4e-9a88-dd9cfc3cebb1 service nova] [instance: 6509707e-afc6-4c4a-88f9-a471865d510c] Updating instance_info_cache with network_info: [{"id": "7df2736b-10d0-4f9d-83aa-e2bbdf1ef45b", "address": "fa:16:3e:2e:da:a9", "network": {"id": "6fbccc17-6674-4cd5-8f1b-aa73570baac7", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-394544811-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1dbdc699741a48af93c52068d88a5357", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "20e3f794-c7a3-4696-9488-ecf34c570ef9", "external-id": "nsx-vlan-transportzone-509", "segmentation_id": 509, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7df2736b-10", "ovs_interfaceid": "7df2736b-10d0-4f9d-83aa-e2bbdf1ef45b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 994.135372] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-a5bdcddf-64c6-4110-8e30-a016fe726a7a tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] [instance: f65458b2-0db6-4d98-8c82-b27b6945a3fd] Powering off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 994.135677] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a4af8ead-7e0b-4532-bae0-574371771937 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.146995] env[61957]: DEBUG oslo_vmware.api [None req-a5bdcddf-64c6-4110-8e30-a016fe726a7a tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Waiting for the task: (returnval){ [ 994.146995] env[61957]: value = "task-1277905" [ 994.146995] env[61957]: _type = "Task" [ 994.146995] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.157454] env[61957]: DEBUG oslo_vmware.api [None req-a5bdcddf-64c6-4110-8e30-a016fe726a7a tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Task: {'id': task-1277905, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.210758] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277902, 'name': CreateVM_Task, 'duration_secs': 0.558432} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.211087] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6509707e-afc6-4c4a-88f9-a471865d510c] Created VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 994.211650] env[61957]: DEBUG oslo_concurrency.lockutils [None req-bc1354f0-0ce9-464c-89e1-e1e216546ebe tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 994.211885] env[61957]: DEBUG oslo_concurrency.lockutils [None req-bc1354f0-0ce9-464c-89e1-e1e216546ebe tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Acquired lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 994.212283] env[61957]: DEBUG oslo_concurrency.lockutils [None req-bc1354f0-0ce9-464c-89e1-e1e216546ebe tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 994.212566] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b388fc60-ddea-4c34-87a0-bed701b64b31 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.218436] env[61957]: DEBUG oslo_vmware.api [None req-bc1354f0-0ce9-464c-89e1-e1e216546ebe tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Waiting for the task: (returnval){ [ 994.218436] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52a7ab02-9560-de48-d161-518eb95e7b11" [ 994.218436] env[61957]: _type = "Task" [ 994.218436] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.228145] env[61957]: DEBUG oslo_vmware.api [None req-bc1354f0-0ce9-464c-89e1-e1e216546ebe tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52a7ab02-9560-de48-d161-518eb95e7b11, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.290600] env[61957]: DEBUG oslo_vmware.api [None req-4614d338-72f4-4519-9a7d-0a36e19fbd39 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': task-1277903, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.896276} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.290870] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-4614d338-72f4-4519-9a7d-0a36e19fbd39 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Deleted the datastore file {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 994.291070] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-4614d338-72f4-4519-9a7d-0a36e19fbd39 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: d4674fbc-47b1-42d1-aaba-e86d46c51e8f] Deleted contents of the VM from datastore datastore2 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 994.291264] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-4614d338-72f4-4519-9a7d-0a36e19fbd39 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: d4674fbc-47b1-42d1-aaba-e86d46c51e8f] Instance destroyed {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 994.291440] env[61957]: INFO nova.compute.manager [None req-4614d338-72f4-4519-9a7d-0a36e19fbd39 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: d4674fbc-47b1-42d1-aaba-e86d46c51e8f] Took 1.71 seconds to destroy the instance on the hypervisor. [ 994.291691] env[61957]: DEBUG oslo.service.loopingcall [None req-4614d338-72f4-4519-9a7d-0a36e19fbd39 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 994.292045] env[61957]: DEBUG nova.compute.manager [-] [instance: d4674fbc-47b1-42d1-aaba-e86d46c51e8f] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 994.292169] env[61957]: DEBUG nova.network.neutron [-] [instance: d4674fbc-47b1-42d1-aaba-e86d46c51e8f] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 994.355860] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ab927ca0-dc5c-48f5-8ef0-a84245f36801 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 994.356430] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ab927ca0-dc5c-48f5-8ef0-a84245f36801 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 994.358037] env[61957]: INFO nova.compute.claims [None req-ab927ca0-dc5c-48f5-8ef0-a84245f36801 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: e2107d5c-8453-4d24-b364-2b5c1d125062] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 994.489591] env[61957]: DEBUG nova.compute.manager [None req-8cf153c7-e439-4990-9fa2-f6d040c68c3c tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 84cc9489-2a11-4510-8ad6-13dc41814856] Starting instance... {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 994.529152] env[61957]: DEBUG oslo_vmware.api [None req-8b32736b-e3a6-485c-a95e-afec23b58d86 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Task: {'id': task-1277904, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.140429} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.533020] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-8b32736b-e3a6-485c-a95e-afec23b58d86 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Deleted the datastore file {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 994.533262] env[61957]: DEBUG oslo_concurrency.lockutils [None req-8b32736b-e3a6-485c-a95e-afec23b58d86 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0018d8e9-1d6e-4ca4-a224-31fc9e9f9d2e/0018d8e9-1d6e-4ca4-a224-31fc9e9f9d2e.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 994.533513] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-8b32736b-e3a6-485c-a95e-afec23b58d86 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0018d8e9-1d6e-4ca4-a224-31fc9e9f9d2e/0018d8e9-1d6e-4ca4-a224-31fc9e9f9d2e.vmdk to [datastore2] 736eabcf-70f8-4e1e-9830-270d9971fc9a/736eabcf-70f8-4e1e-9830-270d9971fc9a.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 994.534061] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-42a3c740-314e-4cf2-9c88-54a8b8056391 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.542067] env[61957]: DEBUG oslo_vmware.api [None req-6d13ccf5-9ccd-48d0-a35e-c6df97be9c15 tempest-ServerAddressesTestJSON-2054963291 tempest-ServerAddressesTestJSON-2054963291-project-member] Task: {'id': task-1277899, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.634291} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.543416] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d13ccf5-9ccd-48d0-a35e-c6df97be9c15 tempest-ServerAddressesTestJSON-2054963291 tempest-ServerAddressesTestJSON-2054963291-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore2] fa4c150e-7a85-41a4-8890-9f2deb3b34b1/fa4c150e-7a85-41a4-8890-9f2deb3b34b1.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 994.543652] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-6d13ccf5-9ccd-48d0-a35e-c6df97be9c15 tempest-ServerAddressesTestJSON-2054963291 tempest-ServerAddressesTestJSON-2054963291-project-member] [instance: fa4c150e-7a85-41a4-8890-9f2deb3b34b1] Extending root virtual disk to 1048576 {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 994.543949] env[61957]: DEBUG oslo_vmware.api [None req-8b32736b-e3a6-485c-a95e-afec23b58d86 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Waiting for the task: (returnval){ [ 994.543949] env[61957]: value = "task-1277906" [ 994.543949] env[61957]: _type = "Task" [ 994.543949] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.544295] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f3e8e7a6-b87d-45ab-9df1-945dbc35a151 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.554843] env[61957]: DEBUG oslo_vmware.api [None req-8b32736b-e3a6-485c-a95e-afec23b58d86 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Task: {'id': task-1277906, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.556247] env[61957]: DEBUG oslo_vmware.api [None req-6d13ccf5-9ccd-48d0-a35e-c6df97be9c15 tempest-ServerAddressesTestJSON-2054963291 tempest-ServerAddressesTestJSON-2054963291-project-member] Waiting for the task: (returnval){ [ 994.556247] env[61957]: value = "task-1277907" [ 994.556247] env[61957]: _type = "Task" [ 994.556247] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.566500] env[61957]: DEBUG oslo_vmware.api [None req-6d13ccf5-9ccd-48d0-a35e-c6df97be9c15 tempest-ServerAddressesTestJSON-2054963291 tempest-ServerAddressesTestJSON-2054963291-project-member] Task: {'id': task-1277907, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.618047] env[61957]: DEBUG oslo_concurrency.lockutils [req-d51d4229-0fd5-46b1-a8e6-a3c02e676be0 req-3a0bbc78-cd07-4f4e-9a88-dd9cfc3cebb1 service nova] Releasing lock "refresh_cache-6509707e-afc6-4c4a-88f9-a471865d510c" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 994.657642] env[61957]: DEBUG oslo_vmware.api [None req-a5bdcddf-64c6-4110-8e30-a016fe726a7a tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Task: {'id': task-1277905, 'name': PowerOffVM_Task, 'duration_secs': 0.305679} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.660577] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-a5bdcddf-64c6-4110-8e30-a016fe726a7a tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] [instance: f65458b2-0db6-4d98-8c82-b27b6945a3fd] Powered off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 994.662875] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a364f102-5b36-4ffc-9275-39c31a2968b0 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.669612] env[61957]: DEBUG oslo_vmware.rw_handles [None req-b10a76eb-ac59-4162-9083-648b4df991d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52ae571d-e901-2583-cb41-ba12bc62eac2/disk-0.vmdk. {{(pid=61957) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 994.670444] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43be014f-4108-4a08-b927-53dfed14fb68 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.687306] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1950a04-7898-4c22-85fc-c6178915bf55 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.704359] env[61957]: DEBUG oslo_vmware.rw_handles [None req-b10a76eb-ac59-4162-9083-648b4df991d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52ae571d-e901-2583-cb41-ba12bc62eac2/disk-0.vmdk is in state: ready. {{(pid=61957) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 994.704359] env[61957]: ERROR oslo_vmware.rw_handles [None req-b10a76eb-ac59-4162-9083-648b4df991d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52ae571d-e901-2583-cb41-ba12bc62eac2/disk-0.vmdk due to incomplete transfer. [ 994.707025] env[61957]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-7ffa5bf1-9422-42b2-b54e-a0e6a94a585f {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.716770] env[61957]: DEBUG oslo_vmware.rw_handles [None req-b10a76eb-ac59-4162-9083-648b4df991d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52ae571d-e901-2583-cb41-ba12bc62eac2/disk-0.vmdk. {{(pid=61957) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 994.716770] env[61957]: DEBUG nova.virt.vmwareapi.images [None req-b10a76eb-ac59-4162-9083-648b4df991d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 98929ed3-d420-45cd-9cde-9738b2ac8251] Uploaded image 74e02d56-c151-4bb1-b998-cd7ba6872fac to the Glance image server {{(pid=61957) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 994.719256] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-b10a76eb-ac59-4162-9083-648b4df991d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 98929ed3-d420-45cd-9cde-9738b2ac8251] Destroying the VM {{(pid=61957) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 994.719401] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-ed79f5ce-3ada-44b7-b8d6-ffc3cae858d3 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.733828] env[61957]: DEBUG oslo_vmware.api [None req-bc1354f0-0ce9-464c-89e1-e1e216546ebe tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52a7ab02-9560-de48-d161-518eb95e7b11, 'name': SearchDatastore_Task, 'duration_secs': 0.011092} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.736033] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-a5bdcddf-64c6-4110-8e30-a016fe726a7a tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] [instance: f65458b2-0db6-4d98-8c82-b27b6945a3fd] Powering off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 994.737693] env[61957]: DEBUG oslo_concurrency.lockutils [None req-bc1354f0-0ce9-464c-89e1-e1e216546ebe tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Releasing lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 994.737952] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-bc1354f0-0ce9-464c-89e1-e1e216546ebe tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: 6509707e-afc6-4c4a-88f9-a471865d510c] Processing image 11c76a2c-f705-470a-ba9d-4657858bab38 {{(pid=61957) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 994.738222] env[61957]: DEBUG oslo_concurrency.lockutils [None req-bc1354f0-0ce9-464c-89e1-e1e216546ebe tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 994.738373] env[61957]: DEBUG oslo_concurrency.lockutils [None req-bc1354f0-0ce9-464c-89e1-e1e216546ebe tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Acquired lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 994.738574] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-bc1354f0-0ce9-464c-89e1-e1e216546ebe tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 994.738844] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a17635f4-86d5-4de2-ae4f-8eb1c55051c0 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.741038] env[61957]: DEBUG oslo_vmware.api [None req-b10a76eb-ac59-4162-9083-648b4df991d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Waiting for the task: (returnval){ [ 994.741038] env[61957]: value = "task-1277908" [ 994.741038] env[61957]: _type = "Task" [ 994.741038] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.741275] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a8a3a000-6e17-4023-8004-9f263a4e71b8 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.753616] env[61957]: DEBUG oslo_vmware.api [None req-b10a76eb-ac59-4162-9083-648b4df991d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1277908, 'name': Destroy_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.756870] env[61957]: DEBUG oslo_vmware.api [None req-a5bdcddf-64c6-4110-8e30-a016fe726a7a tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Waiting for the task: (returnval){ [ 994.756870] env[61957]: value = "task-1277909" [ 994.756870] env[61957]: _type = "Task" [ 994.756870] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.757267] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-bc1354f0-0ce9-464c-89e1-e1e216546ebe tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 994.757544] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-bc1354f0-0ce9-464c-89e1-e1e216546ebe tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61957) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 994.758691] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6bc4bcde-0927-4a42-98d5-83f1064f8ce7 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.768344] env[61957]: DEBUG oslo_vmware.api [None req-bc1354f0-0ce9-464c-89e1-e1e216546ebe tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Waiting for the task: (returnval){ [ 994.768344] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]5228b9f4-44a9-61de-6f42-00c26ce20b87" [ 994.768344] env[61957]: _type = "Task" [ 994.768344] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.775302] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-a5bdcddf-64c6-4110-8e30-a016fe726a7a tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] [instance: f65458b2-0db6-4d98-8c82-b27b6945a3fd] VM already powered off {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 994.775553] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-a5bdcddf-64c6-4110-8e30-a016fe726a7a tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] [instance: f65458b2-0db6-4d98-8c82-b27b6945a3fd] Processing image 11c76a2c-f705-470a-ba9d-4657858bab38 {{(pid=61957) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 994.775800] env[61957]: DEBUG oslo_concurrency.lockutils [None req-a5bdcddf-64c6-4110-8e30-a016fe726a7a tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 994.783090] env[61957]: DEBUG oslo_vmware.api [None req-bc1354f0-0ce9-464c-89e1-e1e216546ebe tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]5228b9f4-44a9-61de-6f42-00c26ce20b87, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.013262] env[61957]: DEBUG oslo_concurrency.lockutils [None req-8cf153c7-e439-4990-9fa2-f6d040c68c3c tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 995.059131] env[61957]: DEBUG oslo_vmware.api [None req-8b32736b-e3a6-485c-a95e-afec23b58d86 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Task: {'id': task-1277906, 'name': CopyVirtualDisk_Task} progress is 21%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.070014] env[61957]: DEBUG oslo_vmware.api [None req-6d13ccf5-9ccd-48d0-a35e-c6df97be9c15 tempest-ServerAddressesTestJSON-2054963291 tempest-ServerAddressesTestJSON-2054963291-project-member] Task: {'id': task-1277907, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.079394} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.070325] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-6d13ccf5-9ccd-48d0-a35e-c6df97be9c15 tempest-ServerAddressesTestJSON-2054963291 tempest-ServerAddressesTestJSON-2054963291-project-member] [instance: fa4c150e-7a85-41a4-8890-9f2deb3b34b1] Extended root virtual disk {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 995.072127] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8487f976-ddf5-43b1-b78a-f4463d9a494d {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.087674] env[61957]: DEBUG nova.network.neutron [-] [instance: d4674fbc-47b1-42d1-aaba-e86d46c51e8f] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 995.104214] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-6d13ccf5-9ccd-48d0-a35e-c6df97be9c15 tempest-ServerAddressesTestJSON-2054963291 tempest-ServerAddressesTestJSON-2054963291-project-member] [instance: fa4c150e-7a85-41a4-8890-9f2deb3b34b1] Reconfiguring VM instance instance-0000005c to attach disk [datastore2] fa4c150e-7a85-41a4-8890-9f2deb3b34b1/fa4c150e-7a85-41a4-8890-9f2deb3b34b1.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 995.104214] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2fcd2c6e-a893-4679-8c15-ffd47ff1dd9f {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.129311] env[61957]: DEBUG oslo_vmware.api [None req-6d13ccf5-9ccd-48d0-a35e-c6df97be9c15 tempest-ServerAddressesTestJSON-2054963291 tempest-ServerAddressesTestJSON-2054963291-project-member] Waiting for the task: (returnval){ [ 995.129311] env[61957]: value = "task-1277910" [ 995.129311] env[61957]: _type = "Task" [ 995.129311] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.141753] env[61957]: DEBUG oslo_vmware.api [None req-6d13ccf5-9ccd-48d0-a35e-c6df97be9c15 tempest-ServerAddressesTestJSON-2054963291 tempest-ServerAddressesTestJSON-2054963291-project-member] Task: {'id': task-1277910, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.256977] env[61957]: DEBUG oslo_vmware.api [None req-b10a76eb-ac59-4162-9083-648b4df991d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1277908, 'name': Destroy_Task} progress is 33%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.284956] env[61957]: DEBUG oslo_vmware.api [None req-bc1354f0-0ce9-464c-89e1-e1e216546ebe tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]5228b9f4-44a9-61de-6f42-00c26ce20b87, 'name': SearchDatastore_Task, 'duration_secs': 0.015548} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.285879] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f43f4911-6ca4-4c2e-ac36-1f5334c9dc51 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.295021] env[61957]: DEBUG oslo_vmware.api [None req-bc1354f0-0ce9-464c-89e1-e1e216546ebe tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Waiting for the task: (returnval){ [ 995.295021] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52f17021-ac7c-39f3-f70f-fdacc29a950a" [ 995.295021] env[61957]: _type = "Task" [ 995.295021] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.306906] env[61957]: DEBUG oslo_vmware.api [None req-bc1354f0-0ce9-464c-89e1-e1e216546ebe tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52f17021-ac7c-39f3-f70f-fdacc29a950a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.374027] env[61957]: DEBUG nova.compute.manager [req-4d8aeeac-a693-429d-9297-53421c6d0c36 req-d66c9d57-5134-41df-b275-3fe19465c680 service nova] [instance: d4674fbc-47b1-42d1-aaba-e86d46c51e8f] Received event network-vif-deleted-7e1da2f5-45f0-4991-bd1f-8d7115b1adbe {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 995.558893] env[61957]: DEBUG oslo_vmware.api [None req-8b32736b-e3a6-485c-a95e-afec23b58d86 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Task: {'id': task-1277906, 'name': CopyVirtualDisk_Task} progress is 40%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.573874] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f713a68-da2a-48ae-9990-1905d135eeba {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.584187] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87ad59f5-7c59-4718-adc6-1d61107619f4 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.619198] env[61957]: INFO nova.compute.manager [-] [instance: d4674fbc-47b1-42d1-aaba-e86d46c51e8f] Took 1.33 seconds to deallocate network for instance. [ 995.621933] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45a8210e-9428-4501-92de-fac913568d33 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.640280] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79c0a2cd-dff1-46ff-a8b1-4930657fe1c1 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.651497] env[61957]: DEBUG oslo_vmware.api [None req-6d13ccf5-9ccd-48d0-a35e-c6df97be9c15 tempest-ServerAddressesTestJSON-2054963291 tempest-ServerAddressesTestJSON-2054963291-project-member] Task: {'id': task-1277910, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.665509] env[61957]: DEBUG nova.compute.provider_tree [None req-ab927ca0-dc5c-48f5-8ef0-a84245f36801 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 995.757798] env[61957]: DEBUG oslo_vmware.api [None req-b10a76eb-ac59-4162-9083-648b4df991d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1277908, 'name': Destroy_Task, 'duration_secs': 0.601171} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.757798] env[61957]: INFO nova.virt.vmwareapi.vm_util [None req-b10a76eb-ac59-4162-9083-648b4df991d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 98929ed3-d420-45cd-9cde-9738b2ac8251] Destroyed the VM [ 995.758057] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-b10a76eb-ac59-4162-9083-648b4df991d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 98929ed3-d420-45cd-9cde-9738b2ac8251] Deleting Snapshot of the VM instance {{(pid=61957) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 995.758244] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-013baf38-68d6-46fe-87da-a2e9c43657bd {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.768154] env[61957]: DEBUG oslo_vmware.api [None req-b10a76eb-ac59-4162-9083-648b4df991d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Waiting for the task: (returnval){ [ 995.768154] env[61957]: value = "task-1277911" [ 995.768154] env[61957]: _type = "Task" [ 995.768154] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.779966] env[61957]: DEBUG oslo_vmware.api [None req-b10a76eb-ac59-4162-9083-648b4df991d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1277911, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.809576] env[61957]: DEBUG oslo_vmware.api [None req-bc1354f0-0ce9-464c-89e1-e1e216546ebe tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52f17021-ac7c-39f3-f70f-fdacc29a950a, 'name': SearchDatastore_Task, 'duration_secs': 0.015014} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.809880] env[61957]: DEBUG oslo_concurrency.lockutils [None req-bc1354f0-0ce9-464c-89e1-e1e216546ebe tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Releasing lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 995.810184] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-bc1354f0-0ce9-464c-89e1-e1e216546ebe tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore1] 6509707e-afc6-4c4a-88f9-a471865d510c/6509707e-afc6-4c4a-88f9-a471865d510c.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 995.810488] env[61957]: DEBUG oslo_concurrency.lockutils [None req-a5bdcddf-64c6-4110-8e30-a016fe726a7a tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Acquired lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 995.810682] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-a5bdcddf-64c6-4110-8e30-a016fe726a7a tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 995.810919] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-577ba9f0-823c-43cc-81c7-67b82feb6a15 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.813360] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-30ca0d17-1b85-4ba7-93ed-c7a8ef37b150 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.828018] env[61957]: DEBUG oslo_vmware.api [None req-bc1354f0-0ce9-464c-89e1-e1e216546ebe tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Waiting for the task: (returnval){ [ 995.828018] env[61957]: value = "task-1277912" [ 995.828018] env[61957]: _type = "Task" [ 995.828018] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.828018] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-a5bdcddf-64c6-4110-8e30-a016fe726a7a tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 995.828018] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-a5bdcddf-64c6-4110-8e30-a016fe726a7a tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61957) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 995.828018] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-76056965-fbe2-43eb-ac04-f0be86e9f311 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.836291] env[61957]: DEBUG oslo_vmware.api [None req-bc1354f0-0ce9-464c-89e1-e1e216546ebe tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': task-1277912, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.837833] env[61957]: DEBUG oslo_vmware.api [None req-a5bdcddf-64c6-4110-8e30-a016fe726a7a tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Waiting for the task: (returnval){ [ 995.837833] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52875c78-832f-1b7c-b961-3f38927f5516" [ 995.837833] env[61957]: _type = "Task" [ 995.837833] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.848986] env[61957]: DEBUG oslo_vmware.api [None req-a5bdcddf-64c6-4110-8e30-a016fe726a7a tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52875c78-832f-1b7c-b961-3f38927f5516, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.064078] env[61957]: DEBUG oslo_vmware.api [None req-8b32736b-e3a6-485c-a95e-afec23b58d86 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Task: {'id': task-1277906, 'name': CopyVirtualDisk_Task} progress is 63%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.138145] env[61957]: DEBUG oslo_concurrency.lockutils [None req-4614d338-72f4-4519-9a7d-0a36e19fbd39 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 996.146242] env[61957]: DEBUG oslo_vmware.api [None req-6d13ccf5-9ccd-48d0-a35e-c6df97be9c15 tempest-ServerAddressesTestJSON-2054963291 tempest-ServerAddressesTestJSON-2054963291-project-member] Task: {'id': task-1277910, 'name': ReconfigVM_Task, 'duration_secs': 0.677857} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.148058] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-6d13ccf5-9ccd-48d0-a35e-c6df97be9c15 tempest-ServerAddressesTestJSON-2054963291 tempest-ServerAddressesTestJSON-2054963291-project-member] [instance: fa4c150e-7a85-41a4-8890-9f2deb3b34b1] Reconfigured VM instance instance-0000005c to attach disk [datastore2] fa4c150e-7a85-41a4-8890-9f2deb3b34b1/fa4c150e-7a85-41a4-8890-9f2deb3b34b1.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 996.148058] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5e91e417-7991-4d46-b8eb-e091433f1d0f {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.162262] env[61957]: DEBUG oslo_vmware.api [None req-6d13ccf5-9ccd-48d0-a35e-c6df97be9c15 tempest-ServerAddressesTestJSON-2054963291 tempest-ServerAddressesTestJSON-2054963291-project-member] Waiting for the task: (returnval){ [ 996.162262] env[61957]: value = "task-1277913" [ 996.162262] env[61957]: _type = "Task" [ 996.162262] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.169397] env[61957]: DEBUG nova.scheduler.client.report [None req-ab927ca0-dc5c-48f5-8ef0-a84245f36801 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 996.180443] env[61957]: DEBUG oslo_vmware.api [None req-6d13ccf5-9ccd-48d0-a35e-c6df97be9c15 tempest-ServerAddressesTestJSON-2054963291 tempest-ServerAddressesTestJSON-2054963291-project-member] Task: {'id': task-1277913, 'name': Rename_Task} progress is 5%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.284677] env[61957]: DEBUG oslo_vmware.api [None req-b10a76eb-ac59-4162-9083-648b4df991d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1277911, 'name': RemoveSnapshot_Task, 'duration_secs': 0.465651} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.284991] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-b10a76eb-ac59-4162-9083-648b4df991d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 98929ed3-d420-45cd-9cde-9738b2ac8251] Deleted Snapshot of the VM instance {{(pid=61957) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 996.285310] env[61957]: DEBUG nova.compute.manager [None req-b10a76eb-ac59-4162-9083-648b4df991d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 98929ed3-d420-45cd-9cde-9738b2ac8251] Checking state {{(pid=61957) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 996.286198] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fed94d6f-c8f5-4c87-adbb-d376929d2b6c {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.335099] env[61957]: DEBUG oslo_vmware.api [None req-bc1354f0-0ce9-464c-89e1-e1e216546ebe tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': task-1277912, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.352178] env[61957]: DEBUG oslo_vmware.api [None req-a5bdcddf-64c6-4110-8e30-a016fe726a7a tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52875c78-832f-1b7c-b961-3f38927f5516, 'name': SearchDatastore_Task, 'duration_secs': 0.01499} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.353220] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2f693bf2-9ead-4f2f-87e1-ed74a4556235 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.362775] env[61957]: DEBUG oslo_vmware.api [None req-a5bdcddf-64c6-4110-8e30-a016fe726a7a tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Waiting for the task: (returnval){ [ 996.362775] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]520ffe02-f77d-59c7-ebbd-22780c91d452" [ 996.362775] env[61957]: _type = "Task" [ 996.362775] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.374533] env[61957]: DEBUG oslo_vmware.api [None req-a5bdcddf-64c6-4110-8e30-a016fe726a7a tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]520ffe02-f77d-59c7-ebbd-22780c91d452, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.561094] env[61957]: DEBUG oslo_vmware.api [None req-8b32736b-e3a6-485c-a95e-afec23b58d86 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Task: {'id': task-1277906, 'name': CopyVirtualDisk_Task} progress is 85%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.672078] env[61957]: DEBUG oslo_vmware.api [None req-6d13ccf5-9ccd-48d0-a35e-c6df97be9c15 tempest-ServerAddressesTestJSON-2054963291 tempest-ServerAddressesTestJSON-2054963291-project-member] Task: {'id': task-1277913, 'name': Rename_Task, 'duration_secs': 0.246489} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.672519] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d13ccf5-9ccd-48d0-a35e-c6df97be9c15 tempest-ServerAddressesTestJSON-2054963291 tempest-ServerAddressesTestJSON-2054963291-project-member] [instance: fa4c150e-7a85-41a4-8890-9f2deb3b34b1] Powering on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 996.672650] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-db1943fa-2b25-42f0-ae32-327d2b3a68c5 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.682109] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ab927ca0-dc5c-48f5-8ef0-a84245f36801 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.326s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 996.682637] env[61957]: DEBUG nova.compute.manager [None req-ab927ca0-dc5c-48f5-8ef0-a84245f36801 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: e2107d5c-8453-4d24-b364-2b5c1d125062] Start building networks asynchronously for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 996.685298] env[61957]: DEBUG oslo_vmware.api [None req-6d13ccf5-9ccd-48d0-a35e-c6df97be9c15 tempest-ServerAddressesTestJSON-2054963291 tempest-ServerAddressesTestJSON-2054963291-project-member] Waiting for the task: (returnval){ [ 996.685298] env[61957]: value = "task-1277914" [ 996.685298] env[61957]: _type = "Task" [ 996.685298] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.685542] env[61957]: DEBUG oslo_concurrency.lockutils [None req-8cf153c7-e439-4990-9fa2-f6d040c68c3c tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.673s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 996.687029] env[61957]: INFO nova.compute.claims [None req-8cf153c7-e439-4990-9fa2-f6d040c68c3c tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 84cc9489-2a11-4510-8ad6-13dc41814856] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 996.708316] env[61957]: DEBUG oslo_vmware.api [None req-6d13ccf5-9ccd-48d0-a35e-c6df97be9c15 tempest-ServerAddressesTestJSON-2054963291 tempest-ServerAddressesTestJSON-2054963291-project-member] Task: {'id': task-1277914, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.801819] env[61957]: INFO nova.compute.manager [None req-b10a76eb-ac59-4162-9083-648b4df991d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 98929ed3-d420-45cd-9cde-9738b2ac8251] Shelve offloading [ 996.803576] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-b10a76eb-ac59-4162-9083-648b4df991d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 98929ed3-d420-45cd-9cde-9738b2ac8251] Powering off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 996.803840] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-085ff86d-386e-4577-83f6-cf0c70cba84d {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.813694] env[61957]: DEBUG oslo_vmware.api [None req-b10a76eb-ac59-4162-9083-648b4df991d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Waiting for the task: (returnval){ [ 996.813694] env[61957]: value = "task-1277915" [ 996.813694] env[61957]: _type = "Task" [ 996.813694] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.828167] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-b10a76eb-ac59-4162-9083-648b4df991d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 98929ed3-d420-45cd-9cde-9738b2ac8251] VM already powered off {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 996.828370] env[61957]: DEBUG nova.compute.manager [None req-b10a76eb-ac59-4162-9083-648b4df991d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 98929ed3-d420-45cd-9cde-9738b2ac8251] Checking state {{(pid=61957) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 996.829168] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61fdc333-bd66-4254-a71e-4ebd4ac86454 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.836403] env[61957]: DEBUG oslo_concurrency.lockutils [None req-b10a76eb-ac59-4162-9083-648b4df991d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Acquiring lock "refresh_cache-98929ed3-d420-45cd-9cde-9738b2ac8251" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 996.836462] env[61957]: DEBUG oslo_concurrency.lockutils [None req-b10a76eb-ac59-4162-9083-648b4df991d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Acquired lock "refresh_cache-98929ed3-d420-45cd-9cde-9738b2ac8251" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 996.836592] env[61957]: DEBUG nova.network.neutron [None req-b10a76eb-ac59-4162-9083-648b4df991d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 98929ed3-d420-45cd-9cde-9738b2ac8251] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 996.840923] env[61957]: DEBUG oslo_vmware.api [None req-bc1354f0-0ce9-464c-89e1-e1e216546ebe tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': task-1277912, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.558503} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.841469] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-bc1354f0-0ce9-464c-89e1-e1e216546ebe tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore1] 6509707e-afc6-4c4a-88f9-a471865d510c/6509707e-afc6-4c4a-88f9-a471865d510c.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 996.841692] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-bc1354f0-0ce9-464c-89e1-e1e216546ebe tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: 6509707e-afc6-4c4a-88f9-a471865d510c] Extending root virtual disk to 1048576 {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 996.842093] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2e41b9e3-050e-4616-b1c9-e4149f4dc687 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.852160] env[61957]: DEBUG oslo_vmware.api [None req-bc1354f0-0ce9-464c-89e1-e1e216546ebe tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Waiting for the task: (returnval){ [ 996.852160] env[61957]: value = "task-1277916" [ 996.852160] env[61957]: _type = "Task" [ 996.852160] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.863218] env[61957]: DEBUG oslo_vmware.api [None req-bc1354f0-0ce9-464c-89e1-e1e216546ebe tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': task-1277916, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.874297] env[61957]: DEBUG oslo_vmware.api [None req-a5bdcddf-64c6-4110-8e30-a016fe726a7a tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]520ffe02-f77d-59c7-ebbd-22780c91d452, 'name': SearchDatastore_Task, 'duration_secs': 0.012622} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.874575] env[61957]: DEBUG oslo_concurrency.lockutils [None req-a5bdcddf-64c6-4110-8e30-a016fe726a7a tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Releasing lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 996.874856] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-a5bdcddf-64c6-4110-8e30-a016fe726a7a tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Copying virtual disk from [datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore1] f65458b2-0db6-4d98-8c82-b27b6945a3fd/11c76a2c-f705-470a-ba9d-4657858bab38-rescue.vmdk. {{(pid=61957) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 996.875151] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-63f35fcb-58cf-4d2e-b6e2-ea76b23b7648 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.882561] env[61957]: DEBUG oslo_vmware.api [None req-a5bdcddf-64c6-4110-8e30-a016fe726a7a tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Waiting for the task: (returnval){ [ 996.882561] env[61957]: value = "task-1277917" [ 996.882561] env[61957]: _type = "Task" [ 996.882561] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.892122] env[61957]: DEBUG oslo_vmware.api [None req-a5bdcddf-64c6-4110-8e30-a016fe726a7a tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Task: {'id': task-1277917, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.973902] env[61957]: DEBUG oslo_concurrency.lockutils [None req-be7ed37f-5147-4a2e-a55e-6d6b5d896537 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Acquiring lock "interface-08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 996.974214] env[61957]: DEBUG oslo_concurrency.lockutils [None req-be7ed37f-5147-4a2e-a55e-6d6b5d896537 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Lock "interface-08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 996.974542] env[61957]: DEBUG nova.objects.instance [None req-be7ed37f-5147-4a2e-a55e-6d6b5d896537 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Lazy-loading 'flavor' on Instance uuid 08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2 {{(pid=61957) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 997.060801] env[61957]: DEBUG oslo_vmware.api [None req-8b32736b-e3a6-485c-a95e-afec23b58d86 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Task: {'id': task-1277906, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.3383} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.060801] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-8b32736b-e3a6-485c-a95e-afec23b58d86 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0018d8e9-1d6e-4ca4-a224-31fc9e9f9d2e/0018d8e9-1d6e-4ca4-a224-31fc9e9f9d2e.vmdk to [datastore2] 736eabcf-70f8-4e1e-9830-270d9971fc9a/736eabcf-70f8-4e1e-9830-270d9971fc9a.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 997.061064] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36a976e0-015f-4922-a529-e305a17b4145 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.084290] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-8b32736b-e3a6-485c-a95e-afec23b58d86 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] [instance: 736eabcf-70f8-4e1e-9830-270d9971fc9a] Reconfiguring VM instance instance-0000004a to attach disk [datastore2] 736eabcf-70f8-4e1e-9830-270d9971fc9a/736eabcf-70f8-4e1e-9830-270d9971fc9a.vmdk or device None with type streamOptimized {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 997.084576] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a71efcc4-3a28-430a-a27c-e5c29aeb12e7 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.105219] env[61957]: DEBUG oslo_vmware.api [None req-8b32736b-e3a6-485c-a95e-afec23b58d86 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Waiting for the task: (returnval){ [ 997.105219] env[61957]: value = "task-1277918" [ 997.105219] env[61957]: _type = "Task" [ 997.105219] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.114614] env[61957]: DEBUG oslo_vmware.api [None req-8b32736b-e3a6-485c-a95e-afec23b58d86 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Task: {'id': task-1277918, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.192429] env[61957]: DEBUG nova.compute.utils [None req-ab927ca0-dc5c-48f5-8ef0-a84245f36801 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Using /dev/sd instead of None {{(pid=61957) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 997.198016] env[61957]: DEBUG nova.compute.manager [None req-ab927ca0-dc5c-48f5-8ef0-a84245f36801 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: e2107d5c-8453-4d24-b364-2b5c1d125062] Allocating IP information in the background. {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 997.198400] env[61957]: DEBUG nova.network.neutron [None req-ab927ca0-dc5c-48f5-8ef0-a84245f36801 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: e2107d5c-8453-4d24-b364-2b5c1d125062] allocate_for_instance() {{(pid=61957) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 997.210982] env[61957]: DEBUG oslo_vmware.api [None req-6d13ccf5-9ccd-48d0-a35e-c6df97be9c15 tempest-ServerAddressesTestJSON-2054963291 tempest-ServerAddressesTestJSON-2054963291-project-member] Task: {'id': task-1277914, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.252770] env[61957]: DEBUG nova.policy [None req-ab927ca0-dc5c-48f5-8ef0-a84245f36801 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '76e1754f96ed438ea0ed91ad337419f1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '975189ebb3cc4cdb9391880f0c9ba6ba', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61957) authorize /opt/stack/nova/nova/policy.py:203}} [ 997.363585] env[61957]: DEBUG oslo_vmware.api [None req-bc1354f0-0ce9-464c-89e1-e1e216546ebe tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': task-1277916, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.07419} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.363979] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-bc1354f0-0ce9-464c-89e1-e1e216546ebe tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: 6509707e-afc6-4c4a-88f9-a471865d510c] Extended root virtual disk {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 997.364845] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0529ef27-86ca-4993-a5e0-9dd11b79b105 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.388793] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-bc1354f0-0ce9-464c-89e1-e1e216546ebe tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: 6509707e-afc6-4c4a-88f9-a471865d510c] Reconfiguring VM instance instance-0000005d to attach disk [datastore1] 6509707e-afc6-4c4a-88f9-a471865d510c/6509707e-afc6-4c4a-88f9-a471865d510c.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 997.389323] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4f6dab63-0cf2-46ae-9cfd-68ee5a1643f9 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.420194] env[61957]: DEBUG oslo_vmware.api [None req-a5bdcddf-64c6-4110-8e30-a016fe726a7a tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Task: {'id': task-1277917, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.421888] env[61957]: DEBUG oslo_vmware.api [None req-bc1354f0-0ce9-464c-89e1-e1e216546ebe tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Waiting for the task: (returnval){ [ 997.421888] env[61957]: value = "task-1277919" [ 997.421888] env[61957]: _type = "Task" [ 997.421888] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.431591] env[61957]: DEBUG oslo_vmware.api [None req-bc1354f0-0ce9-464c-89e1-e1e216546ebe tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': task-1277919, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.478358] env[61957]: DEBUG nova.objects.instance [None req-be7ed37f-5147-4a2e-a55e-6d6b5d896537 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Lazy-loading 'pci_requests' on Instance uuid 08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2 {{(pid=61957) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 997.616914] env[61957]: DEBUG oslo_vmware.api [None req-8b32736b-e3a6-485c-a95e-afec23b58d86 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Task: {'id': task-1277918, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.708169] env[61957]: DEBUG nova.compute.manager [None req-ab927ca0-dc5c-48f5-8ef0-a84245f36801 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: e2107d5c-8453-4d24-b364-2b5c1d125062] Start building block device mappings for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 997.712490] env[61957]: DEBUG oslo_vmware.api [None req-6d13ccf5-9ccd-48d0-a35e-c6df97be9c15 tempest-ServerAddressesTestJSON-2054963291 tempest-ServerAddressesTestJSON-2054963291-project-member] Task: {'id': task-1277914, 'name': PowerOnVM_Task, 'duration_secs': 0.538102} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.713686] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d13ccf5-9ccd-48d0-a35e-c6df97be9c15 tempest-ServerAddressesTestJSON-2054963291 tempest-ServerAddressesTestJSON-2054963291-project-member] [instance: fa4c150e-7a85-41a4-8890-9f2deb3b34b1] Powered on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 997.715300] env[61957]: INFO nova.compute.manager [None req-6d13ccf5-9ccd-48d0-a35e-c6df97be9c15 tempest-ServerAddressesTestJSON-2054963291 tempest-ServerAddressesTestJSON-2054963291-project-member] [instance: fa4c150e-7a85-41a4-8890-9f2deb3b34b1] Took 9.12 seconds to spawn the instance on the hypervisor. [ 997.715300] env[61957]: DEBUG nova.compute.manager [None req-6d13ccf5-9ccd-48d0-a35e-c6df97be9c15 tempest-ServerAddressesTestJSON-2054963291 tempest-ServerAddressesTestJSON-2054963291-project-member] [instance: fa4c150e-7a85-41a4-8890-9f2deb3b34b1] Checking state {{(pid=61957) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 997.715300] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-027d0048-d8d4-422c-b9fd-1627e689d0ca {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.722793] env[61957]: DEBUG nova.network.neutron [None req-b10a76eb-ac59-4162-9083-648b4df991d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 98929ed3-d420-45cd-9cde-9738b2ac8251] Updating instance_info_cache with network_info: [{"id": "bcb69e39-31e7-4ef0-837a-fa6d08bd0449", "address": "fa:16:3e:d4:d9:2a", "network": {"id": "efe08823-51fe-4620-a849-5345e262cc71", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1035436819-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e16f6dce3f0e44fb96de516f17d4c6f5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "667a2e97-c1be-421d-9941-6b84c2629b43", "external-id": "nsx-vlan-transportzone-484", "segmentation_id": 484, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbcb69e39-31", "ovs_interfaceid": "bcb69e39-31e7-4ef0-837a-fa6d08bd0449", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 997.727323] env[61957]: DEBUG nova.network.neutron [None req-ab927ca0-dc5c-48f5-8ef0-a84245f36801 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: e2107d5c-8453-4d24-b364-2b5c1d125062] Successfully created port: f17ed646-b16b-40e9-a98f-a1324721caa1 {{(pid=61957) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 997.902427] env[61957]: DEBUG oslo_vmware.api [None req-a5bdcddf-64c6-4110-8e30-a016fe726a7a tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Task: {'id': task-1277917, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.594573} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.902427] env[61957]: INFO nova.virt.vmwareapi.ds_util [None req-a5bdcddf-64c6-4110-8e30-a016fe726a7a tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Copied virtual disk from [datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore1] f65458b2-0db6-4d98-8c82-b27b6945a3fd/11c76a2c-f705-470a-ba9d-4657858bab38-rescue.vmdk. [ 997.903236] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c33b76f-7dbe-46ba-b5d1-2093b615dff5 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.928960] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-a5bdcddf-64c6-4110-8e30-a016fe726a7a tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] [instance: f65458b2-0db6-4d98-8c82-b27b6945a3fd] Reconfiguring VM instance instance-0000005b to attach disk [datastore1] f65458b2-0db6-4d98-8c82-b27b6945a3fd/11c76a2c-f705-470a-ba9d-4657858bab38-rescue.vmdk or device None with type thin {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 997.930334] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9caf8d3c-eace-4fda-a4e1-fb4db6a397c8 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.944525] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9132e37d-51f1-4bea-9cc8-66996d69d455 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.957417] env[61957]: DEBUG oslo_vmware.api [None req-bc1354f0-0ce9-464c-89e1-e1e216546ebe tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': task-1277919, 'name': ReconfigVM_Task, 'duration_secs': 0.348588} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.959900] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-bc1354f0-0ce9-464c-89e1-e1e216546ebe tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: 6509707e-afc6-4c4a-88f9-a471865d510c] Reconfigured VM instance instance-0000005d to attach disk [datastore1] 6509707e-afc6-4c4a-88f9-a471865d510c/6509707e-afc6-4c4a-88f9-a471865d510c.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 997.960653] env[61957]: DEBUG oslo_vmware.api [None req-a5bdcddf-64c6-4110-8e30-a016fe726a7a tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Waiting for the task: (returnval){ [ 997.960653] env[61957]: value = "task-1277920" [ 997.960653] env[61957]: _type = "Task" [ 997.960653] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.960851] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-030e144e-d441-488c-8e6d-44bb05e80462 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.963133] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0aee8fca-2997-4d9f-9f68-e0191eeccaca {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.977505] env[61957]: DEBUG oslo_vmware.api [None req-a5bdcddf-64c6-4110-8e30-a016fe726a7a tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Task: {'id': task-1277920, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.007679] env[61957]: DEBUG nova.objects.base [None req-be7ed37f-5147-4a2e-a55e-6d6b5d896537 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Object Instance<08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2> lazy-loaded attributes: flavor,pci_requests {{(pid=61957) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 998.008105] env[61957]: DEBUG nova.network.neutron [None req-be7ed37f-5147-4a2e-a55e-6d6b5d896537 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2] allocate_for_instance() {{(pid=61957) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 998.011497] env[61957]: DEBUG oslo_vmware.api [None req-bc1354f0-0ce9-464c-89e1-e1e216546ebe tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Waiting for the task: (returnval){ [ 998.011497] env[61957]: value = "task-1277921" [ 998.011497] env[61957]: _type = "Task" [ 998.011497] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.012724] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-057c79e5-2d56-4a8e-a83b-18b73c726f8d {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.036074] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93f17e07-ef55-4a5f-aff9-edb9dfa6952a {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.042812] env[61957]: DEBUG oslo_vmware.api [None req-bc1354f0-0ce9-464c-89e1-e1e216546ebe tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': task-1277921, 'name': Rename_Task} progress is 14%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.060350] env[61957]: DEBUG nova.compute.provider_tree [None req-8cf153c7-e439-4990-9fa2-f6d040c68c3c tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 998.075937] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ad83dc8d-b942-4b5a-8561-778a1542e9de tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Acquiring lock "9b0053fe-420e-4b92-86a8-ed44f6a6ec49" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 998.076344] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ad83dc8d-b942-4b5a-8561-778a1542e9de tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Lock "9b0053fe-420e-4b92-86a8-ed44f6a6ec49" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.001s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 998.076640] env[61957]: INFO nova.compute.manager [None req-ad83dc8d-b942-4b5a-8561-778a1542e9de tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] [instance: 9b0053fe-420e-4b92-86a8-ed44f6a6ec49] Shelving [ 998.102542] env[61957]: DEBUG oslo_concurrency.lockutils [None req-be7ed37f-5147-4a2e-a55e-6d6b5d896537 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Lock "interface-08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 1.128s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 998.121013] env[61957]: DEBUG oslo_vmware.api [None req-8b32736b-e3a6-485c-a95e-afec23b58d86 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Task: {'id': task-1277918, 'name': ReconfigVM_Task, 'duration_secs': 0.536857} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.121487] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-8b32736b-e3a6-485c-a95e-afec23b58d86 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] [instance: 736eabcf-70f8-4e1e-9830-270d9971fc9a] Reconfigured VM instance instance-0000004a to attach disk [datastore2] 736eabcf-70f8-4e1e-9830-270d9971fc9a/736eabcf-70f8-4e1e-9830-270d9971fc9a.vmdk or device None with type streamOptimized {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 998.122383] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3c8970e3-4be4-4722-945a-077663af2626 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.130920] env[61957]: DEBUG oslo_vmware.api [None req-8b32736b-e3a6-485c-a95e-afec23b58d86 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Waiting for the task: (returnval){ [ 998.130920] env[61957]: value = "task-1277922" [ 998.130920] env[61957]: _type = "Task" [ 998.130920] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.144117] env[61957]: DEBUG oslo_vmware.api [None req-8b32736b-e3a6-485c-a95e-afec23b58d86 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Task: {'id': task-1277922, 'name': Rename_Task} progress is 5%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.233880] env[61957]: DEBUG oslo_concurrency.lockutils [None req-b10a76eb-ac59-4162-9083-648b4df991d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Releasing lock "refresh_cache-98929ed3-d420-45cd-9cde-9738b2ac8251" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 998.253186] env[61957]: INFO nova.compute.manager [None req-6d13ccf5-9ccd-48d0-a35e-c6df97be9c15 tempest-ServerAddressesTestJSON-2054963291 tempest-ServerAddressesTestJSON-2054963291-project-member] [instance: fa4c150e-7a85-41a4-8890-9f2deb3b34b1] Took 15.85 seconds to build instance. [ 998.477235] env[61957]: DEBUG oslo_vmware.api [None req-a5bdcddf-64c6-4110-8e30-a016fe726a7a tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Task: {'id': task-1277920, 'name': ReconfigVM_Task, 'duration_secs': 0.335053} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.477543] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-a5bdcddf-64c6-4110-8e30-a016fe726a7a tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] [instance: f65458b2-0db6-4d98-8c82-b27b6945a3fd] Reconfigured VM instance instance-0000005b to attach disk [datastore1] f65458b2-0db6-4d98-8c82-b27b6945a3fd/11c76a2c-f705-470a-ba9d-4657858bab38-rescue.vmdk or device None with type thin {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 998.478432] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8352a2a8-7ed1-40d0-b04a-9a1915181af3 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.504921] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-64f5ea1d-73ea-44c8-9139-2a8df3247622 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.521566] env[61957]: DEBUG oslo_vmware.api [None req-a5bdcddf-64c6-4110-8e30-a016fe726a7a tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Waiting for the task: (returnval){ [ 998.521566] env[61957]: value = "task-1277923" [ 998.521566] env[61957]: _type = "Task" [ 998.521566] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.532128] env[61957]: DEBUG nova.compute.manager [req-d78ce18a-dc41-4095-bf98-f110427979b6 req-ebf5a601-adf7-4588-a500-f6ecbf6662ba service nova] [instance: 98929ed3-d420-45cd-9cde-9738b2ac8251] Received event network-vif-unplugged-bcb69e39-31e7-4ef0-837a-fa6d08bd0449 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 998.532427] env[61957]: DEBUG oslo_concurrency.lockutils [req-d78ce18a-dc41-4095-bf98-f110427979b6 req-ebf5a601-adf7-4588-a500-f6ecbf6662ba service nova] Acquiring lock "98929ed3-d420-45cd-9cde-9738b2ac8251-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 998.532748] env[61957]: DEBUG oslo_concurrency.lockutils [req-d78ce18a-dc41-4095-bf98-f110427979b6 req-ebf5a601-adf7-4588-a500-f6ecbf6662ba service nova] Lock "98929ed3-d420-45cd-9cde-9738b2ac8251-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 998.532748] env[61957]: DEBUG oslo_concurrency.lockutils [req-d78ce18a-dc41-4095-bf98-f110427979b6 req-ebf5a601-adf7-4588-a500-f6ecbf6662ba service nova] Lock "98929ed3-d420-45cd-9cde-9738b2ac8251-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 998.533060] env[61957]: DEBUG nova.compute.manager [req-d78ce18a-dc41-4095-bf98-f110427979b6 req-ebf5a601-adf7-4588-a500-f6ecbf6662ba service nova] [instance: 98929ed3-d420-45cd-9cde-9738b2ac8251] No waiting events found dispatching network-vif-unplugged-bcb69e39-31e7-4ef0-837a-fa6d08bd0449 {{(pid=61957) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 998.533060] env[61957]: WARNING nova.compute.manager [req-d78ce18a-dc41-4095-bf98-f110427979b6 req-ebf5a601-adf7-4588-a500-f6ecbf6662ba service nova] [instance: 98929ed3-d420-45cd-9cde-9738b2ac8251] Received unexpected event network-vif-unplugged-bcb69e39-31e7-4ef0-837a-fa6d08bd0449 for instance with vm_state shelved and task_state shelving_offloading. [ 998.533636] env[61957]: DEBUG oslo_vmware.api [None req-bc1354f0-0ce9-464c-89e1-e1e216546ebe tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': task-1277921, 'name': Rename_Task, 'duration_secs': 0.244812} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.534259] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-bc1354f0-0ce9-464c-89e1-e1e216546ebe tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: 6509707e-afc6-4c4a-88f9-a471865d510c] Powering on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 998.534498] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-64674f02-32e3-48c9-be0d-dd6971ffb195 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.539518] env[61957]: DEBUG oslo_vmware.api [None req-a5bdcddf-64c6-4110-8e30-a016fe726a7a tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Task: {'id': task-1277923, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.545223] env[61957]: DEBUG oslo_vmware.api [None req-bc1354f0-0ce9-464c-89e1-e1e216546ebe tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Waiting for the task: (returnval){ [ 998.545223] env[61957]: value = "task-1277924" [ 998.545223] env[61957]: _type = "Task" [ 998.545223] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.553634] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-b10a76eb-ac59-4162-9083-648b4df991d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 98929ed3-d420-45cd-9cde-9738b2ac8251] Destroying instance {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 998.554777] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd985565-f836-45f3-848a-e8035a1e0ce3 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.561287] env[61957]: DEBUG oslo_vmware.api [None req-bc1354f0-0ce9-464c-89e1-e1e216546ebe tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': task-1277924, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.566505] env[61957]: DEBUG nova.scheduler.client.report [None req-8cf153c7-e439-4990-9fa2-f6d040c68c3c tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 998.569921] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-b10a76eb-ac59-4162-9083-648b4df991d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 98929ed3-d420-45cd-9cde-9738b2ac8251] Unregistering the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 998.570194] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-774b8b4c-a0ce-4cd1-9b08-4a9cb2f681c2 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.588299] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-ad83dc8d-b942-4b5a-8561-778a1542e9de tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] [instance: 9b0053fe-420e-4b92-86a8-ed44f6a6ec49] Powering off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 998.588596] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-17350ce0-fb56-4d3e-9929-3d8343fa4e3b {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.596726] env[61957]: DEBUG oslo_vmware.api [None req-ad83dc8d-b942-4b5a-8561-778a1542e9de tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Waiting for the task: (returnval){ [ 998.596726] env[61957]: value = "task-1277926" [ 998.596726] env[61957]: _type = "Task" [ 998.596726] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.607427] env[61957]: DEBUG oslo_vmware.api [None req-ad83dc8d-b942-4b5a-8561-778a1542e9de tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Task: {'id': task-1277926, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.647996] env[61957]: DEBUG oslo_vmware.api [None req-8b32736b-e3a6-485c-a95e-afec23b58d86 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Task: {'id': task-1277922, 'name': Rename_Task, 'duration_secs': 0.262788} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.649590] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-8b32736b-e3a6-485c-a95e-afec23b58d86 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] [instance: 736eabcf-70f8-4e1e-9830-270d9971fc9a] Powering on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 998.649960] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-b10a76eb-ac59-4162-9083-648b4df991d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 98929ed3-d420-45cd-9cde-9738b2ac8251] Unregistered the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 998.650227] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-b10a76eb-ac59-4162-9083-648b4df991d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 98929ed3-d420-45cd-9cde-9738b2ac8251] Deleting contents of the VM from datastore datastore1 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 998.650458] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-b10a76eb-ac59-4162-9083-648b4df991d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Deleting the datastore file [datastore1] 98929ed3-d420-45cd-9cde-9738b2ac8251 {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 998.650750] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1c66ed07-ed03-4d16-b24c-7368589be747 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.652793] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-aac61394-0bd0-4672-ac94-8b1cd0662146 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.662093] env[61957]: DEBUG oslo_vmware.api [None req-b10a76eb-ac59-4162-9083-648b4df991d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Waiting for the task: (returnval){ [ 998.662093] env[61957]: value = "task-1277928" [ 998.662093] env[61957]: _type = "Task" [ 998.662093] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.662496] env[61957]: DEBUG oslo_vmware.api [None req-8b32736b-e3a6-485c-a95e-afec23b58d86 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Waiting for the task: (returnval){ [ 998.662496] env[61957]: value = "task-1277927" [ 998.662496] env[61957]: _type = "Task" [ 998.662496] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.676946] env[61957]: DEBUG oslo_vmware.api [None req-b10a76eb-ac59-4162-9083-648b4df991d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1277928, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.679571] env[61957]: DEBUG oslo_vmware.api [None req-8b32736b-e3a6-485c-a95e-afec23b58d86 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Task: {'id': task-1277927, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.721049] env[61957]: DEBUG nova.compute.manager [None req-ab927ca0-dc5c-48f5-8ef0-a84245f36801 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: e2107d5c-8453-4d24-b364-2b5c1d125062] Start spawning the instance on the hypervisor. {{(pid=61957) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 998.751220] env[61957]: DEBUG nova.virt.hardware [None req-ab927ca0-dc5c-48f5-8ef0-a84245f36801 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T17:22:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T17:21:50Z,direct_url=,disk_format='vmdk',id=11c76a2c-f705-470a-ba9d-4657858bab38,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='81d3096d0f094373913fc3dc8f5c3c6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T17:21:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 998.751387] env[61957]: DEBUG nova.virt.hardware [None req-ab927ca0-dc5c-48f5-8ef0-a84245f36801 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Flavor limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 998.751541] env[61957]: DEBUG nova.virt.hardware [None req-ab927ca0-dc5c-48f5-8ef0-a84245f36801 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Image limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 998.751829] env[61957]: DEBUG nova.virt.hardware [None req-ab927ca0-dc5c-48f5-8ef0-a84245f36801 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Flavor pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 998.752177] env[61957]: DEBUG nova.virt.hardware [None req-ab927ca0-dc5c-48f5-8ef0-a84245f36801 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Image pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 998.752296] env[61957]: DEBUG nova.virt.hardware [None req-ab927ca0-dc5c-48f5-8ef0-a84245f36801 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 998.752457] env[61957]: DEBUG nova.virt.hardware [None req-ab927ca0-dc5c-48f5-8ef0-a84245f36801 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 998.752621] env[61957]: DEBUG nova.virt.hardware [None req-ab927ca0-dc5c-48f5-8ef0-a84245f36801 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 998.752803] env[61957]: DEBUG nova.virt.hardware [None req-ab927ca0-dc5c-48f5-8ef0-a84245f36801 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Got 1 possible topologies {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 998.753025] env[61957]: DEBUG nova.virt.hardware [None req-ab927ca0-dc5c-48f5-8ef0-a84245f36801 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 998.753231] env[61957]: DEBUG nova.virt.hardware [None req-ab927ca0-dc5c-48f5-8ef0-a84245f36801 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 998.754196] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10d2405c-13bf-4608-9419-88847533e111 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.757231] env[61957]: DEBUG oslo_concurrency.lockutils [None req-6d13ccf5-9ccd-48d0-a35e-c6df97be9c15 tempest-ServerAddressesTestJSON-2054963291 tempest-ServerAddressesTestJSON-2054963291-project-member] Lock "fa4c150e-7a85-41a4-8890-9f2deb3b34b1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.861s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 998.764501] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7322c82a-85fb-4d6a-9f7c-60bf0738ff21 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.032886] env[61957]: DEBUG oslo_vmware.api [None req-a5bdcddf-64c6-4110-8e30-a016fe726a7a tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Task: {'id': task-1277923, 'name': ReconfigVM_Task, 'duration_secs': 0.203771} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.033185] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-a5bdcddf-64c6-4110-8e30-a016fe726a7a tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] [instance: f65458b2-0db6-4d98-8c82-b27b6945a3fd] Powering on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 999.033448] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0df7fece-61c4-474d-8448-acf1ea66801e {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.041794] env[61957]: DEBUG oslo_vmware.api [None req-a5bdcddf-64c6-4110-8e30-a016fe726a7a tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Waiting for the task: (returnval){ [ 999.041794] env[61957]: value = "task-1277929" [ 999.041794] env[61957]: _type = "Task" [ 999.041794] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.055125] env[61957]: DEBUG oslo_vmware.api [None req-a5bdcddf-64c6-4110-8e30-a016fe726a7a tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Task: {'id': task-1277929, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.061277] env[61957]: DEBUG oslo_vmware.api [None req-bc1354f0-0ce9-464c-89e1-e1e216546ebe tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': task-1277924, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.073411] env[61957]: DEBUG oslo_concurrency.lockutils [None req-8cf153c7-e439-4990-9fa2-f6d040c68c3c tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.388s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 999.074169] env[61957]: DEBUG nova.compute.manager [None req-8cf153c7-e439-4990-9fa2-f6d040c68c3c tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 84cc9489-2a11-4510-8ad6-13dc41814856] Start building networks asynchronously for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 999.077208] env[61957]: DEBUG oslo_concurrency.lockutils [None req-4614d338-72f4-4519-9a7d-0a36e19fbd39 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.939s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 999.077503] env[61957]: DEBUG nova.objects.instance [None req-4614d338-72f4-4519-9a7d-0a36e19fbd39 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Lazy-loading 'resources' on Instance uuid d4674fbc-47b1-42d1-aaba-e86d46c51e8f {{(pid=61957) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 999.107297] env[61957]: DEBUG oslo_vmware.api [None req-ad83dc8d-b942-4b5a-8561-778a1542e9de tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Task: {'id': task-1277926, 'name': PowerOffVM_Task, 'duration_secs': 0.17891} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.107590] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-ad83dc8d-b942-4b5a-8561-778a1542e9de tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] [instance: 9b0053fe-420e-4b92-86a8-ed44f6a6ec49] Powered off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 999.108338] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe29cd2a-4acb-42e4-b040-39822695181a {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.133286] env[61957]: DEBUG oslo_concurrency.lockutils [None req-89f56e18-80ff-4394-b23a-cf5db06d4c7c tempest-ServerAddressesTestJSON-2054963291 tempest-ServerAddressesTestJSON-2054963291-project-member] Acquiring lock "fa4c150e-7a85-41a4-8890-9f2deb3b34b1" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 999.133286] env[61957]: DEBUG oslo_concurrency.lockutils [None req-89f56e18-80ff-4394-b23a-cf5db06d4c7c tempest-ServerAddressesTestJSON-2054963291 tempest-ServerAddressesTestJSON-2054963291-project-member] Lock "fa4c150e-7a85-41a4-8890-9f2deb3b34b1" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 999.133286] env[61957]: DEBUG oslo_concurrency.lockutils [None req-89f56e18-80ff-4394-b23a-cf5db06d4c7c tempest-ServerAddressesTestJSON-2054963291 tempest-ServerAddressesTestJSON-2054963291-project-member] Acquiring lock "fa4c150e-7a85-41a4-8890-9f2deb3b34b1-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 999.133286] env[61957]: DEBUG oslo_concurrency.lockutils [None req-89f56e18-80ff-4394-b23a-cf5db06d4c7c tempest-ServerAddressesTestJSON-2054963291 tempest-ServerAddressesTestJSON-2054963291-project-member] Lock "fa4c150e-7a85-41a4-8890-9f2deb3b34b1-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 999.133458] env[61957]: DEBUG oslo_concurrency.lockutils [None req-89f56e18-80ff-4394-b23a-cf5db06d4c7c tempest-ServerAddressesTestJSON-2054963291 tempest-ServerAddressesTestJSON-2054963291-project-member] Lock "fa4c150e-7a85-41a4-8890-9f2deb3b34b1-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 999.135725] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1cb3fae5-03f6-4343-9d09-66faf1e5f2cd {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.138612] env[61957]: INFO nova.compute.manager [None req-89f56e18-80ff-4394-b23a-cf5db06d4c7c tempest-ServerAddressesTestJSON-2054963291 tempest-ServerAddressesTestJSON-2054963291-project-member] [instance: fa4c150e-7a85-41a4-8890-9f2deb3b34b1] Terminating instance [ 999.140750] env[61957]: DEBUG nova.compute.manager [None req-89f56e18-80ff-4394-b23a-cf5db06d4c7c tempest-ServerAddressesTestJSON-2054963291 tempest-ServerAddressesTestJSON-2054963291-project-member] [instance: fa4c150e-7a85-41a4-8890-9f2deb3b34b1] Start destroying the instance on the hypervisor. {{(pid=61957) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 999.140949] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-89f56e18-80ff-4394-b23a-cf5db06d4c7c tempest-ServerAddressesTestJSON-2054963291 tempest-ServerAddressesTestJSON-2054963291-project-member] [instance: fa4c150e-7a85-41a4-8890-9f2deb3b34b1] Destroying instance {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 999.141976] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fdb0ba5-e360-4ec0-9452-77398acda438 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.154669] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-89f56e18-80ff-4394-b23a-cf5db06d4c7c tempest-ServerAddressesTestJSON-2054963291 tempest-ServerAddressesTestJSON-2054963291-project-member] [instance: fa4c150e-7a85-41a4-8890-9f2deb3b34b1] Powering off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 999.154960] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-574ccdc6-ad10-4587-b959-6e51dfe50a7b {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.162388] env[61957]: DEBUG oslo_vmware.api [None req-89f56e18-80ff-4394-b23a-cf5db06d4c7c tempest-ServerAddressesTestJSON-2054963291 tempest-ServerAddressesTestJSON-2054963291-project-member] Waiting for the task: (returnval){ [ 999.162388] env[61957]: value = "task-1277930" [ 999.162388] env[61957]: _type = "Task" [ 999.162388] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.178614] env[61957]: DEBUG oslo_vmware.api [None req-89f56e18-80ff-4394-b23a-cf5db06d4c7c tempest-ServerAddressesTestJSON-2054963291 tempest-ServerAddressesTestJSON-2054963291-project-member] Task: {'id': task-1277930, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.186173] env[61957]: DEBUG oslo_vmware.api [None req-b10a76eb-ac59-4162-9083-648b4df991d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1277928, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.248073} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.186478] env[61957]: DEBUG oslo_vmware.api [None req-8b32736b-e3a6-485c-a95e-afec23b58d86 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Task: {'id': task-1277927, 'name': PowerOnVM_Task} progress is 94%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.187491] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-b10a76eb-ac59-4162-9083-648b4df991d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Deleted the datastore file {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 999.187688] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-b10a76eb-ac59-4162-9083-648b4df991d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 98929ed3-d420-45cd-9cde-9738b2ac8251] Deleted contents of the VM from datastore datastore1 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 999.187872] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-b10a76eb-ac59-4162-9083-648b4df991d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 98929ed3-d420-45cd-9cde-9738b2ac8251] Instance destroyed {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 999.220897] env[61957]: INFO nova.scheduler.client.report [None req-b10a76eb-ac59-4162-9083-648b4df991d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Deleted allocations for instance 98929ed3-d420-45cd-9cde-9738b2ac8251 [ 999.251176] env[61957]: DEBUG nova.network.neutron [None req-ab927ca0-dc5c-48f5-8ef0-a84245f36801 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: e2107d5c-8453-4d24-b364-2b5c1d125062] Successfully updated port: f17ed646-b16b-40e9-a98f-a1324721caa1 {{(pid=61957) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 999.560161] env[61957]: DEBUG oslo_vmware.api [None req-a5bdcddf-64c6-4110-8e30-a016fe726a7a tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Task: {'id': task-1277929, 'name': PowerOnVM_Task, 'duration_secs': 0.422931} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.560485] env[61957]: DEBUG oslo_vmware.api [None req-bc1354f0-0ce9-464c-89e1-e1e216546ebe tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': task-1277924, 'name': PowerOnVM_Task, 'duration_secs': 0.578357} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.560807] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-a5bdcddf-64c6-4110-8e30-a016fe726a7a tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] [instance: f65458b2-0db6-4d98-8c82-b27b6945a3fd] Powered on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 999.562960] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-bc1354f0-0ce9-464c-89e1-e1e216546ebe tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: 6509707e-afc6-4c4a-88f9-a471865d510c] Powered on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 999.563068] env[61957]: INFO nova.compute.manager [None req-bc1354f0-0ce9-464c-89e1-e1e216546ebe tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: 6509707e-afc6-4c4a-88f9-a471865d510c] Took 8.56 seconds to spawn the instance on the hypervisor. [ 999.563241] env[61957]: DEBUG nova.compute.manager [None req-bc1354f0-0ce9-464c-89e1-e1e216546ebe tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: 6509707e-afc6-4c4a-88f9-a471865d510c] Checking state {{(pid=61957) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 999.564638] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eaa85f1a-4c1b-471a-be57-440afbb5dc0b {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.568404] env[61957]: DEBUG nova.compute.manager [None req-a5bdcddf-64c6-4110-8e30-a016fe726a7a tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] [instance: f65458b2-0db6-4d98-8c82-b27b6945a3fd] Checking state {{(pid=61957) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 999.568827] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17bb95c7-6e50-4325-b9af-4498c192069f {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.581613] env[61957]: DEBUG nova.compute.utils [None req-8cf153c7-e439-4990-9fa2-f6d040c68c3c tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Using /dev/sd instead of None {{(pid=61957) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 999.594752] env[61957]: DEBUG nova.compute.manager [None req-8cf153c7-e439-4990-9fa2-f6d040c68c3c tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 84cc9489-2a11-4510-8ad6-13dc41814856] Allocating IP information in the background. {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 999.594752] env[61957]: DEBUG nova.network.neutron [None req-8cf153c7-e439-4990-9fa2-f6d040c68c3c tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 84cc9489-2a11-4510-8ad6-13dc41814856] allocate_for_instance() {{(pid=61957) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 999.644868] env[61957]: DEBUG nova.policy [None req-8cf153c7-e439-4990-9fa2-f6d040c68c3c tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e2f13689624b483d903ce12ef290db8b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '65fd71bcd1bf41238a9cc3a5d6dd4924', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61957) authorize /opt/stack/nova/nova/policy.py:203}} [ 999.651337] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-ad83dc8d-b942-4b5a-8561-778a1542e9de tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] [instance: 9b0053fe-420e-4b92-86a8-ed44f6a6ec49] Creating Snapshot of the VM instance {{(pid=61957) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 999.651993] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-5473c6d7-a49b-4611-9f6b-d0b6f864639c {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.665912] env[61957]: DEBUG oslo_vmware.api [None req-ad83dc8d-b942-4b5a-8561-778a1542e9de tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Waiting for the task: (returnval){ [ 999.665912] env[61957]: value = "task-1277931" [ 999.665912] env[61957]: _type = "Task" [ 999.665912] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.686584] env[61957]: DEBUG oslo_vmware.api [None req-89f56e18-80ff-4394-b23a-cf5db06d4c7c tempest-ServerAddressesTestJSON-2054963291 tempest-ServerAddressesTestJSON-2054963291-project-member] Task: {'id': task-1277930, 'name': PowerOffVM_Task, 'duration_secs': 0.22277} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.693240] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-89f56e18-80ff-4394-b23a-cf5db06d4c7c tempest-ServerAddressesTestJSON-2054963291 tempest-ServerAddressesTestJSON-2054963291-project-member] [instance: fa4c150e-7a85-41a4-8890-9f2deb3b34b1] Powered off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 999.693468] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-89f56e18-80ff-4394-b23a-cf5db06d4c7c tempest-ServerAddressesTestJSON-2054963291 tempest-ServerAddressesTestJSON-2054963291-project-member] [instance: fa4c150e-7a85-41a4-8890-9f2deb3b34b1] Unregistering the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 999.694157] env[61957]: DEBUG oslo_vmware.api [None req-8b32736b-e3a6-485c-a95e-afec23b58d86 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Task: {'id': task-1277927, 'name': PowerOnVM_Task, 'duration_secs': 0.575473} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.695020] env[61957]: DEBUG oslo_vmware.api [None req-ad83dc8d-b942-4b5a-8561-778a1542e9de tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Task: {'id': task-1277931, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.697293] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-08e9a210-8fe3-49d1-9e6a-a88ecbee2116 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.698945] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-8b32736b-e3a6-485c-a95e-afec23b58d86 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] [instance: 736eabcf-70f8-4e1e-9830-270d9971fc9a] Powered on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 999.726459] env[61957]: DEBUG oslo_concurrency.lockutils [None req-b10a76eb-ac59-4162-9083-648b4df991d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 999.754570] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ab927ca0-dc5c-48f5-8ef0-a84245f36801 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Acquiring lock "refresh_cache-e2107d5c-8453-4d24-b364-2b5c1d125062" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 999.754570] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ab927ca0-dc5c-48f5-8ef0-a84245f36801 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Acquired lock "refresh_cache-e2107d5c-8453-4d24-b364-2b5c1d125062" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 999.754570] env[61957]: DEBUG nova.network.neutron [None req-ab927ca0-dc5c-48f5-8ef0-a84245f36801 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: e2107d5c-8453-4d24-b364-2b5c1d125062] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 999.784559] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-89f56e18-80ff-4394-b23a-cf5db06d4c7c tempest-ServerAddressesTestJSON-2054963291 tempest-ServerAddressesTestJSON-2054963291-project-member] [instance: fa4c150e-7a85-41a4-8890-9f2deb3b34b1] Unregistered the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 999.784559] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-89f56e18-80ff-4394-b23a-cf5db06d4c7c tempest-ServerAddressesTestJSON-2054963291 tempest-ServerAddressesTestJSON-2054963291-project-member] [instance: fa4c150e-7a85-41a4-8890-9f2deb3b34b1] Deleting contents of the VM from datastore datastore2 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 999.784793] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-89f56e18-80ff-4394-b23a-cf5db06d4c7c tempest-ServerAddressesTestJSON-2054963291 tempest-ServerAddressesTestJSON-2054963291-project-member] Deleting the datastore file [datastore2] fa4c150e-7a85-41a4-8890-9f2deb3b34b1 {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 999.785401] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-dd5b4370-07af-4b25-b878-8c16161a2692 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.793752] env[61957]: DEBUG oslo_vmware.api [None req-89f56e18-80ff-4394-b23a-cf5db06d4c7c tempest-ServerAddressesTestJSON-2054963291 tempest-ServerAddressesTestJSON-2054963291-project-member] Waiting for the task: (returnval){ [ 999.793752] env[61957]: value = "task-1277933" [ 999.793752] env[61957]: _type = "Task" [ 999.793752] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.802380] env[61957]: DEBUG nova.compute.manager [None req-8b32736b-e3a6-485c-a95e-afec23b58d86 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] [instance: 736eabcf-70f8-4e1e-9830-270d9971fc9a] Checking state {{(pid=61957) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 999.802780] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7dc596ac-6a20-4c00-a5f9-e534efb80ef5 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.808470] env[61957]: DEBUG oslo_vmware.api [None req-89f56e18-80ff-4394-b23a-cf5db06d4c7c tempest-ServerAddressesTestJSON-2054963291 tempest-ServerAddressesTestJSON-2054963291-project-member] Task: {'id': task-1277933, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.810850] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfd40b2a-07e2-4d5e-b7db-9daf63539395 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.826364] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-221939d9-cbbe-46f5-b3de-b09ef9bee295 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.859156] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0e12f56-a35a-47f2-a3dd-0b11cda4cbb8 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.868130] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20c3d48f-8466-47d8-8e2d-98788afc95fa {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.885108] env[61957]: DEBUG nova.compute.provider_tree [None req-4614d338-72f4-4519-9a7d-0a36e19fbd39 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 999.925899] env[61957]: DEBUG nova.network.neutron [None req-8cf153c7-e439-4990-9fa2-f6d040c68c3c tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 84cc9489-2a11-4510-8ad6-13dc41814856] Successfully created port: b602c88c-d9b2-44bd-ba43-0ba7d976d8ac {{(pid=61957) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1000.083908] env[61957]: DEBUG nova.compute.manager [None req-8cf153c7-e439-4990-9fa2-f6d040c68c3c tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 84cc9489-2a11-4510-8ad6-13dc41814856] Start building block device mappings for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1000.097868] env[61957]: INFO nova.compute.manager [None req-bc1354f0-0ce9-464c-89e1-e1e216546ebe tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: 6509707e-afc6-4c4a-88f9-a471865d510c] Took 16.55 seconds to build instance. [ 1000.105142] env[61957]: DEBUG oslo_concurrency.lockutils [None req-690abcce-5857-41f4-8c3f-00315918c375 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Acquiring lock "interface-08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1000.105412] env[61957]: DEBUG oslo_concurrency.lockutils [None req-690abcce-5857-41f4-8c3f-00315918c375 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Lock "interface-08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1000.106748] env[61957]: DEBUG nova.objects.instance [None req-690abcce-5857-41f4-8c3f-00315918c375 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Lazy-loading 'flavor' on Instance uuid 08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2 {{(pid=61957) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1000.166272] env[61957]: DEBUG oslo_concurrency.lockutils [None req-58e997cd-a82d-4012-a69d-a48696b8b5f1 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Acquiring lock "98929ed3-d420-45cd-9cde-9738b2ac8251" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1000.179499] env[61957]: DEBUG oslo_vmware.api [None req-ad83dc8d-b942-4b5a-8561-778a1542e9de tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Task: {'id': task-1277931, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.293680] env[61957]: DEBUG nova.network.neutron [None req-ab927ca0-dc5c-48f5-8ef0-a84245f36801 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: e2107d5c-8453-4d24-b364-2b5c1d125062] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1000.305320] env[61957]: DEBUG oslo_vmware.api [None req-89f56e18-80ff-4394-b23a-cf5db06d4c7c tempest-ServerAddressesTestJSON-2054963291 tempest-ServerAddressesTestJSON-2054963291-project-member] Task: {'id': task-1277933, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.209456} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.305598] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-89f56e18-80ff-4394-b23a-cf5db06d4c7c tempest-ServerAddressesTestJSON-2054963291 tempest-ServerAddressesTestJSON-2054963291-project-member] Deleted the datastore file {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1000.305787] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-89f56e18-80ff-4394-b23a-cf5db06d4c7c tempest-ServerAddressesTestJSON-2054963291 tempest-ServerAddressesTestJSON-2054963291-project-member] [instance: fa4c150e-7a85-41a4-8890-9f2deb3b34b1] Deleted contents of the VM from datastore datastore2 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1000.305976] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-89f56e18-80ff-4394-b23a-cf5db06d4c7c tempest-ServerAddressesTestJSON-2054963291 tempest-ServerAddressesTestJSON-2054963291-project-member] [instance: fa4c150e-7a85-41a4-8890-9f2deb3b34b1] Instance destroyed {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1000.306175] env[61957]: INFO nova.compute.manager [None req-89f56e18-80ff-4394-b23a-cf5db06d4c7c tempest-ServerAddressesTestJSON-2054963291 tempest-ServerAddressesTestJSON-2054963291-project-member] [instance: fa4c150e-7a85-41a4-8890-9f2deb3b34b1] Took 1.17 seconds to destroy the instance on the hypervisor. [ 1000.306425] env[61957]: DEBUG oslo.service.loopingcall [None req-89f56e18-80ff-4394-b23a-cf5db06d4c7c tempest-ServerAddressesTestJSON-2054963291 tempest-ServerAddressesTestJSON-2054963291-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1000.306652] env[61957]: DEBUG nova.compute.manager [-] [instance: fa4c150e-7a85-41a4-8890-9f2deb3b34b1] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1000.307516] env[61957]: DEBUG nova.network.neutron [-] [instance: fa4c150e-7a85-41a4-8890-9f2deb3b34b1] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1000.327372] env[61957]: DEBUG oslo_concurrency.lockutils [None req-8b32736b-e3a6-485c-a95e-afec23b58d86 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Lock "736eabcf-70f8-4e1e-9830-270d9971fc9a" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 20.201s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1000.387996] env[61957]: DEBUG nova.scheduler.client.report [None req-4614d338-72f4-4519-9a7d-0a36e19fbd39 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1000.573118] env[61957]: DEBUG nova.network.neutron [None req-ab927ca0-dc5c-48f5-8ef0-a84245f36801 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: e2107d5c-8453-4d24-b364-2b5c1d125062] Updating instance_info_cache with network_info: [{"id": "f17ed646-b16b-40e9-a98f-a1324721caa1", "address": "fa:16:3e:a4:aa:08", "network": {"id": "dc7d4c3b-1f91-4937-a2b0-1084e5c478f0", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1637370260-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "975189ebb3cc4cdb9391880f0c9ba6ba", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b71230ae-e879-4384-88ce-fe64c86fce22", "external-id": "nsx-vlan-transportzone-473", "segmentation_id": 473, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf17ed646-b1", "ovs_interfaceid": "f17ed646-b16b-40e9-a98f-a1324721caa1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1000.600107] env[61957]: DEBUG oslo_concurrency.lockutils [None req-bc1354f0-0ce9-464c-89e1-e1e216546ebe tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Lock "6509707e-afc6-4c4a-88f9-a471865d510c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.068s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1000.679935] env[61957]: DEBUG oslo_vmware.api [None req-ad83dc8d-b942-4b5a-8561-778a1542e9de tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Task: {'id': task-1277931, 'name': CreateSnapshot_Task, 'duration_secs': 1.006826} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.680670] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-ad83dc8d-b942-4b5a-8561-778a1542e9de tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] [instance: 9b0053fe-420e-4b92-86a8-ed44f6a6ec49] Created Snapshot of the VM instance {{(pid=61957) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1000.681122] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5f24cb5-d34f-4fc2-bd95-cf87a4e7a04d {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.722684] env[61957]: DEBUG nova.objects.instance [None req-690abcce-5857-41f4-8c3f-00315918c375 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Lazy-loading 'pci_requests' on Instance uuid 08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2 {{(pid=61957) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1000.841223] env[61957]: DEBUG nova.compute.manager [req-c1321c9d-d08a-4826-8061-e5ec571ad0d6 req-c94d95d3-cc5f-4aab-86de-0688851e28bf service nova] [instance: 98929ed3-d420-45cd-9cde-9738b2ac8251] Received event network-changed-bcb69e39-31e7-4ef0-837a-fa6d08bd0449 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1000.841507] env[61957]: DEBUG nova.compute.manager [req-c1321c9d-d08a-4826-8061-e5ec571ad0d6 req-c94d95d3-cc5f-4aab-86de-0688851e28bf service nova] [instance: 98929ed3-d420-45cd-9cde-9738b2ac8251] Refreshing instance network info cache due to event network-changed-bcb69e39-31e7-4ef0-837a-fa6d08bd0449. {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1000.841705] env[61957]: DEBUG oslo_concurrency.lockutils [req-c1321c9d-d08a-4826-8061-e5ec571ad0d6 req-c94d95d3-cc5f-4aab-86de-0688851e28bf service nova] Acquiring lock "refresh_cache-98929ed3-d420-45cd-9cde-9738b2ac8251" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1000.841935] env[61957]: DEBUG oslo_concurrency.lockutils [req-c1321c9d-d08a-4826-8061-e5ec571ad0d6 req-c94d95d3-cc5f-4aab-86de-0688851e28bf service nova] Acquired lock "refresh_cache-98929ed3-d420-45cd-9cde-9738b2ac8251" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1000.842837] env[61957]: DEBUG nova.network.neutron [req-c1321c9d-d08a-4826-8061-e5ec571ad0d6 req-c94d95d3-cc5f-4aab-86de-0688851e28bf service nova] [instance: 98929ed3-d420-45cd-9cde-9738b2ac8251] Refreshing network info cache for port bcb69e39-31e7-4ef0-837a-fa6d08bd0449 {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1000.870176] env[61957]: INFO nova.compute.manager [None req-c17a70a2-b114-4298-9696-2f9ae04e6d89 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] [instance: f65458b2-0db6-4d98-8c82-b27b6945a3fd] Unrescuing [ 1000.870490] env[61957]: DEBUG oslo_concurrency.lockutils [None req-c17a70a2-b114-4298-9696-2f9ae04e6d89 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Acquiring lock "refresh_cache-f65458b2-0db6-4d98-8c82-b27b6945a3fd" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1000.870707] env[61957]: DEBUG oslo_concurrency.lockutils [None req-c17a70a2-b114-4298-9696-2f9ae04e6d89 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Acquired lock "refresh_cache-f65458b2-0db6-4d98-8c82-b27b6945a3fd" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1000.870818] env[61957]: DEBUG nova.network.neutron [None req-c17a70a2-b114-4298-9696-2f9ae04e6d89 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] [instance: f65458b2-0db6-4d98-8c82-b27b6945a3fd] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1000.893395] env[61957]: DEBUG oslo_concurrency.lockutils [None req-4614d338-72f4-4519-9a7d-0a36e19fbd39 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.816s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1000.895548] env[61957]: DEBUG oslo_concurrency.lockutils [None req-b10a76eb-ac59-4162-9083-648b4df991d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.169s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1000.895786] env[61957]: DEBUG nova.objects.instance [None req-b10a76eb-ac59-4162-9083-648b4df991d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Lazy-loading 'resources' on Instance uuid 98929ed3-d420-45cd-9cde-9738b2ac8251 {{(pid=61957) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1000.914234] env[61957]: INFO nova.scheduler.client.report [None req-4614d338-72f4-4519-9a7d-0a36e19fbd39 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Deleted allocations for instance d4674fbc-47b1-42d1-aaba-e86d46c51e8f [ 1001.075943] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ab927ca0-dc5c-48f5-8ef0-a84245f36801 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Releasing lock "refresh_cache-e2107d5c-8453-4d24-b364-2b5c1d125062" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1001.076147] env[61957]: DEBUG nova.compute.manager [None req-ab927ca0-dc5c-48f5-8ef0-a84245f36801 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: e2107d5c-8453-4d24-b364-2b5c1d125062] Instance network_info: |[{"id": "f17ed646-b16b-40e9-a98f-a1324721caa1", "address": "fa:16:3e:a4:aa:08", "network": {"id": "dc7d4c3b-1f91-4937-a2b0-1084e5c478f0", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1637370260-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "975189ebb3cc4cdb9391880f0c9ba6ba", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b71230ae-e879-4384-88ce-fe64c86fce22", "external-id": "nsx-vlan-transportzone-473", "segmentation_id": 473, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf17ed646-b1", "ovs_interfaceid": "f17ed646-b16b-40e9-a98f-a1324721caa1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1001.076600] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-ab927ca0-dc5c-48f5-8ef0-a84245f36801 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: e2107d5c-8453-4d24-b364-2b5c1d125062] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a4:aa:08', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'b71230ae-e879-4384-88ce-fe64c86fce22', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f17ed646-b16b-40e9-a98f-a1324721caa1', 'vif_model': 'vmxnet3'}] {{(pid=61957) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1001.084284] env[61957]: DEBUG oslo.service.loopingcall [None req-ab927ca0-dc5c-48f5-8ef0-a84245f36801 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1001.084505] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e2107d5c-8453-4d24-b364-2b5c1d125062] Creating VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1001.084727] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-405a139b-e71d-4e0c-b0e3-e0c2b398d336 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.099819] env[61957]: DEBUG nova.compute.manager [None req-8cf153c7-e439-4990-9fa2-f6d040c68c3c tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 84cc9489-2a11-4510-8ad6-13dc41814856] Start spawning the instance on the hypervisor. {{(pid=61957) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1001.109567] env[61957]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1001.109567] env[61957]: value = "task-1277934" [ 1001.109567] env[61957]: _type = "Task" [ 1001.109567] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.120250] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277934, 'name': CreateVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.129274] env[61957]: DEBUG nova.virt.hardware [None req-8cf153c7-e439-4990-9fa2-f6d040c68c3c tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T17:22:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T17:21:50Z,direct_url=,disk_format='vmdk',id=11c76a2c-f705-470a-ba9d-4657858bab38,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='81d3096d0f094373913fc3dc8f5c3c6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T17:21:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1001.129515] env[61957]: DEBUG nova.virt.hardware [None req-8cf153c7-e439-4990-9fa2-f6d040c68c3c tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Flavor limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1001.129678] env[61957]: DEBUG nova.virt.hardware [None req-8cf153c7-e439-4990-9fa2-f6d040c68c3c tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Image limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1001.129863] env[61957]: DEBUG nova.virt.hardware [None req-8cf153c7-e439-4990-9fa2-f6d040c68c3c tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Flavor pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1001.130015] env[61957]: DEBUG nova.virt.hardware [None req-8cf153c7-e439-4990-9fa2-f6d040c68c3c tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Image pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1001.130181] env[61957]: DEBUG nova.virt.hardware [None req-8cf153c7-e439-4990-9fa2-f6d040c68c3c tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1001.130393] env[61957]: DEBUG nova.virt.hardware [None req-8cf153c7-e439-4990-9fa2-f6d040c68c3c tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1001.130558] env[61957]: DEBUG nova.virt.hardware [None req-8cf153c7-e439-4990-9fa2-f6d040c68c3c tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1001.130726] env[61957]: DEBUG nova.virt.hardware [None req-8cf153c7-e439-4990-9fa2-f6d040c68c3c tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Got 1 possible topologies {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1001.130890] env[61957]: DEBUG nova.virt.hardware [None req-8cf153c7-e439-4990-9fa2-f6d040c68c3c tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1001.131078] env[61957]: DEBUG nova.virt.hardware [None req-8cf153c7-e439-4990-9fa2-f6d040c68c3c tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1001.131933] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-191aa76b-f81b-4f5a-8fc6-24b5296e0c80 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.140299] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2c7860b-bd7d-4082-a728-2a93e17a99bf {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.218815] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-ad83dc8d-b942-4b5a-8561-778a1542e9de tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] [instance: 9b0053fe-420e-4b92-86a8-ed44f6a6ec49] Creating linked-clone VM from snapshot {{(pid=61957) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1001.219189] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-6ad6acb7-8990-4119-8d4b-71d3e720d384 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.225205] env[61957]: DEBUG nova.objects.base [None req-690abcce-5857-41f4-8c3f-00315918c375 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Object Instance<08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2> lazy-loaded attributes: flavor,pci_requests {{(pid=61957) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1001.225425] env[61957]: DEBUG nova.network.neutron [None req-690abcce-5857-41f4-8c3f-00315918c375 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2] allocate_for_instance() {{(pid=61957) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1001.228855] env[61957]: DEBUG oslo_vmware.api [None req-ad83dc8d-b942-4b5a-8561-778a1542e9de tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Waiting for the task: (returnval){ [ 1001.228855] env[61957]: value = "task-1277935" [ 1001.228855] env[61957]: _type = "Task" [ 1001.228855] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.238970] env[61957]: DEBUG oslo_vmware.api [None req-ad83dc8d-b942-4b5a-8561-778a1542e9de tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Task: {'id': task-1277935, 'name': CloneVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.245089] env[61957]: DEBUG nova.network.neutron [-] [instance: fa4c150e-7a85-41a4-8890-9f2deb3b34b1] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1001.309058] env[61957]: DEBUG nova.policy [None req-690abcce-5857-41f4-8c3f-00315918c375 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6fd1c19dc3a44212ada44445e0919106', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8713aa35bcb24b86ad0b58ca9fc991ba', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61957) authorize /opt/stack/nova/nova/policy.py:203}} [ 1001.398754] env[61957]: DEBUG nova.objects.instance [None req-b10a76eb-ac59-4162-9083-648b4df991d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Lazy-loading 'numa_topology' on Instance uuid 98929ed3-d420-45cd-9cde-9738b2ac8251 {{(pid=61957) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1001.423947] env[61957]: DEBUG oslo_concurrency.lockutils [None req-4614d338-72f4-4519-9a7d-0a36e19fbd39 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Lock "d4674fbc-47b1-42d1-aaba-e86d46c51e8f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.850s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1001.622416] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277934, 'name': CreateVM_Task, 'duration_secs': 0.380418} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.622619] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e2107d5c-8453-4d24-b364-2b5c1d125062] Created VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1001.623737] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ab927ca0-dc5c-48f5-8ef0-a84245f36801 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1001.624114] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ab927ca0-dc5c-48f5-8ef0-a84245f36801 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1001.624478] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ab927ca0-dc5c-48f5-8ef0-a84245f36801 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1001.624895] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-88748a69-3aca-42d4-a21d-50e07449a276 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.631725] env[61957]: DEBUG oslo_vmware.api [None req-ab927ca0-dc5c-48f5-8ef0-a84245f36801 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Waiting for the task: (returnval){ [ 1001.631725] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52b83fdd-8573-ad66-edba-44970866917c" [ 1001.631725] env[61957]: _type = "Task" [ 1001.631725] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.647781] env[61957]: DEBUG oslo_vmware.api [None req-ab927ca0-dc5c-48f5-8ef0-a84245f36801 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52b83fdd-8573-ad66-edba-44970866917c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.742585] env[61957]: DEBUG oslo_vmware.api [None req-ad83dc8d-b942-4b5a-8561-778a1542e9de tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Task: {'id': task-1277935, 'name': CloneVM_Task} progress is 94%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.747390] env[61957]: INFO nova.compute.manager [-] [instance: fa4c150e-7a85-41a4-8890-9f2deb3b34b1] Took 1.44 seconds to deallocate network for instance. [ 1001.904360] env[61957]: DEBUG nova.objects.base [None req-b10a76eb-ac59-4162-9083-648b4df991d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Object Instance<98929ed3-d420-45cd-9cde-9738b2ac8251> lazy-loaded attributes: resources,numa_topology {{(pid=61957) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1002.013515] env[61957]: DEBUG nova.compute.manager [None req-0b16b4fe-c18a-4e4e-8833-61daba0fafce tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: 6509707e-afc6-4c4a-88f9-a471865d510c] Stashing vm_state: active {{(pid=61957) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1002.115215] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d47410b6-2d7f-445e-a7e6-df46b780744b {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.125266] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91625f9f-340c-43b9-85f9-f9ecba4687f6 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.160365] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcb30a16-71ca-468d-92c8-7948a2f243a9 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.171467] env[61957]: DEBUG oslo_vmware.api [None req-ab927ca0-dc5c-48f5-8ef0-a84245f36801 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52b83fdd-8573-ad66-edba-44970866917c, 'name': SearchDatastore_Task, 'duration_secs': 0.012748} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.175139] env[61957]: DEBUG nova.network.neutron [None req-c17a70a2-b114-4298-9696-2f9ae04e6d89 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] [instance: f65458b2-0db6-4d98-8c82-b27b6945a3fd] Updating instance_info_cache with network_info: [{"id": "bb715e7d-14c5-418b-b3ee-1ac91aba709d", "address": "fa:16:3e:fd:f1:f4", "network": {"id": "d55c72c6-107d-4965-b4e1-d18383813953", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-193304765-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.2", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "7758bdd10505455b8cecb8129d70ebd5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c24464bb-bb6b-43a2-bdcd-8086ad1a307f", "external-id": "nsx-vlan-transportzone-781", "segmentation_id": 781, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbb715e7d-14", "ovs_interfaceid": "bb715e7d-14c5-418b-b3ee-1ac91aba709d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1002.176644] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ab927ca0-dc5c-48f5-8ef0-a84245f36801 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1002.176961] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-ab927ca0-dc5c-48f5-8ef0-a84245f36801 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: e2107d5c-8453-4d24-b364-2b5c1d125062] Processing image 11c76a2c-f705-470a-ba9d-4657858bab38 {{(pid=61957) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1002.177297] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ab927ca0-dc5c-48f5-8ef0-a84245f36801 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1002.177445] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ab927ca0-dc5c-48f5-8ef0-a84245f36801 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1002.177661] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-ab927ca0-dc5c-48f5-8ef0-a84245f36801 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1002.178252] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-be2800f6-196e-4ed6-bc07-0f248b5d4546 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.182892] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ffbffe2-b736-4c75-9329-73c42f670d65 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.189521] env[61957]: DEBUG nova.network.neutron [req-c1321c9d-d08a-4826-8061-e5ec571ad0d6 req-c94d95d3-cc5f-4aab-86de-0688851e28bf service nova] [instance: 98929ed3-d420-45cd-9cde-9738b2ac8251] Updated VIF entry in instance network info cache for port bcb69e39-31e7-4ef0-837a-fa6d08bd0449. {{(pid=61957) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1002.189905] env[61957]: DEBUG nova.network.neutron [req-c1321c9d-d08a-4826-8061-e5ec571ad0d6 req-c94d95d3-cc5f-4aab-86de-0688851e28bf service nova] [instance: 98929ed3-d420-45cd-9cde-9738b2ac8251] Updating instance_info_cache with network_info: [{"id": "bcb69e39-31e7-4ef0-837a-fa6d08bd0449", "address": "fa:16:3e:d4:d9:2a", "network": {"id": "efe08823-51fe-4620-a849-5345e262cc71", "bridge": null, "label": "tempest-DeleteServersTestJSON-1035436819-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e16f6dce3f0e44fb96de516f17d4c6f5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tapbcb69e39-31", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1002.206306] env[61957]: DEBUG nova.compute.provider_tree [None req-b10a76eb-ac59-4162-9083-648b4df991d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1002.209169] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-ab927ca0-dc5c-48f5-8ef0-a84245f36801 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1002.209413] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-ab927ca0-dc5c-48f5-8ef0-a84245f36801 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61957) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1002.210378] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ac9a41ba-7edc-4501-9a82-c9ba11d5361b {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.217468] env[61957]: DEBUG oslo_vmware.api [None req-ab927ca0-dc5c-48f5-8ef0-a84245f36801 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Waiting for the task: (returnval){ [ 1002.217468] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52f13934-bdc4-49e1-0151-ef650096e717" [ 1002.217468] env[61957]: _type = "Task" [ 1002.217468] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.229665] env[61957]: DEBUG oslo_vmware.api [None req-ab927ca0-dc5c-48f5-8ef0-a84245f36801 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52f13934-bdc4-49e1-0151-ef650096e717, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.240168] env[61957]: DEBUG oslo_vmware.api [None req-ad83dc8d-b942-4b5a-8561-778a1542e9de tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Task: {'id': task-1277935, 'name': CloneVM_Task} progress is 95%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.254721] env[61957]: DEBUG oslo_concurrency.lockutils [None req-89f56e18-80ff-4394-b23a-cf5db06d4c7c tempest-ServerAddressesTestJSON-2054963291 tempest-ServerAddressesTestJSON-2054963291-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1002.274603] env[61957]: DEBUG nova.network.neutron [None req-8cf153c7-e439-4990-9fa2-f6d040c68c3c tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 84cc9489-2a11-4510-8ad6-13dc41814856] Successfully updated port: b602c88c-d9b2-44bd-ba43-0ba7d976d8ac {{(pid=61957) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1002.312430] env[61957]: DEBUG nova.network.neutron [None req-690abcce-5857-41f4-8c3f-00315918c375 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2] Successfully created port: 6a70445e-1266-453e-a6ff-50e8cbdcfd08 {{(pid=61957) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1002.535175] env[61957]: DEBUG oslo_concurrency.lockutils [None req-0b16b4fe-c18a-4e4e-8833-61daba0fafce tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1002.678971] env[61957]: DEBUG oslo_concurrency.lockutils [None req-c17a70a2-b114-4298-9696-2f9ae04e6d89 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Releasing lock "refresh_cache-f65458b2-0db6-4d98-8c82-b27b6945a3fd" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1002.679983] env[61957]: DEBUG nova.objects.instance [None req-c17a70a2-b114-4298-9696-2f9ae04e6d89 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Lazy-loading 'flavor' on Instance uuid f65458b2-0db6-4d98-8c82-b27b6945a3fd {{(pid=61957) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1002.693235] env[61957]: DEBUG oslo_concurrency.lockutils [req-c1321c9d-d08a-4826-8061-e5ec571ad0d6 req-c94d95d3-cc5f-4aab-86de-0688851e28bf service nova] Releasing lock "refresh_cache-98929ed3-d420-45cd-9cde-9738b2ac8251" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1002.693402] env[61957]: DEBUG nova.compute.manager [req-c1321c9d-d08a-4826-8061-e5ec571ad0d6 req-c94d95d3-cc5f-4aab-86de-0688851e28bf service nova] [instance: e2107d5c-8453-4d24-b364-2b5c1d125062] Received event network-vif-plugged-f17ed646-b16b-40e9-a98f-a1324721caa1 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1002.693604] env[61957]: DEBUG oslo_concurrency.lockutils [req-c1321c9d-d08a-4826-8061-e5ec571ad0d6 req-c94d95d3-cc5f-4aab-86de-0688851e28bf service nova] Acquiring lock "e2107d5c-8453-4d24-b364-2b5c1d125062-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1002.693814] env[61957]: DEBUG oslo_concurrency.lockutils [req-c1321c9d-d08a-4826-8061-e5ec571ad0d6 req-c94d95d3-cc5f-4aab-86de-0688851e28bf service nova] Lock "e2107d5c-8453-4d24-b364-2b5c1d125062-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1002.693980] env[61957]: DEBUG oslo_concurrency.lockutils [req-c1321c9d-d08a-4826-8061-e5ec571ad0d6 req-c94d95d3-cc5f-4aab-86de-0688851e28bf service nova] Lock "e2107d5c-8453-4d24-b364-2b5c1d125062-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1002.694191] env[61957]: DEBUG nova.compute.manager [req-c1321c9d-d08a-4826-8061-e5ec571ad0d6 req-c94d95d3-cc5f-4aab-86de-0688851e28bf service nova] [instance: e2107d5c-8453-4d24-b364-2b5c1d125062] No waiting events found dispatching network-vif-plugged-f17ed646-b16b-40e9-a98f-a1324721caa1 {{(pid=61957) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1002.694385] env[61957]: WARNING nova.compute.manager [req-c1321c9d-d08a-4826-8061-e5ec571ad0d6 req-c94d95d3-cc5f-4aab-86de-0688851e28bf service nova] [instance: e2107d5c-8453-4d24-b364-2b5c1d125062] Received unexpected event network-vif-plugged-f17ed646-b16b-40e9-a98f-a1324721caa1 for instance with vm_state building and task_state spawning. [ 1002.694555] env[61957]: DEBUG nova.compute.manager [req-c1321c9d-d08a-4826-8061-e5ec571ad0d6 req-c94d95d3-cc5f-4aab-86de-0688851e28bf service nova] [instance: e2107d5c-8453-4d24-b364-2b5c1d125062] Received event network-changed-f17ed646-b16b-40e9-a98f-a1324721caa1 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1002.694715] env[61957]: DEBUG nova.compute.manager [req-c1321c9d-d08a-4826-8061-e5ec571ad0d6 req-c94d95d3-cc5f-4aab-86de-0688851e28bf service nova] [instance: e2107d5c-8453-4d24-b364-2b5c1d125062] Refreshing instance network info cache due to event network-changed-f17ed646-b16b-40e9-a98f-a1324721caa1. {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1002.694908] env[61957]: DEBUG oslo_concurrency.lockutils [req-c1321c9d-d08a-4826-8061-e5ec571ad0d6 req-c94d95d3-cc5f-4aab-86de-0688851e28bf service nova] Acquiring lock "refresh_cache-e2107d5c-8453-4d24-b364-2b5c1d125062" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1002.695058] env[61957]: DEBUG oslo_concurrency.lockutils [req-c1321c9d-d08a-4826-8061-e5ec571ad0d6 req-c94d95d3-cc5f-4aab-86de-0688851e28bf service nova] Acquired lock "refresh_cache-e2107d5c-8453-4d24-b364-2b5c1d125062" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1002.695222] env[61957]: DEBUG nova.network.neutron [req-c1321c9d-d08a-4826-8061-e5ec571ad0d6 req-c94d95d3-cc5f-4aab-86de-0688851e28bf service nova] [instance: e2107d5c-8453-4d24-b364-2b5c1d125062] Refreshing network info cache for port f17ed646-b16b-40e9-a98f-a1324721caa1 {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1002.711278] env[61957]: DEBUG nova.scheduler.client.report [None req-b10a76eb-ac59-4162-9083-648b4df991d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1002.728306] env[61957]: DEBUG oslo_vmware.api [None req-ab927ca0-dc5c-48f5-8ef0-a84245f36801 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52f13934-bdc4-49e1-0151-ef650096e717, 'name': SearchDatastore_Task, 'duration_secs': 0.016536} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.729673] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1ade28bd-fb91-4e0c-883c-c4bf1fb39016 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.739212] env[61957]: DEBUG oslo_vmware.api [None req-ab927ca0-dc5c-48f5-8ef0-a84245f36801 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Waiting for the task: (returnval){ [ 1002.739212] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52f572e8-3d16-ed93-241c-36428a5cee15" [ 1002.739212] env[61957]: _type = "Task" [ 1002.739212] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.742576] env[61957]: DEBUG oslo_vmware.api [None req-ad83dc8d-b942-4b5a-8561-778a1542e9de tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Task: {'id': task-1277935, 'name': CloneVM_Task, 'duration_secs': 1.289252} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.745571] env[61957]: INFO nova.virt.vmwareapi.vmops [None req-ad83dc8d-b942-4b5a-8561-778a1542e9de tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] [instance: 9b0053fe-420e-4b92-86a8-ed44f6a6ec49] Created linked-clone VM from snapshot [ 1002.746265] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc8a740a-9f49-4a7c-a6e8-bfca3dc124bb {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.753863] env[61957]: DEBUG nova.virt.vmwareapi.images [None req-ad83dc8d-b942-4b5a-8561-778a1542e9de tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] [instance: 9b0053fe-420e-4b92-86a8-ed44f6a6ec49] Uploading image f04c1641-6ff1-4b2c-9748-9cf176307db0 {{(pid=61957) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 1002.758642] env[61957]: DEBUG oslo_vmware.api [None req-ab927ca0-dc5c-48f5-8ef0-a84245f36801 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52f572e8-3d16-ed93-241c-36428a5cee15, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.777025] env[61957]: DEBUG oslo_concurrency.lockutils [None req-8cf153c7-e439-4990-9fa2-f6d040c68c3c tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Acquiring lock "refresh_cache-84cc9489-2a11-4510-8ad6-13dc41814856" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1002.777025] env[61957]: DEBUG oslo_concurrency.lockutils [None req-8cf153c7-e439-4990-9fa2-f6d040c68c3c tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Acquired lock "refresh_cache-84cc9489-2a11-4510-8ad6-13dc41814856" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1002.777025] env[61957]: DEBUG nova.network.neutron [None req-8cf153c7-e439-4990-9fa2-f6d040c68c3c tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 84cc9489-2a11-4510-8ad6-13dc41814856] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1002.786925] env[61957]: DEBUG oslo_vmware.rw_handles [None req-ad83dc8d-b942-4b5a-8561-778a1542e9de tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1002.786925] env[61957]: value = "vm-274591" [ 1002.786925] env[61957]: _type = "VirtualMachine" [ 1002.786925] env[61957]: }. {{(pid=61957) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1002.787458] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-729c5b50-8f9e-4b2d-b209-c3ee1a0e89a5 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.796268] env[61957]: DEBUG oslo_vmware.rw_handles [None req-ad83dc8d-b942-4b5a-8561-778a1542e9de tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Lease: (returnval){ [ 1002.796268] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]520e133f-02c6-2431-511f-d1dc6cf59e70" [ 1002.796268] env[61957]: _type = "HttpNfcLease" [ 1002.796268] env[61957]: } obtained for exporting VM: (result){ [ 1002.796268] env[61957]: value = "vm-274591" [ 1002.796268] env[61957]: _type = "VirtualMachine" [ 1002.796268] env[61957]: }. {{(pid=61957) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1002.796578] env[61957]: DEBUG oslo_vmware.api [None req-ad83dc8d-b942-4b5a-8561-778a1542e9de tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Waiting for the lease: (returnval){ [ 1002.796578] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]520e133f-02c6-2431-511f-d1dc6cf59e70" [ 1002.796578] env[61957]: _type = "HttpNfcLease" [ 1002.796578] env[61957]: } to be ready. {{(pid=61957) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1002.803652] env[61957]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1002.803652] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]520e133f-02c6-2431-511f-d1dc6cf59e70" [ 1002.803652] env[61957]: _type = "HttpNfcLease" [ 1002.803652] env[61957]: } is initializing. {{(pid=61957) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1002.910134] env[61957]: DEBUG nova.compute.manager [req-2313b73d-1368-497f-ade4-572589a0e0cd req-12788b97-57b7-4f81-961f-55699b647171 service nova] [instance: fa4c150e-7a85-41a4-8890-9f2deb3b34b1] Received event network-vif-deleted-4b2ccba5-48f1-44f1-8746-3792d3695e58 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1002.910398] env[61957]: DEBUG nova.compute.manager [req-2313b73d-1368-497f-ade4-572589a0e0cd req-12788b97-57b7-4f81-961f-55699b647171 service nova] [instance: 84cc9489-2a11-4510-8ad6-13dc41814856] Received event network-vif-plugged-b602c88c-d9b2-44bd-ba43-0ba7d976d8ac {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1002.910568] env[61957]: DEBUG oslo_concurrency.lockutils [req-2313b73d-1368-497f-ade4-572589a0e0cd req-12788b97-57b7-4f81-961f-55699b647171 service nova] Acquiring lock "84cc9489-2a11-4510-8ad6-13dc41814856-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1002.910791] env[61957]: DEBUG oslo_concurrency.lockutils [req-2313b73d-1368-497f-ade4-572589a0e0cd req-12788b97-57b7-4f81-961f-55699b647171 service nova] Lock "84cc9489-2a11-4510-8ad6-13dc41814856-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1002.910965] env[61957]: DEBUG oslo_concurrency.lockutils [req-2313b73d-1368-497f-ade4-572589a0e0cd req-12788b97-57b7-4f81-961f-55699b647171 service nova] Lock "84cc9489-2a11-4510-8ad6-13dc41814856-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1002.912133] env[61957]: DEBUG nova.compute.manager [req-2313b73d-1368-497f-ade4-572589a0e0cd req-12788b97-57b7-4f81-961f-55699b647171 service nova] [instance: 84cc9489-2a11-4510-8ad6-13dc41814856] No waiting events found dispatching network-vif-plugged-b602c88c-d9b2-44bd-ba43-0ba7d976d8ac {{(pid=61957) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1002.912133] env[61957]: WARNING nova.compute.manager [req-2313b73d-1368-497f-ade4-572589a0e0cd req-12788b97-57b7-4f81-961f-55699b647171 service nova] [instance: 84cc9489-2a11-4510-8ad6-13dc41814856] Received unexpected event network-vif-plugged-b602c88c-d9b2-44bd-ba43-0ba7d976d8ac for instance with vm_state building and task_state spawning. [ 1002.912133] env[61957]: DEBUG nova.compute.manager [req-2313b73d-1368-497f-ade4-572589a0e0cd req-12788b97-57b7-4f81-961f-55699b647171 service nova] [instance: 84cc9489-2a11-4510-8ad6-13dc41814856] Received event network-changed-b602c88c-d9b2-44bd-ba43-0ba7d976d8ac {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1002.912133] env[61957]: DEBUG nova.compute.manager [req-2313b73d-1368-497f-ade4-572589a0e0cd req-12788b97-57b7-4f81-961f-55699b647171 service nova] [instance: 84cc9489-2a11-4510-8ad6-13dc41814856] Refreshing instance network info cache due to event network-changed-b602c88c-d9b2-44bd-ba43-0ba7d976d8ac. {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1002.912133] env[61957]: DEBUG oslo_concurrency.lockutils [req-2313b73d-1368-497f-ade4-572589a0e0cd req-12788b97-57b7-4f81-961f-55699b647171 service nova] Acquiring lock "refresh_cache-84cc9489-2a11-4510-8ad6-13dc41814856" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1003.185324] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e149abdd-1f3e-40fa-b73b-bf830d3ae686 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.212065] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-c17a70a2-b114-4298-9696-2f9ae04e6d89 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] [instance: f65458b2-0db6-4d98-8c82-b27b6945a3fd] Powering off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1003.212396] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e7512d63-399c-45fa-b85a-867863b788fd {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.216101] env[61957]: DEBUG oslo_concurrency.lockutils [None req-b10a76eb-ac59-4162-9083-648b4df991d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.320s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1003.218382] env[61957]: DEBUG oslo_concurrency.lockutils [None req-89f56e18-80ff-4394-b23a-cf5db06d4c7c tempest-ServerAddressesTestJSON-2054963291 tempest-ServerAddressesTestJSON-2054963291-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.964s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1003.218614] env[61957]: DEBUG nova.objects.instance [None req-89f56e18-80ff-4394-b23a-cf5db06d4c7c tempest-ServerAddressesTestJSON-2054963291 tempest-ServerAddressesTestJSON-2054963291-project-member] Lazy-loading 'resources' on Instance uuid fa4c150e-7a85-41a4-8890-9f2deb3b34b1 {{(pid=61957) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1003.220629] env[61957]: DEBUG oslo_vmware.api [None req-c17a70a2-b114-4298-9696-2f9ae04e6d89 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Waiting for the task: (returnval){ [ 1003.220629] env[61957]: value = "task-1277937" [ 1003.220629] env[61957]: _type = "Task" [ 1003.220629] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.230369] env[61957]: DEBUG oslo_vmware.api [None req-c17a70a2-b114-4298-9696-2f9ae04e6d89 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Task: {'id': task-1277937, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.252750] env[61957]: DEBUG oslo_vmware.api [None req-ab927ca0-dc5c-48f5-8ef0-a84245f36801 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52f572e8-3d16-ed93-241c-36428a5cee15, 'name': SearchDatastore_Task, 'duration_secs': 0.029621} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.253087] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ab927ca0-dc5c-48f5-8ef0-a84245f36801 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1003.253380] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-ab927ca0-dc5c-48f5-8ef0-a84245f36801 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore2] e2107d5c-8453-4d24-b364-2b5c1d125062/e2107d5c-8453-4d24-b364-2b5c1d125062.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1003.253664] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-91a09256-bf80-4090-be64-c457ce17b57e {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.261686] env[61957]: DEBUG oslo_vmware.api [None req-ab927ca0-dc5c-48f5-8ef0-a84245f36801 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Waiting for the task: (returnval){ [ 1003.261686] env[61957]: value = "task-1277938" [ 1003.261686] env[61957]: _type = "Task" [ 1003.261686] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.271351] env[61957]: DEBUG oslo_vmware.api [None req-ab927ca0-dc5c-48f5-8ef0-a84245f36801 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': task-1277938, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.305171] env[61957]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1003.305171] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]520e133f-02c6-2431-511f-d1dc6cf59e70" [ 1003.305171] env[61957]: _type = "HttpNfcLease" [ 1003.305171] env[61957]: } is ready. {{(pid=61957) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1003.305482] env[61957]: DEBUG oslo_vmware.rw_handles [None req-ad83dc8d-b942-4b5a-8561-778a1542e9de tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1003.305482] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]520e133f-02c6-2431-511f-d1dc6cf59e70" [ 1003.305482] env[61957]: _type = "HttpNfcLease" [ 1003.305482] env[61957]: }. {{(pid=61957) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1003.306294] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd985054-754a-40b4-8c7b-53b589eb706e {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.315344] env[61957]: DEBUG oslo_vmware.rw_handles [None req-ad83dc8d-b942-4b5a-8561-778a1542e9de tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52b02e8b-f180-7868-5405-b594f5954d48/disk-0.vmdk from lease info. {{(pid=61957) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1003.315563] env[61957]: DEBUG oslo_vmware.rw_handles [None req-ad83dc8d-b942-4b5a-8561-778a1542e9de tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52b02e8b-f180-7868-5405-b594f5954d48/disk-0.vmdk for reading. {{(pid=61957) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1003.382113] env[61957]: DEBUG nova.network.neutron [None req-8cf153c7-e439-4990-9fa2-f6d040c68c3c tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 84cc9489-2a11-4510-8ad6-13dc41814856] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1003.426897] env[61957]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-1c46fbe8-0846-4ccd-a88a-7be7f4d7fc31 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.591811] env[61957]: DEBUG nova.network.neutron [req-c1321c9d-d08a-4826-8061-e5ec571ad0d6 req-c94d95d3-cc5f-4aab-86de-0688851e28bf service nova] [instance: e2107d5c-8453-4d24-b364-2b5c1d125062] Updated VIF entry in instance network info cache for port f17ed646-b16b-40e9-a98f-a1324721caa1. {{(pid=61957) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1003.592305] env[61957]: DEBUG nova.network.neutron [req-c1321c9d-d08a-4826-8061-e5ec571ad0d6 req-c94d95d3-cc5f-4aab-86de-0688851e28bf service nova] [instance: e2107d5c-8453-4d24-b364-2b5c1d125062] Updating instance_info_cache with network_info: [{"id": "f17ed646-b16b-40e9-a98f-a1324721caa1", "address": "fa:16:3e:a4:aa:08", "network": {"id": "dc7d4c3b-1f91-4937-a2b0-1084e5c478f0", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1637370260-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "975189ebb3cc4cdb9391880f0c9ba6ba", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b71230ae-e879-4384-88ce-fe64c86fce22", "external-id": "nsx-vlan-transportzone-473", "segmentation_id": 473, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf17ed646-b1", "ovs_interfaceid": "f17ed646-b16b-40e9-a98f-a1324721caa1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1003.639024] env[61957]: DEBUG nova.network.neutron [None req-8cf153c7-e439-4990-9fa2-f6d040c68c3c tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 84cc9489-2a11-4510-8ad6-13dc41814856] Updating instance_info_cache with network_info: [{"id": "b602c88c-d9b2-44bd-ba43-0ba7d976d8ac", "address": "fa:16:3e:c1:41:6c", "network": {"id": "4f0e57be-8906-48b5-865f-59ff12250a8b", "bridge": "br-int", "label": "tempest-ServersTestJSON-1916182687-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "65fd71bcd1bf41238a9cc3a5d6dd4924", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "939c05b6-8f31-4f3a-95ac-6297e0bd243e", "external-id": "nsx-vlan-transportzone-825", "segmentation_id": 825, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb602c88c-d9", "ovs_interfaceid": "b602c88c-d9b2-44bd-ba43-0ba7d976d8ac", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1003.736182] env[61957]: DEBUG oslo_concurrency.lockutils [None req-b10a76eb-ac59-4162-9083-648b4df991d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Lock "98929ed3-d420-45cd-9cde-9738b2ac8251" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 23.328s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1003.736182] env[61957]: DEBUG oslo_concurrency.lockutils [None req-58e997cd-a82d-4012-a69d-a48696b8b5f1 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Lock "98929ed3-d420-45cd-9cde-9738b2ac8251" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 3.568s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1003.736182] env[61957]: DEBUG oslo_concurrency.lockutils [None req-58e997cd-a82d-4012-a69d-a48696b8b5f1 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Acquiring lock "98929ed3-d420-45cd-9cde-9738b2ac8251-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1003.736182] env[61957]: DEBUG oslo_concurrency.lockutils [None req-58e997cd-a82d-4012-a69d-a48696b8b5f1 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Lock "98929ed3-d420-45cd-9cde-9738b2ac8251-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1003.736182] env[61957]: DEBUG oslo_concurrency.lockutils [None req-58e997cd-a82d-4012-a69d-a48696b8b5f1 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Lock "98929ed3-d420-45cd-9cde-9738b2ac8251-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1003.741313] env[61957]: DEBUG oslo_vmware.api [None req-c17a70a2-b114-4298-9696-2f9ae04e6d89 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Task: {'id': task-1277937, 'name': PowerOffVM_Task, 'duration_secs': 0.211596} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.742818] env[61957]: INFO nova.compute.manager [None req-58e997cd-a82d-4012-a69d-a48696b8b5f1 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 98929ed3-d420-45cd-9cde-9738b2ac8251] Terminating instance [ 1003.745467] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-c17a70a2-b114-4298-9696-2f9ae04e6d89 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] [instance: f65458b2-0db6-4d98-8c82-b27b6945a3fd] Powered off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1003.752892] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-c17a70a2-b114-4298-9696-2f9ae04e6d89 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] [instance: f65458b2-0db6-4d98-8c82-b27b6945a3fd] Reconfiguring VM instance instance-0000005b to detach disk 2001 {{(pid=61957) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1003.754019] env[61957]: DEBUG nova.compute.manager [None req-58e997cd-a82d-4012-a69d-a48696b8b5f1 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 98929ed3-d420-45cd-9cde-9738b2ac8251] Start destroying the instance on the hypervisor. {{(pid=61957) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1003.754389] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-58e997cd-a82d-4012-a69d-a48696b8b5f1 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 98929ed3-d420-45cd-9cde-9738b2ac8251] Destroying instance {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1003.758339] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ccdebdd6-9059-4e6c-9d22-bf80f9bfb9e3 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.773169] env[61957]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6f79eadb-5cb6-499a-a909-0bb1513e7a13 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.800147] env[61957]: DEBUG oslo_vmware.api [None req-ab927ca0-dc5c-48f5-8ef0-a84245f36801 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': task-1277938, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.487816} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.800463] env[61957]: DEBUG oslo_vmware.api [None req-c17a70a2-b114-4298-9696-2f9ae04e6d89 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Waiting for the task: (returnval){ [ 1003.800463] env[61957]: value = "task-1277939" [ 1003.800463] env[61957]: _type = "Task" [ 1003.800463] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.800795] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-ab927ca0-dc5c-48f5-8ef0-a84245f36801 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore2] e2107d5c-8453-4d24-b364-2b5c1d125062/e2107d5c-8453-4d24-b364-2b5c1d125062.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1003.801953] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-ab927ca0-dc5c-48f5-8ef0-a84245f36801 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: e2107d5c-8453-4d24-b364-2b5c1d125062] Extending root virtual disk to 1048576 {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1003.805447] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f24a25c-101f-449c-88b0-efa94f4302d7 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.824764] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-bc2e24fe-d689-4b64-8962-deff3fd8974b {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.836702] env[61957]: DEBUG oslo_vmware.api [None req-c17a70a2-b114-4298-9696-2f9ae04e6d89 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Task: {'id': task-1277939, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.839077] env[61957]: DEBUG oslo_vmware.api [None req-ab927ca0-dc5c-48f5-8ef0-a84245f36801 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Waiting for the task: (returnval){ [ 1003.839077] env[61957]: value = "task-1277940" [ 1003.839077] env[61957]: _type = "Task" [ 1003.839077] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.864888] env[61957]: DEBUG oslo_vmware.api [None req-ab927ca0-dc5c-48f5-8ef0-a84245f36801 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': task-1277940, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.865248] env[61957]: WARNING nova.virt.vmwareapi.vmops [None req-58e997cd-a82d-4012-a69d-a48696b8b5f1 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 98929ed3-d420-45cd-9cde-9738b2ac8251] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 98929ed3-d420-45cd-9cde-9738b2ac8251 could not be found. [ 1003.865502] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-58e997cd-a82d-4012-a69d-a48696b8b5f1 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 98929ed3-d420-45cd-9cde-9738b2ac8251] Instance destroyed {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1003.865761] env[61957]: INFO nova.compute.manager [None req-58e997cd-a82d-4012-a69d-a48696b8b5f1 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 98929ed3-d420-45cd-9cde-9738b2ac8251] Took 0.11 seconds to destroy the instance on the hypervisor. [ 1003.866363] env[61957]: DEBUG oslo.service.loopingcall [None req-58e997cd-a82d-4012-a69d-a48696b8b5f1 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1003.870267] env[61957]: DEBUG nova.compute.manager [-] [instance: 98929ed3-d420-45cd-9cde-9738b2ac8251] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1003.870267] env[61957]: DEBUG nova.network.neutron [-] [instance: 98929ed3-d420-45cd-9cde-9738b2ac8251] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1003.876097] env[61957]: DEBUG nova.network.neutron [None req-690abcce-5857-41f4-8c3f-00315918c375 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2] Successfully updated port: 6a70445e-1266-453e-a6ff-50e8cbdcfd08 {{(pid=61957) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1004.010702] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1fc008b-c055-41d7-a505-24108f04293a {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.021451] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55f01104-4469-4d50-9cda-6329fdfeca10 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.054650] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57bb09be-2054-480f-a88b-5ad5f0773846 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.063250] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6781c2c2-d6e9-41f6-af9c-7c7f53a47ff9 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.080491] env[61957]: DEBUG nova.compute.provider_tree [None req-89f56e18-80ff-4394-b23a-cf5db06d4c7c tempest-ServerAddressesTestJSON-2054963291 tempest-ServerAddressesTestJSON-2054963291-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1004.094795] env[61957]: DEBUG oslo_concurrency.lockutils [req-c1321c9d-d08a-4826-8061-e5ec571ad0d6 req-c94d95d3-cc5f-4aab-86de-0688851e28bf service nova] Releasing lock "refresh_cache-e2107d5c-8453-4d24-b364-2b5c1d125062" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1004.144284] env[61957]: DEBUG oslo_concurrency.lockutils [None req-8cf153c7-e439-4990-9fa2-f6d040c68c3c tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Releasing lock "refresh_cache-84cc9489-2a11-4510-8ad6-13dc41814856" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1004.144745] env[61957]: DEBUG nova.compute.manager [None req-8cf153c7-e439-4990-9fa2-f6d040c68c3c tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 84cc9489-2a11-4510-8ad6-13dc41814856] Instance network_info: |[{"id": "b602c88c-d9b2-44bd-ba43-0ba7d976d8ac", "address": "fa:16:3e:c1:41:6c", "network": {"id": "4f0e57be-8906-48b5-865f-59ff12250a8b", "bridge": "br-int", "label": "tempest-ServersTestJSON-1916182687-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "65fd71bcd1bf41238a9cc3a5d6dd4924", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "939c05b6-8f31-4f3a-95ac-6297e0bd243e", "external-id": "nsx-vlan-transportzone-825", "segmentation_id": 825, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb602c88c-d9", "ovs_interfaceid": "b602c88c-d9b2-44bd-ba43-0ba7d976d8ac", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1004.148487] env[61957]: DEBUG oslo_concurrency.lockutils [req-2313b73d-1368-497f-ade4-572589a0e0cd req-12788b97-57b7-4f81-961f-55699b647171 service nova] Acquired lock "refresh_cache-84cc9489-2a11-4510-8ad6-13dc41814856" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1004.148742] env[61957]: DEBUG nova.network.neutron [req-2313b73d-1368-497f-ade4-572589a0e0cd req-12788b97-57b7-4f81-961f-55699b647171 service nova] [instance: 84cc9489-2a11-4510-8ad6-13dc41814856] Refreshing network info cache for port b602c88c-d9b2-44bd-ba43-0ba7d976d8ac {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1004.150033] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-8cf153c7-e439-4990-9fa2-f6d040c68c3c tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 84cc9489-2a11-4510-8ad6-13dc41814856] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c1:41:6c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '939c05b6-8f31-4f3a-95ac-6297e0bd243e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b602c88c-d9b2-44bd-ba43-0ba7d976d8ac', 'vif_model': 'vmxnet3'}] {{(pid=61957) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1004.158571] env[61957]: DEBUG oslo.service.loopingcall [None req-8cf153c7-e439-4990-9fa2-f6d040c68c3c tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1004.160793] env[61957]: DEBUG nova.network.neutron [-] [instance: 98929ed3-d420-45cd-9cde-9738b2ac8251] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1004.164236] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 84cc9489-2a11-4510-8ad6-13dc41814856] Creating VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1004.164883] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3f949771-1cea-48fc-a13e-2dabab03cf9a {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.192185] env[61957]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1004.192185] env[61957]: value = "task-1277941" [ 1004.192185] env[61957]: _type = "Task" [ 1004.192185] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.202154] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277941, 'name': CreateVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.314686] env[61957]: DEBUG oslo_vmware.api [None req-c17a70a2-b114-4298-9696-2f9ae04e6d89 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Task: {'id': task-1277939, 'name': ReconfigVM_Task, 'duration_secs': 0.504103} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.314894] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-c17a70a2-b114-4298-9696-2f9ae04e6d89 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] [instance: f65458b2-0db6-4d98-8c82-b27b6945a3fd] Reconfigured VM instance instance-0000005b to detach disk 2001 {{(pid=61957) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1004.315455] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-c17a70a2-b114-4298-9696-2f9ae04e6d89 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] [instance: f65458b2-0db6-4d98-8c82-b27b6945a3fd] Powering on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1004.315561] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5ef8646a-d14b-4c6e-94d5-c5563ca61dee {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.323540] env[61957]: DEBUG oslo_vmware.api [None req-c17a70a2-b114-4298-9696-2f9ae04e6d89 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Waiting for the task: (returnval){ [ 1004.323540] env[61957]: value = "task-1277942" [ 1004.323540] env[61957]: _type = "Task" [ 1004.323540] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.335352] env[61957]: DEBUG oslo_vmware.api [None req-c17a70a2-b114-4298-9696-2f9ae04e6d89 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Task: {'id': task-1277942, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.349824] env[61957]: DEBUG oslo_vmware.api [None req-ab927ca0-dc5c-48f5-8ef0-a84245f36801 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': task-1277940, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072422} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.350163] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-ab927ca0-dc5c-48f5-8ef0-a84245f36801 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: e2107d5c-8453-4d24-b364-2b5c1d125062] Extended root virtual disk {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1004.351969] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4d17547-1e94-4f11-83e9-2ee796185106 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.378826] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-ab927ca0-dc5c-48f5-8ef0-a84245f36801 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: e2107d5c-8453-4d24-b364-2b5c1d125062] Reconfiguring VM instance instance-0000005e to attach disk [datastore2] e2107d5c-8453-4d24-b364-2b5c1d125062/e2107d5c-8453-4d24-b364-2b5c1d125062.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1004.382991] env[61957]: DEBUG oslo_concurrency.lockutils [None req-690abcce-5857-41f4-8c3f-00315918c375 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Acquiring lock "refresh_cache-08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1004.382991] env[61957]: DEBUG oslo_concurrency.lockutils [None req-690abcce-5857-41f4-8c3f-00315918c375 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Acquired lock "refresh_cache-08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1004.382991] env[61957]: DEBUG nova.network.neutron [None req-690abcce-5857-41f4-8c3f-00315918c375 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1004.385386] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fde5125c-aa4e-4373-a5e2-9626eae1b326 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.409713] env[61957]: DEBUG oslo_vmware.api [None req-ab927ca0-dc5c-48f5-8ef0-a84245f36801 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Waiting for the task: (returnval){ [ 1004.409713] env[61957]: value = "task-1277943" [ 1004.409713] env[61957]: _type = "Task" [ 1004.409713] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.419510] env[61957]: DEBUG oslo_vmware.api [None req-ab927ca0-dc5c-48f5-8ef0-a84245f36801 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': task-1277943, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.447102] env[61957]: DEBUG nova.network.neutron [req-2313b73d-1368-497f-ade4-572589a0e0cd req-12788b97-57b7-4f81-961f-55699b647171 service nova] [instance: 84cc9489-2a11-4510-8ad6-13dc41814856] Updated VIF entry in instance network info cache for port b602c88c-d9b2-44bd-ba43-0ba7d976d8ac. {{(pid=61957) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1004.448034] env[61957]: DEBUG nova.network.neutron [req-2313b73d-1368-497f-ade4-572589a0e0cd req-12788b97-57b7-4f81-961f-55699b647171 service nova] [instance: 84cc9489-2a11-4510-8ad6-13dc41814856] Updating instance_info_cache with network_info: [{"id": "b602c88c-d9b2-44bd-ba43-0ba7d976d8ac", "address": "fa:16:3e:c1:41:6c", "network": {"id": "4f0e57be-8906-48b5-865f-59ff12250a8b", "bridge": "br-int", "label": "tempest-ServersTestJSON-1916182687-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "65fd71bcd1bf41238a9cc3a5d6dd4924", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "939c05b6-8f31-4f3a-95ac-6297e0bd243e", "external-id": "nsx-vlan-transportzone-825", "segmentation_id": 825, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb602c88c-d9", "ovs_interfaceid": "b602c88c-d9b2-44bd-ba43-0ba7d976d8ac", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1004.584147] env[61957]: DEBUG nova.scheduler.client.report [None req-89f56e18-80ff-4394-b23a-cf5db06d4c7c tempest-ServerAddressesTestJSON-2054963291 tempest-ServerAddressesTestJSON-2054963291-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1004.665414] env[61957]: INFO nova.compute.manager [-] [instance: 98929ed3-d420-45cd-9cde-9738b2ac8251] Took 0.79 seconds to deallocate network for instance. [ 1004.704461] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277941, 'name': CreateVM_Task, 'duration_secs': 0.395795} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.705498] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 84cc9489-2a11-4510-8ad6-13dc41814856] Created VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1004.706637] env[61957]: DEBUG oslo_concurrency.lockutils [None req-8cf153c7-e439-4990-9fa2-f6d040c68c3c tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1004.706845] env[61957]: DEBUG oslo_concurrency.lockutils [None req-8cf153c7-e439-4990-9fa2-f6d040c68c3c tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Acquired lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1004.707792] env[61957]: DEBUG oslo_concurrency.lockutils [None req-8cf153c7-e439-4990-9fa2-f6d040c68c3c tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1004.707909] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1e397585-4ef9-400e-ad26-3e99ef1db88a {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.714032] env[61957]: DEBUG oslo_vmware.api [None req-8cf153c7-e439-4990-9fa2-f6d040c68c3c tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Waiting for the task: (returnval){ [ 1004.714032] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]523b847c-c7f0-4361-08c6-49a7b8b58b62" [ 1004.714032] env[61957]: _type = "Task" [ 1004.714032] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.722941] env[61957]: DEBUG oslo_vmware.api [None req-8cf153c7-e439-4990-9fa2-f6d040c68c3c tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]523b847c-c7f0-4361-08c6-49a7b8b58b62, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.833435] env[61957]: DEBUG oslo_vmware.api [None req-c17a70a2-b114-4298-9696-2f9ae04e6d89 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Task: {'id': task-1277942, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.921664] env[61957]: DEBUG oslo_vmware.api [None req-ab927ca0-dc5c-48f5-8ef0-a84245f36801 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': task-1277943, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.922555] env[61957]: WARNING nova.network.neutron [None req-690abcce-5857-41f4-8c3f-00315918c375 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2] 364cabcb-046d-4f91-b1ee-5dd3adfae6d7 already exists in list: networks containing: ['364cabcb-046d-4f91-b1ee-5dd3adfae6d7']. ignoring it [ 1004.937734] env[61957]: DEBUG nova.compute.manager [req-494b367b-4b4f-41ab-88d0-e44d5604dada req-2b1cd18e-1ec7-4ac2-bcb4-fc2785a8b8f1 service nova] [instance: 08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2] Received event network-vif-plugged-6a70445e-1266-453e-a6ff-50e8cbdcfd08 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1004.937999] env[61957]: DEBUG oslo_concurrency.lockutils [req-494b367b-4b4f-41ab-88d0-e44d5604dada req-2b1cd18e-1ec7-4ac2-bcb4-fc2785a8b8f1 service nova] Acquiring lock "08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1004.945785] env[61957]: DEBUG oslo_concurrency.lockutils [req-494b367b-4b4f-41ab-88d0-e44d5604dada req-2b1cd18e-1ec7-4ac2-bcb4-fc2785a8b8f1 service nova] Lock "08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1004.945785] env[61957]: DEBUG oslo_concurrency.lockutils [req-494b367b-4b4f-41ab-88d0-e44d5604dada req-2b1cd18e-1ec7-4ac2-bcb4-fc2785a8b8f1 service nova] Lock "08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.003s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1004.945785] env[61957]: DEBUG nova.compute.manager [req-494b367b-4b4f-41ab-88d0-e44d5604dada req-2b1cd18e-1ec7-4ac2-bcb4-fc2785a8b8f1 service nova] [instance: 08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2] No waiting events found dispatching network-vif-plugged-6a70445e-1266-453e-a6ff-50e8cbdcfd08 {{(pid=61957) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1004.947544] env[61957]: WARNING nova.compute.manager [req-494b367b-4b4f-41ab-88d0-e44d5604dada req-2b1cd18e-1ec7-4ac2-bcb4-fc2785a8b8f1 service nova] [instance: 08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2] Received unexpected event network-vif-plugged-6a70445e-1266-453e-a6ff-50e8cbdcfd08 for instance with vm_state active and task_state None. [ 1004.947544] env[61957]: DEBUG nova.compute.manager [req-494b367b-4b4f-41ab-88d0-e44d5604dada req-2b1cd18e-1ec7-4ac2-bcb4-fc2785a8b8f1 service nova] [instance: 08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2] Received event network-changed-6a70445e-1266-453e-a6ff-50e8cbdcfd08 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1004.947728] env[61957]: DEBUG nova.compute.manager [req-494b367b-4b4f-41ab-88d0-e44d5604dada req-2b1cd18e-1ec7-4ac2-bcb4-fc2785a8b8f1 service nova] [instance: 08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2] Refreshing instance network info cache due to event network-changed-6a70445e-1266-453e-a6ff-50e8cbdcfd08. {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1004.948125] env[61957]: DEBUG oslo_concurrency.lockutils [req-494b367b-4b4f-41ab-88d0-e44d5604dada req-2b1cd18e-1ec7-4ac2-bcb4-fc2785a8b8f1 service nova] Acquiring lock "refresh_cache-08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1004.951185] env[61957]: DEBUG oslo_concurrency.lockutils [req-2313b73d-1368-497f-ade4-572589a0e0cd req-12788b97-57b7-4f81-961f-55699b647171 service nova] Releasing lock "refresh_cache-84cc9489-2a11-4510-8ad6-13dc41814856" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1005.089568] env[61957]: DEBUG oslo_concurrency.lockutils [None req-89f56e18-80ff-4394-b23a-cf5db06d4c7c tempest-ServerAddressesTestJSON-2054963291 tempest-ServerAddressesTestJSON-2054963291-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.871s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1005.092842] env[61957]: DEBUG oslo_concurrency.lockutils [None req-0b16b4fe-c18a-4e4e-8833-61daba0fafce tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 2.557s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1005.112813] env[61957]: INFO nova.scheduler.client.report [None req-89f56e18-80ff-4394-b23a-cf5db06d4c7c tempest-ServerAddressesTestJSON-2054963291 tempest-ServerAddressesTestJSON-2054963291-project-member] Deleted allocations for instance fa4c150e-7a85-41a4-8890-9f2deb3b34b1 [ 1005.224823] env[61957]: DEBUG oslo_vmware.api [None req-8cf153c7-e439-4990-9fa2-f6d040c68c3c tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]523b847c-c7f0-4361-08c6-49a7b8b58b62, 'name': SearchDatastore_Task, 'duration_secs': 0.025892} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.225830] env[61957]: DEBUG nova.network.neutron [None req-690abcce-5857-41f4-8c3f-00315918c375 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2] Updating instance_info_cache with network_info: [{"id": "f62a9f12-0b55-4138-a40e-b431855487fe", "address": "fa:16:3e:af:48:d3", "network": {"id": "364cabcb-046d-4f91-b1ee-5dd3adfae6d7", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1432396967-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.219", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8713aa35bcb24b86ad0b58ca9fc991ba", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b7a73c01-1bb9-4612-a1a7-16d71b732e81", "external-id": "nsx-vlan-transportzone-711", "segmentation_id": 711, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf62a9f12-0b", "ovs_interfaceid": "f62a9f12-0b55-4138-a40e-b431855487fe", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "6a70445e-1266-453e-a6ff-50e8cbdcfd08", "address": "fa:16:3e:bb:2f:b6", "network": {"id": "364cabcb-046d-4f91-b1ee-5dd3adfae6d7", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1432396967-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8713aa35bcb24b86ad0b58ca9fc991ba", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b7a73c01-1bb9-4612-a1a7-16d71b732e81", "external-id": "nsx-vlan-transportzone-711", "segmentation_id": 711, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6a70445e-12", "ovs_interfaceid": "6a70445e-1266-453e-a6ff-50e8cbdcfd08", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1005.227272] env[61957]: DEBUG oslo_concurrency.lockutils [None req-8cf153c7-e439-4990-9fa2-f6d040c68c3c tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Releasing lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1005.227542] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-8cf153c7-e439-4990-9fa2-f6d040c68c3c tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 84cc9489-2a11-4510-8ad6-13dc41814856] Processing image 11c76a2c-f705-470a-ba9d-4657858bab38 {{(pid=61957) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1005.228086] env[61957]: DEBUG oslo_concurrency.lockutils [None req-8cf153c7-e439-4990-9fa2-f6d040c68c3c tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1005.228086] env[61957]: DEBUG oslo_concurrency.lockutils [None req-8cf153c7-e439-4990-9fa2-f6d040c68c3c tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Acquired lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1005.228263] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-8cf153c7-e439-4990-9fa2-f6d040c68c3c tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1005.228554] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-fa6a098b-67ea-4bc1-9e14-1112e50feaad {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.237995] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-8cf153c7-e439-4990-9fa2-f6d040c68c3c tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1005.238202] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-8cf153c7-e439-4990-9fa2-f6d040c68c3c tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61957) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1005.239799] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2d2b6f43-cb6d-4e69-b2f7-5498531bbc77 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.246020] env[61957]: DEBUG oslo_vmware.api [None req-8cf153c7-e439-4990-9fa2-f6d040c68c3c tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Waiting for the task: (returnval){ [ 1005.246020] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]521f6050-7743-6bc8-4ccc-dfd28c845e6e" [ 1005.246020] env[61957]: _type = "Task" [ 1005.246020] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.254753] env[61957]: DEBUG oslo_vmware.api [None req-8cf153c7-e439-4990-9fa2-f6d040c68c3c tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]521f6050-7743-6bc8-4ccc-dfd28c845e6e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.333970] env[61957]: DEBUG oslo_vmware.api [None req-c17a70a2-b114-4298-9696-2f9ae04e6d89 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Task: {'id': task-1277942, 'name': PowerOnVM_Task, 'duration_secs': 0.536523} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.334298] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-c17a70a2-b114-4298-9696-2f9ae04e6d89 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] [instance: f65458b2-0db6-4d98-8c82-b27b6945a3fd] Powered on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1005.334540] env[61957]: DEBUG nova.compute.manager [None req-c17a70a2-b114-4298-9696-2f9ae04e6d89 tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] [instance: f65458b2-0db6-4d98-8c82-b27b6945a3fd] Checking state {{(pid=61957) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1005.335388] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-877cc285-bb0c-4eb0-89d7-6be404b107e7 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.420998] env[61957]: DEBUG oslo_vmware.api [None req-ab927ca0-dc5c-48f5-8ef0-a84245f36801 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': task-1277943, 'name': ReconfigVM_Task, 'duration_secs': 1.012957} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.421313] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-ab927ca0-dc5c-48f5-8ef0-a84245f36801 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: e2107d5c-8453-4d24-b364-2b5c1d125062] Reconfigured VM instance instance-0000005e to attach disk [datastore2] e2107d5c-8453-4d24-b364-2b5c1d125062/e2107d5c-8453-4d24-b364-2b5c1d125062.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1005.421999] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e849fc3f-0f93-494e-8c11-2af3fcbfaf0d {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.428151] env[61957]: DEBUG oslo_vmware.api [None req-ab927ca0-dc5c-48f5-8ef0-a84245f36801 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Waiting for the task: (returnval){ [ 1005.428151] env[61957]: value = "task-1277944" [ 1005.428151] env[61957]: _type = "Task" [ 1005.428151] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.436590] env[61957]: DEBUG oslo_vmware.api [None req-ab927ca0-dc5c-48f5-8ef0-a84245f36801 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': task-1277944, 'name': Rename_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.598484] env[61957]: INFO nova.compute.claims [None req-0b16b4fe-c18a-4e4e-8833-61daba0fafce tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: 6509707e-afc6-4c4a-88f9-a471865d510c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1005.621032] env[61957]: DEBUG oslo_concurrency.lockutils [None req-89f56e18-80ff-4394-b23a-cf5db06d4c7c tempest-ServerAddressesTestJSON-2054963291 tempest-ServerAddressesTestJSON-2054963291-project-member] Lock "fa4c150e-7a85-41a4-8890-9f2deb3b34b1" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.488s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1005.692749] env[61957]: DEBUG oslo_concurrency.lockutils [None req-58e997cd-a82d-4012-a69d-a48696b8b5f1 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Lock "98929ed3-d420-45cd-9cde-9738b2ac8251" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 1.959s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1005.729570] env[61957]: DEBUG oslo_concurrency.lockutils [None req-690abcce-5857-41f4-8c3f-00315918c375 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Releasing lock "refresh_cache-08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1005.730324] env[61957]: DEBUG oslo_concurrency.lockutils [None req-690abcce-5857-41f4-8c3f-00315918c375 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Acquiring lock "08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1005.730559] env[61957]: DEBUG oslo_concurrency.lockutils [None req-690abcce-5857-41f4-8c3f-00315918c375 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Acquired lock "08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1005.730826] env[61957]: DEBUG oslo_concurrency.lockutils [req-494b367b-4b4f-41ab-88d0-e44d5604dada req-2b1cd18e-1ec7-4ac2-bcb4-fc2785a8b8f1 service nova] Acquired lock "refresh_cache-08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1005.731046] env[61957]: DEBUG nova.network.neutron [req-494b367b-4b4f-41ab-88d0-e44d5604dada req-2b1cd18e-1ec7-4ac2-bcb4-fc2785a8b8f1 service nova] [instance: 08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2] Refreshing network info cache for port 6a70445e-1266-453e-a6ff-50e8cbdcfd08 {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1005.732786] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3de408f6-81dd-4574-b500-fdac3421cdcf {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.751507] env[61957]: DEBUG nova.virt.hardware [None req-690abcce-5857-41f4-8c3f-00315918c375 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T17:22:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1005.751758] env[61957]: DEBUG nova.virt.hardware [None req-690abcce-5857-41f4-8c3f-00315918c375 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Flavor limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1005.751925] env[61957]: DEBUG nova.virt.hardware [None req-690abcce-5857-41f4-8c3f-00315918c375 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Image limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1005.752134] env[61957]: DEBUG nova.virt.hardware [None req-690abcce-5857-41f4-8c3f-00315918c375 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Flavor pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1005.752283] env[61957]: DEBUG nova.virt.hardware [None req-690abcce-5857-41f4-8c3f-00315918c375 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Image pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1005.752429] env[61957]: DEBUG nova.virt.hardware [None req-690abcce-5857-41f4-8c3f-00315918c375 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1005.752633] env[61957]: DEBUG nova.virt.hardware [None req-690abcce-5857-41f4-8c3f-00315918c375 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1005.752793] env[61957]: DEBUG nova.virt.hardware [None req-690abcce-5857-41f4-8c3f-00315918c375 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1005.753008] env[61957]: DEBUG nova.virt.hardware [None req-690abcce-5857-41f4-8c3f-00315918c375 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Got 1 possible topologies {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1005.753209] env[61957]: DEBUG nova.virt.hardware [None req-690abcce-5857-41f4-8c3f-00315918c375 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1005.753385] env[61957]: DEBUG nova.virt.hardware [None req-690abcce-5857-41f4-8c3f-00315918c375 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1005.759623] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-690abcce-5857-41f4-8c3f-00315918c375 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2] Reconfiguring VM to attach interface {{(pid=61957) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1929}} [ 1005.763774] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f92ee369-cf9f-4b3c-883a-c5ef67c716e7 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.785347] env[61957]: DEBUG oslo_vmware.api [None req-8cf153c7-e439-4990-9fa2-f6d040c68c3c tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]521f6050-7743-6bc8-4ccc-dfd28c845e6e, 'name': SearchDatastore_Task, 'duration_secs': 0.03964} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.787324] env[61957]: DEBUG oslo_vmware.api [None req-690abcce-5857-41f4-8c3f-00315918c375 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Waiting for the task: (returnval){ [ 1005.787324] env[61957]: value = "task-1277945" [ 1005.787324] env[61957]: _type = "Task" [ 1005.787324] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.787567] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f4b5b3b5-a468-47cd-8f62-f47b0ec8184b {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.798494] env[61957]: DEBUG oslo_vmware.api [None req-8cf153c7-e439-4990-9fa2-f6d040c68c3c tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Waiting for the task: (returnval){ [ 1005.798494] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52bd5d87-7aae-087f-cbdf-84020eaf474e" [ 1005.798494] env[61957]: _type = "Task" [ 1005.798494] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.802020] env[61957]: DEBUG oslo_vmware.api [None req-690abcce-5857-41f4-8c3f-00315918c375 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': task-1277945, 'name': ReconfigVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.809833] env[61957]: DEBUG oslo_vmware.api [None req-8cf153c7-e439-4990-9fa2-f6d040c68c3c tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52bd5d87-7aae-087f-cbdf-84020eaf474e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.940445] env[61957]: DEBUG oslo_vmware.api [None req-ab927ca0-dc5c-48f5-8ef0-a84245f36801 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': task-1277944, 'name': Rename_Task, 'duration_secs': 0.403265} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.940808] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-ab927ca0-dc5c-48f5-8ef0-a84245f36801 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: e2107d5c-8453-4d24-b364-2b5c1d125062] Powering on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1005.941106] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ec41ca54-2d4c-44e9-ac3d-59f558132cc8 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.947842] env[61957]: DEBUG oslo_vmware.api [None req-ab927ca0-dc5c-48f5-8ef0-a84245f36801 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Waiting for the task: (returnval){ [ 1005.947842] env[61957]: value = "task-1277946" [ 1005.947842] env[61957]: _type = "Task" [ 1005.947842] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.956757] env[61957]: DEBUG oslo_vmware.api [None req-ab927ca0-dc5c-48f5-8ef0-a84245f36801 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': task-1277946, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.106010] env[61957]: INFO nova.compute.resource_tracker [None req-0b16b4fe-c18a-4e4e-8833-61daba0fafce tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: 6509707e-afc6-4c4a-88f9-a471865d510c] Updating resource usage from migration c0538d2d-e6c4-4a67-b30e-dadb1e5d8f8e [ 1006.288607] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9aae2b8-2891-4291-b3a6-da4e83d1fb8a {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.302646] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9dcf01d4-c1b7-4d44-8591-76693d8b775c {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.306023] env[61957]: DEBUG oslo_vmware.api [None req-690abcce-5857-41f4-8c3f-00315918c375 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': task-1277945, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.316139] env[61957]: DEBUG oslo_vmware.api [None req-8cf153c7-e439-4990-9fa2-f6d040c68c3c tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52bd5d87-7aae-087f-cbdf-84020eaf474e, 'name': SearchDatastore_Task, 'duration_secs': 0.021712} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.342471] env[61957]: DEBUG oslo_concurrency.lockutils [None req-8cf153c7-e439-4990-9fa2-f6d040c68c3c tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Releasing lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1006.342471] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-8cf153c7-e439-4990-9fa2-f6d040c68c3c tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore1] 84cc9489-2a11-4510-8ad6-13dc41814856/84cc9489-2a11-4510-8ad6-13dc41814856.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1006.342725] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-eb83d1dd-bfd8-401f-a1fc-7c137556e4e2 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.345675] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67471b09-ad78-450c-8c7f-c8ddd332a13a {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.360318] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af31a634-8150-4e97-9145-d4f1b2d0fa53 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.366785] env[61957]: DEBUG oslo_vmware.api [None req-8cf153c7-e439-4990-9fa2-f6d040c68c3c tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Waiting for the task: (returnval){ [ 1006.366785] env[61957]: value = "task-1277947" [ 1006.366785] env[61957]: _type = "Task" [ 1006.366785] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1006.383796] env[61957]: DEBUG nova.compute.provider_tree [None req-0b16b4fe-c18a-4e4e-8833-61daba0fafce tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1006.395417] env[61957]: DEBUG oslo_vmware.api [None req-8cf153c7-e439-4990-9fa2-f6d040c68c3c tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': task-1277947, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.458965] env[61957]: DEBUG oslo_vmware.api [None req-ab927ca0-dc5c-48f5-8ef0-a84245f36801 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': task-1277946, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.592128] env[61957]: DEBUG nova.network.neutron [req-494b367b-4b4f-41ab-88d0-e44d5604dada req-2b1cd18e-1ec7-4ac2-bcb4-fc2785a8b8f1 service nova] [instance: 08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2] Updated VIF entry in instance network info cache for port 6a70445e-1266-453e-a6ff-50e8cbdcfd08. {{(pid=61957) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1006.593617] env[61957]: DEBUG nova.network.neutron [req-494b367b-4b4f-41ab-88d0-e44d5604dada req-2b1cd18e-1ec7-4ac2-bcb4-fc2785a8b8f1 service nova] [instance: 08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2] Updating instance_info_cache with network_info: [{"id": "f62a9f12-0b55-4138-a40e-b431855487fe", "address": "fa:16:3e:af:48:d3", "network": {"id": "364cabcb-046d-4f91-b1ee-5dd3adfae6d7", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1432396967-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.219", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8713aa35bcb24b86ad0b58ca9fc991ba", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b7a73c01-1bb9-4612-a1a7-16d71b732e81", "external-id": "nsx-vlan-transportzone-711", "segmentation_id": 711, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf62a9f12-0b", "ovs_interfaceid": "f62a9f12-0b55-4138-a40e-b431855487fe", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "6a70445e-1266-453e-a6ff-50e8cbdcfd08", "address": "fa:16:3e:bb:2f:b6", "network": {"id": "364cabcb-046d-4f91-b1ee-5dd3adfae6d7", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1432396967-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8713aa35bcb24b86ad0b58ca9fc991ba", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b7a73c01-1bb9-4612-a1a7-16d71b732e81", "external-id": "nsx-vlan-transportzone-711", "segmentation_id": 711, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6a70445e-12", "ovs_interfaceid": "6a70445e-1266-453e-a6ff-50e8cbdcfd08", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1006.667856] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d92c7007-db9c-429d-aa03-41f7719669de tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Acquiring lock "927af4c3-c7e8-404c-839e-d25af5e384a9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1006.668125] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d92c7007-db9c-429d-aa03-41f7719669de tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Lock "927af4c3-c7e8-404c-839e-d25af5e384a9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1006.802638] env[61957]: DEBUG oslo_vmware.api [None req-690abcce-5857-41f4-8c3f-00315918c375 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': task-1277945, 'name': ReconfigVM_Task, 'duration_secs': 0.745112} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.803352] env[61957]: DEBUG oslo_concurrency.lockutils [None req-690abcce-5857-41f4-8c3f-00315918c375 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Releasing lock "08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1006.803573] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-690abcce-5857-41f4-8c3f-00315918c375 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2] Reconfigured VM to attach interface {{(pid=61957) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1943}} [ 1006.879618] env[61957]: DEBUG oslo_vmware.api [None req-8cf153c7-e439-4990-9fa2-f6d040c68c3c tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': task-1277947, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.482137} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.879978] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-8cf153c7-e439-4990-9fa2-f6d040c68c3c tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore1] 84cc9489-2a11-4510-8ad6-13dc41814856/84cc9489-2a11-4510-8ad6-13dc41814856.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1006.880384] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-8cf153c7-e439-4990-9fa2-f6d040c68c3c tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 84cc9489-2a11-4510-8ad6-13dc41814856] Extending root virtual disk to 1048576 {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1006.880677] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c9fdca7d-2925-4b4d-8d66-476a620a66ae {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.889071] env[61957]: DEBUG oslo_vmware.api [None req-8cf153c7-e439-4990-9fa2-f6d040c68c3c tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Waiting for the task: (returnval){ [ 1006.889071] env[61957]: value = "task-1277948" [ 1006.889071] env[61957]: _type = "Task" [ 1006.889071] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1006.893607] env[61957]: DEBUG nova.scheduler.client.report [None req-0b16b4fe-c18a-4e4e-8833-61daba0fafce tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1006.904742] env[61957]: DEBUG oslo_vmware.api [None req-8cf153c7-e439-4990-9fa2-f6d040c68c3c tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': task-1277948, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.962554] env[61957]: DEBUG oslo_vmware.api [None req-ab927ca0-dc5c-48f5-8ef0-a84245f36801 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': task-1277946, 'name': PowerOnVM_Task, 'duration_secs': 0.56653} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.962554] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-ab927ca0-dc5c-48f5-8ef0-a84245f36801 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: e2107d5c-8453-4d24-b364-2b5c1d125062] Powered on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1006.962554] env[61957]: INFO nova.compute.manager [None req-ab927ca0-dc5c-48f5-8ef0-a84245f36801 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: e2107d5c-8453-4d24-b364-2b5c1d125062] Took 8.24 seconds to spawn the instance on the hypervisor. [ 1006.962554] env[61957]: DEBUG nova.compute.manager [None req-ab927ca0-dc5c-48f5-8ef0-a84245f36801 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: e2107d5c-8453-4d24-b364-2b5c1d125062] Checking state {{(pid=61957) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1006.963515] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6296ffbc-1a32-49c6-848f-5bb92aa5164d {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.091796] env[61957]: DEBUG oslo_concurrency.lockutils [None req-728a701e-f15a-4151-a694-52c914bd7a5f tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Acquiring lock "f65458b2-0db6-4d98-8c82-b27b6945a3fd" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1007.092670] env[61957]: DEBUG oslo_concurrency.lockutils [None req-728a701e-f15a-4151-a694-52c914bd7a5f tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Lock "f65458b2-0db6-4d98-8c82-b27b6945a3fd" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1007.092916] env[61957]: DEBUG oslo_concurrency.lockutils [None req-728a701e-f15a-4151-a694-52c914bd7a5f tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Acquiring lock "f65458b2-0db6-4d98-8c82-b27b6945a3fd-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1007.095033] env[61957]: DEBUG oslo_concurrency.lockutils [None req-728a701e-f15a-4151-a694-52c914bd7a5f tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Lock "f65458b2-0db6-4d98-8c82-b27b6945a3fd-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1007.095033] env[61957]: DEBUG oslo_concurrency.lockutils [None req-728a701e-f15a-4151-a694-52c914bd7a5f tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Lock "f65458b2-0db6-4d98-8c82-b27b6945a3fd-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1007.097097] env[61957]: DEBUG oslo_concurrency.lockutils [req-494b367b-4b4f-41ab-88d0-e44d5604dada req-2b1cd18e-1ec7-4ac2-bcb4-fc2785a8b8f1 service nova] Releasing lock "refresh_cache-08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1007.098894] env[61957]: INFO nova.compute.manager [None req-728a701e-f15a-4151-a694-52c914bd7a5f tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] [instance: f65458b2-0db6-4d98-8c82-b27b6945a3fd] Terminating instance [ 1007.100396] env[61957]: DEBUG nova.compute.manager [None req-728a701e-f15a-4151-a694-52c914bd7a5f tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] [instance: f65458b2-0db6-4d98-8c82-b27b6945a3fd] Start destroying the instance on the hypervisor. {{(pid=61957) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1007.100697] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-728a701e-f15a-4151-a694-52c914bd7a5f tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] [instance: f65458b2-0db6-4d98-8c82-b27b6945a3fd] Destroying instance {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1007.101791] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2ca091e-2751-4d26-bc63-9660d23d292b {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.113085] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-728a701e-f15a-4151-a694-52c914bd7a5f tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] [instance: f65458b2-0db6-4d98-8c82-b27b6945a3fd] Powering off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1007.113469] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5b8f5b1a-c6e0-4a44-814f-8fd7ad90c777 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.122961] env[61957]: DEBUG oslo_vmware.api [None req-728a701e-f15a-4151-a694-52c914bd7a5f tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Waiting for the task: (returnval){ [ 1007.122961] env[61957]: value = "task-1277949" [ 1007.122961] env[61957]: _type = "Task" [ 1007.122961] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1007.133636] env[61957]: DEBUG oslo_vmware.api [None req-728a701e-f15a-4151-a694-52c914bd7a5f tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Task: {'id': task-1277949, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.170811] env[61957]: DEBUG nova.compute.manager [None req-d92c7007-db9c-429d-aa03-41f7719669de tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 927af4c3-c7e8-404c-839e-d25af5e384a9] Starting instance... {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1007.311740] env[61957]: DEBUG oslo_concurrency.lockutils [None req-690abcce-5857-41f4-8c3f-00315918c375 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Lock "interface-08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 7.203s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1007.399509] env[61957]: DEBUG oslo_concurrency.lockutils [None req-0b16b4fe-c18a-4e4e-8833-61daba0fafce tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.307s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1007.399850] env[61957]: INFO nova.compute.manager [None req-0b16b4fe-c18a-4e4e-8833-61daba0fafce tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: 6509707e-afc6-4c4a-88f9-a471865d510c] Migrating [ 1007.410888] env[61957]: DEBUG oslo_vmware.api [None req-8cf153c7-e439-4990-9fa2-f6d040c68c3c tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': task-1277948, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.148336} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1007.415587] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-8cf153c7-e439-4990-9fa2-f6d040c68c3c tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 84cc9489-2a11-4510-8ad6-13dc41814856] Extended root virtual disk {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1007.421670] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cee2e140-0c49-457c-96f4-d73c9ab74f1d {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.447996] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-8cf153c7-e439-4990-9fa2-f6d040c68c3c tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 84cc9489-2a11-4510-8ad6-13dc41814856] Reconfiguring VM instance instance-0000005f to attach disk [datastore1] 84cc9489-2a11-4510-8ad6-13dc41814856/84cc9489-2a11-4510-8ad6-13dc41814856.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1007.449507] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4ddc7d38-7917-4540-839f-f12848313d43 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.472560] env[61957]: DEBUG oslo_vmware.api [None req-8cf153c7-e439-4990-9fa2-f6d040c68c3c tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Waiting for the task: (returnval){ [ 1007.472560] env[61957]: value = "task-1277950" [ 1007.472560] env[61957]: _type = "Task" [ 1007.472560] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1007.486858] env[61957]: DEBUG oslo_vmware.api [None req-8cf153c7-e439-4990-9fa2-f6d040c68c3c tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': task-1277950, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.487948] env[61957]: INFO nova.compute.manager [None req-ab927ca0-dc5c-48f5-8ef0-a84245f36801 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: e2107d5c-8453-4d24-b364-2b5c1d125062] Took 13.15 seconds to build instance. [ 1007.633690] env[61957]: DEBUG oslo_vmware.api [None req-728a701e-f15a-4151-a694-52c914bd7a5f tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Task: {'id': task-1277949, 'name': PowerOffVM_Task, 'duration_secs': 0.293115} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1007.634568] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-728a701e-f15a-4151-a694-52c914bd7a5f tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] [instance: f65458b2-0db6-4d98-8c82-b27b6945a3fd] Powered off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1007.634772] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-728a701e-f15a-4151-a694-52c914bd7a5f tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] [instance: f65458b2-0db6-4d98-8c82-b27b6945a3fd] Unregistering the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1007.635187] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-be0ff14c-e2da-4b27-bf95-c99fadcc229f {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.693724] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d92c7007-db9c-429d-aa03-41f7719669de tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1007.694225] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d92c7007-db9c-429d-aa03-41f7719669de tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1007.695669] env[61957]: INFO nova.compute.claims [None req-d92c7007-db9c-429d-aa03-41f7719669de tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 927af4c3-c7e8-404c-839e-d25af5e384a9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1007.718404] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-728a701e-f15a-4151-a694-52c914bd7a5f tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] [instance: f65458b2-0db6-4d98-8c82-b27b6945a3fd] Unregistered the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1007.718655] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-728a701e-f15a-4151-a694-52c914bd7a5f tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] [instance: f65458b2-0db6-4d98-8c82-b27b6945a3fd] Deleting contents of the VM from datastore datastore1 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1007.719193] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-728a701e-f15a-4151-a694-52c914bd7a5f tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Deleting the datastore file [datastore1] f65458b2-0db6-4d98-8c82-b27b6945a3fd {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1007.719489] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e84b9b8f-1b18-4bae-b23e-d43c9d2a9af8 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.731583] env[61957]: DEBUG oslo_vmware.api [None req-728a701e-f15a-4151-a694-52c914bd7a5f tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Waiting for the task: (returnval){ [ 1007.731583] env[61957]: value = "task-1277952" [ 1007.731583] env[61957]: _type = "Task" [ 1007.731583] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1007.741132] env[61957]: DEBUG oslo_vmware.api [None req-728a701e-f15a-4151-a694-52c914bd7a5f tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Task: {'id': task-1277952, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.924080] env[61957]: DEBUG oslo_concurrency.lockutils [None req-0b16b4fe-c18a-4e4e-8833-61daba0fafce tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Acquiring lock "refresh_cache-6509707e-afc6-4c4a-88f9-a471865d510c" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1007.924255] env[61957]: DEBUG oslo_concurrency.lockutils [None req-0b16b4fe-c18a-4e4e-8833-61daba0fafce tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Acquired lock "refresh_cache-6509707e-afc6-4c4a-88f9-a471865d510c" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1007.924467] env[61957]: DEBUG nova.network.neutron [None req-0b16b4fe-c18a-4e4e-8833-61daba0fafce tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: 6509707e-afc6-4c4a-88f9-a471865d510c] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1007.984650] env[61957]: DEBUG oslo_vmware.api [None req-8cf153c7-e439-4990-9fa2-f6d040c68c3c tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': task-1277950, 'name': ReconfigVM_Task, 'duration_secs': 0.316558} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1007.984965] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-8cf153c7-e439-4990-9fa2-f6d040c68c3c tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 84cc9489-2a11-4510-8ad6-13dc41814856] Reconfigured VM instance instance-0000005f to attach disk [datastore1] 84cc9489-2a11-4510-8ad6-13dc41814856/84cc9489-2a11-4510-8ad6-13dc41814856.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1007.985669] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f8e221e6-5910-4a0a-97cb-ace4f424511a {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.994808] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ab927ca0-dc5c-48f5-8ef0-a84245f36801 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Lock "e2107d5c-8453-4d24-b364-2b5c1d125062" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.664s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1007.995208] env[61957]: DEBUG oslo_vmware.api [None req-8cf153c7-e439-4990-9fa2-f6d040c68c3c tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Waiting for the task: (returnval){ [ 1007.995208] env[61957]: value = "task-1277953" [ 1007.995208] env[61957]: _type = "Task" [ 1007.995208] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1008.010948] env[61957]: DEBUG oslo_vmware.api [None req-8cf153c7-e439-4990-9fa2-f6d040c68c3c tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': task-1277953, 'name': Rename_Task} progress is 5%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.243105] env[61957]: DEBUG oslo_vmware.api [None req-728a701e-f15a-4151-a694-52c914bd7a5f tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Task: {'id': task-1277952, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.284208} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1008.243372] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-728a701e-f15a-4151-a694-52c914bd7a5f tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Deleted the datastore file {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1008.243519] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-728a701e-f15a-4151-a694-52c914bd7a5f tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] [instance: f65458b2-0db6-4d98-8c82-b27b6945a3fd] Deleted contents of the VM from datastore datastore1 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1008.243756] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-728a701e-f15a-4151-a694-52c914bd7a5f tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] [instance: f65458b2-0db6-4d98-8c82-b27b6945a3fd] Instance destroyed {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1008.243940] env[61957]: INFO nova.compute.manager [None req-728a701e-f15a-4151-a694-52c914bd7a5f tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] [instance: f65458b2-0db6-4d98-8c82-b27b6945a3fd] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1008.244260] env[61957]: DEBUG oslo.service.loopingcall [None req-728a701e-f15a-4151-a694-52c914bd7a5f tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1008.244476] env[61957]: DEBUG nova.compute.manager [-] [instance: f65458b2-0db6-4d98-8c82-b27b6945a3fd] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1008.244564] env[61957]: DEBUG nova.network.neutron [-] [instance: f65458b2-0db6-4d98-8c82-b27b6945a3fd] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1008.510441] env[61957]: DEBUG oslo_vmware.api [None req-8cf153c7-e439-4990-9fa2-f6d040c68c3c tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': task-1277953, 'name': Rename_Task, 'duration_secs': 0.167729} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1008.510926] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-8cf153c7-e439-4990-9fa2-f6d040c68c3c tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 84cc9489-2a11-4510-8ad6-13dc41814856] Powering on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1008.511312] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8e7b44c9-1be5-4de5-8799-be908e1380e3 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.521618] env[61957]: DEBUG oslo_vmware.api [None req-8cf153c7-e439-4990-9fa2-f6d040c68c3c tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Waiting for the task: (returnval){ [ 1008.521618] env[61957]: value = "task-1277954" [ 1008.521618] env[61957]: _type = "Task" [ 1008.521618] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1008.535394] env[61957]: DEBUG oslo_vmware.api [None req-8cf153c7-e439-4990-9fa2-f6d040c68c3c tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': task-1277954, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.603842] env[61957]: DEBUG nova.compute.manager [req-1fd9bfac-b05d-409f-8855-2de33115843c req-79e85f1a-b888-483d-b760-faf519fb911a service nova] [instance: f65458b2-0db6-4d98-8c82-b27b6945a3fd] Received event network-vif-deleted-bb715e7d-14c5-418b-b3ee-1ac91aba709d {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1008.604113] env[61957]: INFO nova.compute.manager [req-1fd9bfac-b05d-409f-8855-2de33115843c req-79e85f1a-b888-483d-b760-faf519fb911a service nova] [instance: f65458b2-0db6-4d98-8c82-b27b6945a3fd] Neutron deleted interface bb715e7d-14c5-418b-b3ee-1ac91aba709d; detaching it from the instance and deleting it from the info cache [ 1008.604344] env[61957]: DEBUG nova.network.neutron [req-1fd9bfac-b05d-409f-8855-2de33115843c req-79e85f1a-b888-483d-b760-faf519fb911a service nova] [instance: f65458b2-0db6-4d98-8c82-b27b6945a3fd] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1008.710471] env[61957]: DEBUG nova.network.neutron [None req-0b16b4fe-c18a-4e4e-8833-61daba0fafce tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: 6509707e-afc6-4c4a-88f9-a471865d510c] Updating instance_info_cache with network_info: [{"id": "7df2736b-10d0-4f9d-83aa-e2bbdf1ef45b", "address": "fa:16:3e:2e:da:a9", "network": {"id": "6fbccc17-6674-4cd5-8f1b-aa73570baac7", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-394544811-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1dbdc699741a48af93c52068d88a5357", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "20e3f794-c7a3-4696-9488-ecf34c570ef9", "external-id": "nsx-vlan-transportzone-509", "segmentation_id": 509, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7df2736b-10", "ovs_interfaceid": "7df2736b-10d0-4f9d-83aa-e2bbdf1ef45b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1008.816661] env[61957]: DEBUG nova.compute.manager [None req-d6cf550f-30ee-4518-a1a7-00d815bdcf78 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: e2107d5c-8453-4d24-b364-2b5c1d125062] Checking state {{(pid=61957) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1008.817832] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3fc64ed-413c-4711-82e7-842b31ba9585 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.905639] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28161bf6-4ce2-4f33-8ce8-7c3e7c411516 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.914368] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84646bb7-e1fd-411f-9a35-8af18f02c063 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.957379] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-437b2c7b-1c66-4cbb-96b1-92b7333659ce {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.966235] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e04f800-6ba1-431f-9b1d-a608be196260 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.981702] env[61957]: DEBUG nova.compute.provider_tree [None req-d92c7007-db9c-429d-aa03-41f7719669de tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1009.034112] env[61957]: DEBUG oslo_vmware.api [None req-8cf153c7-e439-4990-9fa2-f6d040c68c3c tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': task-1277954, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.079718] env[61957]: DEBUG nova.network.neutron [-] [instance: f65458b2-0db6-4d98-8c82-b27b6945a3fd] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1009.112029] env[61957]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-71be540f-98d7-4a9d-ad0a-0f8028cb9bba {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.121527] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecb0e388-4599-4700-be83-b0ea17a672af {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.156807] env[61957]: DEBUG nova.compute.manager [req-1fd9bfac-b05d-409f-8855-2de33115843c req-79e85f1a-b888-483d-b760-faf519fb911a service nova] [instance: f65458b2-0db6-4d98-8c82-b27b6945a3fd] Detach interface failed, port_id=bb715e7d-14c5-418b-b3ee-1ac91aba709d, reason: Instance f65458b2-0db6-4d98-8c82-b27b6945a3fd could not be found. {{(pid=61957) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1009.212741] env[61957]: DEBUG oslo_concurrency.lockutils [None req-0b16b4fe-c18a-4e4e-8833-61daba0fafce tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Releasing lock "refresh_cache-6509707e-afc6-4c4a-88f9-a471865d510c" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1009.333868] env[61957]: INFO nova.compute.manager [None req-d6cf550f-30ee-4518-a1a7-00d815bdcf78 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: e2107d5c-8453-4d24-b364-2b5c1d125062] instance snapshotting [ 1009.337238] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92db8e3e-38cf-4b63-9eb9-59ffcfd8bc1f {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.363167] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64d4a593-2bd9-4d62-8f45-a0dcf77d7c31 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.488140] env[61957]: DEBUG nova.scheduler.client.report [None req-d92c7007-db9c-429d-aa03-41f7719669de tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1009.492921] env[61957]: DEBUG oslo_concurrency.lockutils [None req-3409aaf8-ffe7-41b4-a265-d651c9473b75 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Acquiring lock "interface-08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2-00c8220d-5633-4de7-9795-8a6a8543c4ec" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1009.493148] env[61957]: DEBUG oslo_concurrency.lockutils [None req-3409aaf8-ffe7-41b4-a265-d651c9473b75 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Lock "interface-08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2-00c8220d-5633-4de7-9795-8a6a8543c4ec" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1009.493361] env[61957]: DEBUG nova.objects.instance [None req-3409aaf8-ffe7-41b4-a265-d651c9473b75 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Lazy-loading 'flavor' on Instance uuid 08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2 {{(pid=61957) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1009.535136] env[61957]: DEBUG oslo_vmware.api [None req-8cf153c7-e439-4990-9fa2-f6d040c68c3c tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': task-1277954, 'name': PowerOnVM_Task, 'duration_secs': 0.980821} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.535317] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-8cf153c7-e439-4990-9fa2-f6d040c68c3c tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 84cc9489-2a11-4510-8ad6-13dc41814856] Powered on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1009.535555] env[61957]: INFO nova.compute.manager [None req-8cf153c7-e439-4990-9fa2-f6d040c68c3c tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 84cc9489-2a11-4510-8ad6-13dc41814856] Took 8.44 seconds to spawn the instance on the hypervisor. [ 1009.535772] env[61957]: DEBUG nova.compute.manager [None req-8cf153c7-e439-4990-9fa2-f6d040c68c3c tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 84cc9489-2a11-4510-8ad6-13dc41814856] Checking state {{(pid=61957) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1009.536840] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6a8c271-8f89-4376-a335-1286f4d4e1af {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.583068] env[61957]: INFO nova.compute.manager [-] [instance: f65458b2-0db6-4d98-8c82-b27b6945a3fd] Took 1.34 seconds to deallocate network for instance. [ 1009.876257] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-d6cf550f-30ee-4518-a1a7-00d815bdcf78 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: e2107d5c-8453-4d24-b364-2b5c1d125062] Creating Snapshot of the VM instance {{(pid=61957) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1009.876742] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-0abc5bff-3a8e-46c5-8908-78136a802343 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.886790] env[61957]: DEBUG oslo_vmware.api [None req-d6cf550f-30ee-4518-a1a7-00d815bdcf78 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Waiting for the task: (returnval){ [ 1009.886790] env[61957]: value = "task-1277955" [ 1009.886790] env[61957]: _type = "Task" [ 1009.886790] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1009.897117] env[61957]: DEBUG oslo_vmware.api [None req-d6cf550f-30ee-4518-a1a7-00d815bdcf78 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': task-1277955, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.996472] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d92c7007-db9c-429d-aa03-41f7719669de tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.302s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1009.997056] env[61957]: DEBUG nova.compute.manager [None req-d92c7007-db9c-429d-aa03-41f7719669de tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 927af4c3-c7e8-404c-839e-d25af5e384a9] Start building networks asynchronously for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1010.059397] env[61957]: INFO nova.compute.manager [None req-8cf153c7-e439-4990-9fa2-f6d040c68c3c tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 84cc9489-2a11-4510-8ad6-13dc41814856] Took 15.06 seconds to build instance. [ 1010.090582] env[61957]: DEBUG oslo_concurrency.lockutils [None req-728a701e-f15a-4151-a694-52c914bd7a5f tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1010.090848] env[61957]: DEBUG oslo_concurrency.lockutils [None req-728a701e-f15a-4151-a694-52c914bd7a5f tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1010.091132] env[61957]: DEBUG nova.objects.instance [None req-728a701e-f15a-4151-a694-52c914bd7a5f tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Lazy-loading 'resources' on Instance uuid f65458b2-0db6-4d98-8c82-b27b6945a3fd {{(pid=61957) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1010.104089] env[61957]: DEBUG nova.objects.instance [None req-3409aaf8-ffe7-41b4-a265-d651c9473b75 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Lazy-loading 'pci_requests' on Instance uuid 08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2 {{(pid=61957) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1010.398064] env[61957]: DEBUG oslo_vmware.api [None req-d6cf550f-30ee-4518-a1a7-00d815bdcf78 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': task-1277955, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.505594] env[61957]: DEBUG nova.compute.utils [None req-d92c7007-db9c-429d-aa03-41f7719669de tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Using /dev/sd instead of None {{(pid=61957) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1010.507245] env[61957]: DEBUG nova.compute.manager [None req-d92c7007-db9c-429d-aa03-41f7719669de tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 927af4c3-c7e8-404c-839e-d25af5e384a9] Allocating IP information in the background. {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1010.507456] env[61957]: DEBUG nova.network.neutron [None req-d92c7007-db9c-429d-aa03-41f7719669de tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 927af4c3-c7e8-404c-839e-d25af5e384a9] allocate_for_instance() {{(pid=61957) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1010.549323] env[61957]: DEBUG nova.policy [None req-d92c7007-db9c-429d-aa03-41f7719669de tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '85d4f8bb3aa64e7285aac42d14087ff1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e16f6dce3f0e44fb96de516f17d4c6f5', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61957) authorize /opt/stack/nova/nova/policy.py:203}} [ 1010.561082] env[61957]: DEBUG oslo_concurrency.lockutils [None req-8cf153c7-e439-4990-9fa2-f6d040c68c3c tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Lock "84cc9489-2a11-4510-8ad6-13dc41814856" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.574s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1010.606607] env[61957]: DEBUG nova.objects.base [None req-3409aaf8-ffe7-41b4-a265-d651c9473b75 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Object Instance<08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2> lazy-loaded attributes: flavor,pci_requests {{(pid=61957) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1010.606908] env[61957]: DEBUG nova.network.neutron [None req-3409aaf8-ffe7-41b4-a265-d651c9473b75 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2] allocate_for_instance() {{(pid=61957) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1010.729505] env[61957]: DEBUG nova.policy [None req-3409aaf8-ffe7-41b4-a265-d651c9473b75 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6fd1c19dc3a44212ada44445e0919106', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8713aa35bcb24b86ad0b58ca9fc991ba', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61957) authorize /opt/stack/nova/nova/policy.py:203}} [ 1010.733207] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd290a69-9437-4172-8490-45c90d55150c {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.752703] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-0b16b4fe-c18a-4e4e-8833-61daba0fafce tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: 6509707e-afc6-4c4a-88f9-a471865d510c] Updating instance '6509707e-afc6-4c4a-88f9-a471865d510c' progress to 0 {{(pid=61957) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1010.822211] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffd7211e-897d-47e2-8c93-57cf5913fe34 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.829633] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-453cefc3-4e7f-4824-9e8c-139ba9cb5852 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.866972] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59f77e3a-16d0-4edb-ad88-2cc5b9c497c0 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.875728] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae766ea6-993c-43f8-97e9-1574831d52a0 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.894533] env[61957]: DEBUG nova.compute.provider_tree [None req-728a701e-f15a-4151-a694-52c914bd7a5f tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1010.905577] env[61957]: DEBUG oslo_vmware.api [None req-d6cf550f-30ee-4518-a1a7-00d815bdcf78 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': task-1277955, 'name': CreateSnapshot_Task, 'duration_secs': 0.866817} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1010.906494] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-d6cf550f-30ee-4518-a1a7-00d815bdcf78 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: e2107d5c-8453-4d24-b364-2b5c1d125062] Created Snapshot of the VM instance {{(pid=61957) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1010.907319] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-637943da-2c64-4374-92d2-9c0567646400 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.010943] env[61957]: DEBUG nova.compute.manager [None req-d92c7007-db9c-429d-aa03-41f7719669de tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 927af4c3-c7e8-404c-839e-d25af5e384a9] Start building block device mappings for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1011.084110] env[61957]: DEBUG nova.network.neutron [None req-d92c7007-db9c-429d-aa03-41f7719669de tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 927af4c3-c7e8-404c-839e-d25af5e384a9] Successfully created port: 38da1ba9-c5cc-4f76-b516-5ab8fdca9064 {{(pid=61957) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1011.262652] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b16b4fe-c18a-4e4e-8833-61daba0fafce tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: 6509707e-afc6-4c4a-88f9-a471865d510c] Powering off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1011.262652] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2ff3c1c1-534c-4eb6-b54c-4e038035b78c {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.275020] env[61957]: DEBUG oslo_vmware.api [None req-0b16b4fe-c18a-4e4e-8833-61daba0fafce tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Waiting for the task: (returnval){ [ 1011.275020] env[61957]: value = "task-1277956" [ 1011.275020] env[61957]: _type = "Task" [ 1011.275020] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.282058] env[61957]: DEBUG oslo_vmware.api [None req-0b16b4fe-c18a-4e4e-8833-61daba0fafce tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': task-1277956, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.319817] env[61957]: DEBUG oslo_concurrency.lockutils [None req-6a1be883-2588-47a0-91f4-274d9905657e tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Acquiring lock "3f144012-c0cc-43ce-8c73-4422fe8c1fe6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1011.320199] env[61957]: DEBUG oslo_concurrency.lockutils [None req-6a1be883-2588-47a0-91f4-274d9905657e tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Lock "3f144012-c0cc-43ce-8c73-4422fe8c1fe6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1011.400719] env[61957]: DEBUG nova.scheduler.client.report [None req-728a701e-f15a-4151-a694-52c914bd7a5f tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1011.427203] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-d6cf550f-30ee-4518-a1a7-00d815bdcf78 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: e2107d5c-8453-4d24-b364-2b5c1d125062] Creating linked-clone VM from snapshot {{(pid=61957) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1011.428053] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-6c09e2a6-a47a-4ac2-b2de-97c745c9161f {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.438835] env[61957]: DEBUG oslo_vmware.api [None req-d6cf550f-30ee-4518-a1a7-00d815bdcf78 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Waiting for the task: (returnval){ [ 1011.438835] env[61957]: value = "task-1277957" [ 1011.438835] env[61957]: _type = "Task" [ 1011.438835] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.449259] env[61957]: DEBUG oslo_vmware.api [None req-d6cf550f-30ee-4518-a1a7-00d815bdcf78 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': task-1277957, 'name': CloneVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.783355] env[61957]: DEBUG oslo_vmware.api [None req-0b16b4fe-c18a-4e4e-8833-61daba0fafce tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': task-1277956, 'name': PowerOffVM_Task, 'duration_secs': 0.209812} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.783745] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b16b4fe-c18a-4e4e-8833-61daba0fafce tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: 6509707e-afc6-4c4a-88f9-a471865d510c] Powered off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1011.784057] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-0b16b4fe-c18a-4e4e-8833-61daba0fafce tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: 6509707e-afc6-4c4a-88f9-a471865d510c] Updating instance '6509707e-afc6-4c4a-88f9-a471865d510c' progress to 17 {{(pid=61957) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1011.822765] env[61957]: DEBUG nova.compute.manager [None req-6a1be883-2588-47a0-91f4-274d9905657e tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 3f144012-c0cc-43ce-8c73-4422fe8c1fe6] Starting instance... {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1011.906739] env[61957]: DEBUG oslo_concurrency.lockutils [None req-728a701e-f15a-4151-a694-52c914bd7a5f tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.816s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1011.936134] env[61957]: INFO nova.scheduler.client.report [None req-728a701e-f15a-4151-a694-52c914bd7a5f tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Deleted allocations for instance f65458b2-0db6-4d98-8c82-b27b6945a3fd [ 1011.953200] env[61957]: DEBUG oslo_vmware.api [None req-d6cf550f-30ee-4518-a1a7-00d815bdcf78 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': task-1277957, 'name': CloneVM_Task} progress is 94%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.019979] env[61957]: DEBUG nova.compute.manager [None req-d92c7007-db9c-429d-aa03-41f7719669de tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 927af4c3-c7e8-404c-839e-d25af5e384a9] Start spawning the instance on the hypervisor. {{(pid=61957) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1012.047564] env[61957]: DEBUG nova.virt.hardware [None req-d92c7007-db9c-429d-aa03-41f7719669de tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T17:22:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T17:21:50Z,direct_url=,disk_format='vmdk',id=11c76a2c-f705-470a-ba9d-4657858bab38,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='81d3096d0f094373913fc3dc8f5c3c6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T17:21:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1012.047851] env[61957]: DEBUG nova.virt.hardware [None req-d92c7007-db9c-429d-aa03-41f7719669de tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Flavor limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1012.048045] env[61957]: DEBUG nova.virt.hardware [None req-d92c7007-db9c-429d-aa03-41f7719669de tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Image limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1012.048287] env[61957]: DEBUG nova.virt.hardware [None req-d92c7007-db9c-429d-aa03-41f7719669de tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Flavor pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1012.048465] env[61957]: DEBUG nova.virt.hardware [None req-d92c7007-db9c-429d-aa03-41f7719669de tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Image pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1012.048603] env[61957]: DEBUG nova.virt.hardware [None req-d92c7007-db9c-429d-aa03-41f7719669de tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1012.048924] env[61957]: DEBUG nova.virt.hardware [None req-d92c7007-db9c-429d-aa03-41f7719669de tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1012.049131] env[61957]: DEBUG nova.virt.hardware [None req-d92c7007-db9c-429d-aa03-41f7719669de tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1012.049328] env[61957]: DEBUG nova.virt.hardware [None req-d92c7007-db9c-429d-aa03-41f7719669de tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Got 1 possible topologies {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1012.049521] env[61957]: DEBUG nova.virt.hardware [None req-d92c7007-db9c-429d-aa03-41f7719669de tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1012.049750] env[61957]: DEBUG nova.virt.hardware [None req-d92c7007-db9c-429d-aa03-41f7719669de tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1012.051266] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a0568ef-4b4a-434a-a0b9-8bad34ab5357 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.060580] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32b7ebb3-5882-4550-b8ae-444bf0a81c82 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.291729] env[61957]: DEBUG nova.virt.hardware [None req-0b16b4fe-c18a-4e4e-8833-61daba0fafce tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T17:22:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=11c76a2c-f705-470a-ba9d-4657858bab38,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1012.291989] env[61957]: DEBUG nova.virt.hardware [None req-0b16b4fe-c18a-4e4e-8833-61daba0fafce tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Flavor limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1012.292258] env[61957]: DEBUG nova.virt.hardware [None req-0b16b4fe-c18a-4e4e-8833-61daba0fafce tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Image limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1012.292478] env[61957]: DEBUG nova.virt.hardware [None req-0b16b4fe-c18a-4e4e-8833-61daba0fafce tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Flavor pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1012.292635] env[61957]: DEBUG nova.virt.hardware [None req-0b16b4fe-c18a-4e4e-8833-61daba0fafce tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Image pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1012.292831] env[61957]: DEBUG nova.virt.hardware [None req-0b16b4fe-c18a-4e4e-8833-61daba0fafce tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1012.293115] env[61957]: DEBUG nova.virt.hardware [None req-0b16b4fe-c18a-4e4e-8833-61daba0fafce tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1012.293356] env[61957]: DEBUG nova.virt.hardware [None req-0b16b4fe-c18a-4e4e-8833-61daba0fafce tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1012.293575] env[61957]: DEBUG nova.virt.hardware [None req-0b16b4fe-c18a-4e4e-8833-61daba0fafce tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Got 1 possible topologies {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1012.293831] env[61957]: DEBUG nova.virt.hardware [None req-0b16b4fe-c18a-4e4e-8833-61daba0fafce tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1012.294032] env[61957]: DEBUG nova.virt.hardware [None req-0b16b4fe-c18a-4e4e-8833-61daba0fafce tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1012.299997] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-302df159-8484-45f1-a2fc-4aa08636d8ca {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.321099] env[61957]: DEBUG oslo_vmware.api [None req-0b16b4fe-c18a-4e4e-8833-61daba0fafce tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Waiting for the task: (returnval){ [ 1012.321099] env[61957]: value = "task-1277958" [ 1012.321099] env[61957]: _type = "Task" [ 1012.321099] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.333444] env[61957]: DEBUG oslo_vmware.api [None req-0b16b4fe-c18a-4e4e-8833-61daba0fafce tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': task-1277958, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.347354] env[61957]: DEBUG oslo_concurrency.lockutils [None req-6a1be883-2588-47a0-91f4-274d9905657e tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1012.347675] env[61957]: DEBUG oslo_concurrency.lockutils [None req-6a1be883-2588-47a0-91f4-274d9905657e tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1012.349429] env[61957]: INFO nova.compute.claims [None req-6a1be883-2588-47a0-91f4-274d9905657e tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 3f144012-c0cc-43ce-8c73-4422fe8c1fe6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1012.449920] env[61957]: DEBUG oslo_concurrency.lockutils [None req-728a701e-f15a-4151-a694-52c914bd7a5f tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Lock "f65458b2-0db6-4d98-8c82-b27b6945a3fd" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.357s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1012.456072] env[61957]: DEBUG oslo_vmware.api [None req-d6cf550f-30ee-4518-a1a7-00d815bdcf78 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': task-1277957, 'name': CloneVM_Task} progress is 94%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.498937] env[61957]: DEBUG nova.compute.manager [req-0971f1e6-5854-4732-94f5-920bfc5b5339 req-f569340a-49da-4349-b90c-b04c4b060dd4 service nova] [instance: 08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2] Received event network-vif-plugged-00c8220d-5633-4de7-9795-8a6a8543c4ec {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1012.498937] env[61957]: DEBUG oslo_concurrency.lockutils [req-0971f1e6-5854-4732-94f5-920bfc5b5339 req-f569340a-49da-4349-b90c-b04c4b060dd4 service nova] Acquiring lock "08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1012.498937] env[61957]: DEBUG oslo_concurrency.lockutils [req-0971f1e6-5854-4732-94f5-920bfc5b5339 req-f569340a-49da-4349-b90c-b04c4b060dd4 service nova] Lock "08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1012.499175] env[61957]: DEBUG oslo_concurrency.lockutils [req-0971f1e6-5854-4732-94f5-920bfc5b5339 req-f569340a-49da-4349-b90c-b04c4b060dd4 service nova] Lock "08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1012.499217] env[61957]: DEBUG nova.compute.manager [req-0971f1e6-5854-4732-94f5-920bfc5b5339 req-f569340a-49da-4349-b90c-b04c4b060dd4 service nova] [instance: 08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2] No waiting events found dispatching network-vif-plugged-00c8220d-5633-4de7-9795-8a6a8543c4ec {{(pid=61957) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1012.499387] env[61957]: WARNING nova.compute.manager [req-0971f1e6-5854-4732-94f5-920bfc5b5339 req-f569340a-49da-4349-b90c-b04c4b060dd4 service nova] [instance: 08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2] Received unexpected event network-vif-plugged-00c8220d-5633-4de7-9795-8a6a8543c4ec for instance with vm_state active and task_state None. [ 1012.686515] env[61957]: DEBUG nova.network.neutron [None req-3409aaf8-ffe7-41b4-a265-d651c9473b75 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2] Successfully updated port: 00c8220d-5633-4de7-9795-8a6a8543c4ec {{(pid=61957) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1012.832971] env[61957]: DEBUG oslo_vmware.api [None req-0b16b4fe-c18a-4e4e-8833-61daba0fafce tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': task-1277958, 'name': ReconfigVM_Task, 'duration_secs': 0.387482} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1012.833329] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-0b16b4fe-c18a-4e4e-8833-61daba0fafce tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: 6509707e-afc6-4c4a-88f9-a471865d510c] Updating instance '6509707e-afc6-4c4a-88f9-a471865d510c' progress to 33 {{(pid=61957) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1012.844716] env[61957]: DEBUG nova.network.neutron [None req-d92c7007-db9c-429d-aa03-41f7719669de tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 927af4c3-c7e8-404c-839e-d25af5e384a9] Successfully updated port: 38da1ba9-c5cc-4f76-b516-5ab8fdca9064 {{(pid=61957) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1012.950736] env[61957]: DEBUG oslo_vmware.api [None req-d6cf550f-30ee-4518-a1a7-00d815bdcf78 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': task-1277957, 'name': CloneVM_Task} progress is 100%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.021720] env[61957]: DEBUG oslo_vmware.rw_handles [None req-ad83dc8d-b942-4b5a-8561-778a1542e9de tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52b02e8b-f180-7868-5405-b594f5954d48/disk-0.vmdk. {{(pid=61957) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1013.022839] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23179a88-6fdd-4317-87d6-6a1dfdc0671e {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.029424] env[61957]: DEBUG oslo_vmware.rw_handles [None req-ad83dc8d-b942-4b5a-8561-778a1542e9de tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52b02e8b-f180-7868-5405-b594f5954d48/disk-0.vmdk is in state: ready. {{(pid=61957) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1013.029609] env[61957]: ERROR oslo_vmware.rw_handles [None req-ad83dc8d-b942-4b5a-8561-778a1542e9de tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52b02e8b-f180-7868-5405-b594f5954d48/disk-0.vmdk due to incomplete transfer. [ 1013.029833] env[61957]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-72244c7a-f01a-4fae-95a7-1194bac68658 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.036983] env[61957]: DEBUG oslo_vmware.rw_handles [None req-ad83dc8d-b942-4b5a-8561-778a1542e9de tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52b02e8b-f180-7868-5405-b594f5954d48/disk-0.vmdk. {{(pid=61957) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1013.037197] env[61957]: DEBUG nova.virt.vmwareapi.images [None req-ad83dc8d-b942-4b5a-8561-778a1542e9de tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] [instance: 9b0053fe-420e-4b92-86a8-ed44f6a6ec49] Uploaded image f04c1641-6ff1-4b2c-9748-9cf176307db0 to the Glance image server {{(pid=61957) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 1013.039590] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-ad83dc8d-b942-4b5a-8561-778a1542e9de tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] [instance: 9b0053fe-420e-4b92-86a8-ed44f6a6ec49] Destroying the VM {{(pid=61957) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 1013.039870] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-19fa72f8-719c-428a-ba09-f50be0b33198 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.046629] env[61957]: DEBUG oslo_vmware.api [None req-ad83dc8d-b942-4b5a-8561-778a1542e9de tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Waiting for the task: (returnval){ [ 1013.046629] env[61957]: value = "task-1277959" [ 1013.046629] env[61957]: _type = "Task" [ 1013.046629] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.057060] env[61957]: DEBUG oslo_vmware.api [None req-ad83dc8d-b942-4b5a-8561-778a1542e9de tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Task: {'id': task-1277959, 'name': Destroy_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.087757] env[61957]: DEBUG oslo_concurrency.lockutils [None req-1238a3ba-00b0-4ef8-9842-11dbb65e047b tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Acquiring lock "7793aa07-6aa2-459d-8a91-56c6b0412d3b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1013.088071] env[61957]: DEBUG oslo_concurrency.lockutils [None req-1238a3ba-00b0-4ef8-9842-11dbb65e047b tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Lock "7793aa07-6aa2-459d-8a91-56c6b0412d3b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1013.088323] env[61957]: DEBUG oslo_concurrency.lockutils [None req-1238a3ba-00b0-4ef8-9842-11dbb65e047b tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Acquiring lock "7793aa07-6aa2-459d-8a91-56c6b0412d3b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1013.088574] env[61957]: DEBUG oslo_concurrency.lockutils [None req-1238a3ba-00b0-4ef8-9842-11dbb65e047b tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Lock "7793aa07-6aa2-459d-8a91-56c6b0412d3b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1013.088829] env[61957]: DEBUG oslo_concurrency.lockutils [None req-1238a3ba-00b0-4ef8-9842-11dbb65e047b tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Lock "7793aa07-6aa2-459d-8a91-56c6b0412d3b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1013.091149] env[61957]: INFO nova.compute.manager [None req-1238a3ba-00b0-4ef8-9842-11dbb65e047b tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] [instance: 7793aa07-6aa2-459d-8a91-56c6b0412d3b] Terminating instance [ 1013.093150] env[61957]: DEBUG nova.compute.manager [None req-1238a3ba-00b0-4ef8-9842-11dbb65e047b tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] [instance: 7793aa07-6aa2-459d-8a91-56c6b0412d3b] Start destroying the instance on the hypervisor. {{(pid=61957) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1013.093380] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-1238a3ba-00b0-4ef8-9842-11dbb65e047b tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] [instance: 7793aa07-6aa2-459d-8a91-56c6b0412d3b] Destroying instance {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1013.094311] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ca30b9d-2e5d-44d0-964b-51dfe2247ab0 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.102250] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-1238a3ba-00b0-4ef8-9842-11dbb65e047b tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] [instance: 7793aa07-6aa2-459d-8a91-56c6b0412d3b] Powering off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1013.102857] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-47ab7d33-84a2-4ef0-a2d2-832d3bb22185 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.109759] env[61957]: DEBUG oslo_vmware.api [None req-1238a3ba-00b0-4ef8-9842-11dbb65e047b tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Waiting for the task: (returnval){ [ 1013.109759] env[61957]: value = "task-1277960" [ 1013.109759] env[61957]: _type = "Task" [ 1013.109759] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.117891] env[61957]: DEBUG oslo_vmware.api [None req-1238a3ba-00b0-4ef8-9842-11dbb65e047b tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Task: {'id': task-1277960, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.189581] env[61957]: DEBUG oslo_concurrency.lockutils [None req-3409aaf8-ffe7-41b4-a265-d651c9473b75 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Acquiring lock "refresh_cache-08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1013.189842] env[61957]: DEBUG oslo_concurrency.lockutils [None req-3409aaf8-ffe7-41b4-a265-d651c9473b75 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Acquired lock "refresh_cache-08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1013.190054] env[61957]: DEBUG nova.network.neutron [None req-3409aaf8-ffe7-41b4-a265-d651c9473b75 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1013.340646] env[61957]: DEBUG nova.virt.hardware [None req-0b16b4fe-c18a-4e4e-8833-61daba0fafce tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T17:22:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=11c76a2c-f705-470a-ba9d-4657858bab38,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1013.340984] env[61957]: DEBUG nova.virt.hardware [None req-0b16b4fe-c18a-4e4e-8833-61daba0fafce tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Flavor limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1013.341267] env[61957]: DEBUG nova.virt.hardware [None req-0b16b4fe-c18a-4e4e-8833-61daba0fafce tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Image limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1013.341567] env[61957]: DEBUG nova.virt.hardware [None req-0b16b4fe-c18a-4e4e-8833-61daba0fafce tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Flavor pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1013.341749] env[61957]: DEBUG nova.virt.hardware [None req-0b16b4fe-c18a-4e4e-8833-61daba0fafce tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Image pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1013.341917] env[61957]: DEBUG nova.virt.hardware [None req-0b16b4fe-c18a-4e4e-8833-61daba0fafce tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1013.342207] env[61957]: DEBUG nova.virt.hardware [None req-0b16b4fe-c18a-4e4e-8833-61daba0fafce tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1013.342394] env[61957]: DEBUG nova.virt.hardware [None req-0b16b4fe-c18a-4e4e-8833-61daba0fafce tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1013.342560] env[61957]: DEBUG nova.virt.hardware [None req-0b16b4fe-c18a-4e4e-8833-61daba0fafce tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Got 1 possible topologies {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1013.342731] env[61957]: DEBUG nova.virt.hardware [None req-0b16b4fe-c18a-4e4e-8833-61daba0fafce tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1013.342934] env[61957]: DEBUG nova.virt.hardware [None req-0b16b4fe-c18a-4e4e-8833-61daba0fafce tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1013.348382] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-0b16b4fe-c18a-4e4e-8833-61daba0fafce tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: 6509707e-afc6-4c4a-88f9-a471865d510c] Reconfiguring VM instance instance-0000005d to detach disk 2000 {{(pid=61957) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1013.348949] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d92c7007-db9c-429d-aa03-41f7719669de tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Acquiring lock "refresh_cache-927af4c3-c7e8-404c-839e-d25af5e384a9" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1013.349203] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d92c7007-db9c-429d-aa03-41f7719669de tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Acquired lock "refresh_cache-927af4c3-c7e8-404c-839e-d25af5e384a9" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1013.349263] env[61957]: DEBUG nova.network.neutron [None req-d92c7007-db9c-429d-aa03-41f7719669de tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 927af4c3-c7e8-404c-839e-d25af5e384a9] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1013.350528] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cf962301-bb2c-4cfb-9a82-c0bb3e4d515a {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.377599] env[61957]: DEBUG oslo_vmware.api [None req-0b16b4fe-c18a-4e4e-8833-61daba0fafce tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Waiting for the task: (returnval){ [ 1013.377599] env[61957]: value = "task-1277961" [ 1013.377599] env[61957]: _type = "Task" [ 1013.377599] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.390594] env[61957]: DEBUG oslo_vmware.api [None req-0b16b4fe-c18a-4e4e-8833-61daba0fafce tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': task-1277961, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.453233] env[61957]: DEBUG oslo_vmware.api [None req-d6cf550f-30ee-4518-a1a7-00d815bdcf78 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': task-1277957, 'name': CloneVM_Task, 'duration_secs': 1.547544} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.453618] env[61957]: INFO nova.virt.vmwareapi.vmops [None req-d6cf550f-30ee-4518-a1a7-00d815bdcf78 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: e2107d5c-8453-4d24-b364-2b5c1d125062] Created linked-clone VM from snapshot [ 1013.454680] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-839d5594-2aac-4434-bfab-cdf6a0aafb1b {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.465330] env[61957]: DEBUG nova.virt.vmwareapi.images [None req-d6cf550f-30ee-4518-a1a7-00d815bdcf78 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: e2107d5c-8453-4d24-b364-2b5c1d125062] Uploading image 594a9221-8714-4a94-8ce1-b7ac0dead682 {{(pid=61957) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 1013.475223] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-d6cf550f-30ee-4518-a1a7-00d815bdcf78 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: e2107d5c-8453-4d24-b364-2b5c1d125062] Destroying the VM {{(pid=61957) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 1013.475492] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-e4727406-ac38-4702-9a9a-e1d907ca98d5 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.483958] env[61957]: DEBUG oslo_vmware.api [None req-d6cf550f-30ee-4518-a1a7-00d815bdcf78 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Waiting for the task: (returnval){ [ 1013.483958] env[61957]: value = "task-1277962" [ 1013.483958] env[61957]: _type = "Task" [ 1013.483958] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.493975] env[61957]: DEBUG oslo_vmware.api [None req-d6cf550f-30ee-4518-a1a7-00d815bdcf78 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': task-1277962, 'name': Destroy_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.552411] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0361426f-9477-48da-a62c-02a39ae47036 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.560267] env[61957]: DEBUG oslo_vmware.api [None req-ad83dc8d-b942-4b5a-8561-778a1542e9de tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Task: {'id': task-1277959, 'name': Destroy_Task, 'duration_secs': 0.386178} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.562160] env[61957]: INFO nova.virt.vmwareapi.vm_util [None req-ad83dc8d-b942-4b5a-8561-778a1542e9de tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] [instance: 9b0053fe-420e-4b92-86a8-ed44f6a6ec49] Destroyed the VM [ 1013.562416] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-ad83dc8d-b942-4b5a-8561-778a1542e9de tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] [instance: 9b0053fe-420e-4b92-86a8-ed44f6a6ec49] Deleting Snapshot of the VM instance {{(pid=61957) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1013.562708] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-392fde78-fb51-4939-87e2-24dc5e12dc44 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.565024] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06edac5e-7141-4991-b3ef-fd24d1e396d6 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.604285] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5536e4e-4106-44af-ae2b-e05b7312cd93 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.606957] env[61957]: DEBUG oslo_vmware.api [None req-ad83dc8d-b942-4b5a-8561-778a1542e9de tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Waiting for the task: (returnval){ [ 1013.606957] env[61957]: value = "task-1277963" [ 1013.606957] env[61957]: _type = "Task" [ 1013.606957] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.616228] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38b6db89-97dd-4348-b5ad-19cf4c885038 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.623312] env[61957]: DEBUG oslo_vmware.api [None req-ad83dc8d-b942-4b5a-8561-778a1542e9de tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Task: {'id': task-1277963, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.629038] env[61957]: DEBUG oslo_vmware.api [None req-1238a3ba-00b0-4ef8-9842-11dbb65e047b tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Task: {'id': task-1277960, 'name': PowerOffVM_Task, 'duration_secs': 0.233928} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.636824] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-1238a3ba-00b0-4ef8-9842-11dbb65e047b tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] [instance: 7793aa07-6aa2-459d-8a91-56c6b0412d3b] Powered off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1013.637023] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-1238a3ba-00b0-4ef8-9842-11dbb65e047b tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] [instance: 7793aa07-6aa2-459d-8a91-56c6b0412d3b] Unregistering the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1013.637795] env[61957]: DEBUG nova.compute.provider_tree [None req-6a1be883-2588-47a0-91f4-274d9905657e tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1013.639088] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-41520ebf-f5af-4fa6-aaba-db309a325428 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.705222] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-1238a3ba-00b0-4ef8-9842-11dbb65e047b tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] [instance: 7793aa07-6aa2-459d-8a91-56c6b0412d3b] Unregistered the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1013.705513] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-1238a3ba-00b0-4ef8-9842-11dbb65e047b tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] [instance: 7793aa07-6aa2-459d-8a91-56c6b0412d3b] Deleting contents of the VM from datastore datastore2 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1013.705702] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-1238a3ba-00b0-4ef8-9842-11dbb65e047b tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Deleting the datastore file [datastore2] 7793aa07-6aa2-459d-8a91-56c6b0412d3b {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1013.705987] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-50e10d05-4861-485b-8723-deba869c301e {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.714885] env[61957]: DEBUG oslo_vmware.api [None req-1238a3ba-00b0-4ef8-9842-11dbb65e047b tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Waiting for the task: (returnval){ [ 1013.714885] env[61957]: value = "task-1277965" [ 1013.714885] env[61957]: _type = "Task" [ 1013.714885] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.723454] env[61957]: DEBUG oslo_vmware.api [None req-1238a3ba-00b0-4ef8-9842-11dbb65e047b tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Task: {'id': task-1277965, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.731780] env[61957]: WARNING nova.network.neutron [None req-3409aaf8-ffe7-41b4-a265-d651c9473b75 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2] 364cabcb-046d-4f91-b1ee-5dd3adfae6d7 already exists in list: networks containing: ['364cabcb-046d-4f91-b1ee-5dd3adfae6d7']. ignoring it [ 1013.731990] env[61957]: WARNING nova.network.neutron [None req-3409aaf8-ffe7-41b4-a265-d651c9473b75 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2] 364cabcb-046d-4f91-b1ee-5dd3adfae6d7 already exists in list: networks containing: ['364cabcb-046d-4f91-b1ee-5dd3adfae6d7']. ignoring it [ 1013.889263] env[61957]: DEBUG oslo_vmware.api [None req-0b16b4fe-c18a-4e4e-8833-61daba0fafce tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': task-1277961, 'name': ReconfigVM_Task, 'duration_secs': 0.343949} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.889550] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-0b16b4fe-c18a-4e4e-8833-61daba0fafce tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: 6509707e-afc6-4c4a-88f9-a471865d510c] Reconfigured VM instance instance-0000005d to detach disk 2000 {{(pid=61957) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1013.890369] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8bf94f0-9c20-4492-8935-7dc7c0a632d0 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.912876] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-0b16b4fe-c18a-4e4e-8833-61daba0fafce tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: 6509707e-afc6-4c4a-88f9-a471865d510c] Reconfiguring VM instance instance-0000005d to attach disk [datastore1] 6509707e-afc6-4c4a-88f9-a471865d510c/6509707e-afc6-4c4a-88f9-a471865d510c.vmdk or device None with type thin {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1013.913642] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8440f6ed-22dd-4222-b6f7-a7f33e6e1d15 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.936242] env[61957]: DEBUG oslo_vmware.api [None req-0b16b4fe-c18a-4e4e-8833-61daba0fafce tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Waiting for the task: (returnval){ [ 1013.936242] env[61957]: value = "task-1277966" [ 1013.936242] env[61957]: _type = "Task" [ 1013.936242] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.948561] env[61957]: DEBUG oslo_vmware.api [None req-0b16b4fe-c18a-4e4e-8833-61daba0fafce tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': task-1277966, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.995972] env[61957]: DEBUG oslo_vmware.api [None req-d6cf550f-30ee-4518-a1a7-00d815bdcf78 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': task-1277962, 'name': Destroy_Task} progress is 33%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.107980] env[61957]: DEBUG nova.network.neutron [None req-d92c7007-db9c-429d-aa03-41f7719669de tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 927af4c3-c7e8-404c-839e-d25af5e384a9] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1014.120647] env[61957]: DEBUG oslo_vmware.api [None req-ad83dc8d-b942-4b5a-8561-778a1542e9de tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Task: {'id': task-1277963, 'name': RemoveSnapshot_Task, 'duration_secs': 0.367175} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.120921] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-ad83dc8d-b942-4b5a-8561-778a1542e9de tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] [instance: 9b0053fe-420e-4b92-86a8-ed44f6a6ec49] Deleted Snapshot of the VM instance {{(pid=61957) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1014.121221] env[61957]: DEBUG nova.compute.manager [None req-ad83dc8d-b942-4b5a-8561-778a1542e9de tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] [instance: 9b0053fe-420e-4b92-86a8-ed44f6a6ec49] Checking state {{(pid=61957) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1014.125020] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00d6c737-fdbb-472e-913d-3ba2e314591f {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.141705] env[61957]: DEBUG nova.scheduler.client.report [None req-6a1be883-2588-47a0-91f4-274d9905657e tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1014.225144] env[61957]: DEBUG oslo_vmware.api [None req-1238a3ba-00b0-4ef8-9842-11dbb65e047b tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Task: {'id': task-1277965, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.369328} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.225421] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-1238a3ba-00b0-4ef8-9842-11dbb65e047b tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Deleted the datastore file {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1014.225606] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-1238a3ba-00b0-4ef8-9842-11dbb65e047b tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] [instance: 7793aa07-6aa2-459d-8a91-56c6b0412d3b] Deleted contents of the VM from datastore datastore2 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1014.225785] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-1238a3ba-00b0-4ef8-9842-11dbb65e047b tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] [instance: 7793aa07-6aa2-459d-8a91-56c6b0412d3b] Instance destroyed {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1014.225964] env[61957]: INFO nova.compute.manager [None req-1238a3ba-00b0-4ef8-9842-11dbb65e047b tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] [instance: 7793aa07-6aa2-459d-8a91-56c6b0412d3b] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1014.226272] env[61957]: DEBUG oslo.service.loopingcall [None req-1238a3ba-00b0-4ef8-9842-11dbb65e047b tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1014.226473] env[61957]: DEBUG nova.compute.manager [-] [instance: 7793aa07-6aa2-459d-8a91-56c6b0412d3b] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1014.226569] env[61957]: DEBUG nova.network.neutron [-] [instance: 7793aa07-6aa2-459d-8a91-56c6b0412d3b] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1014.437238] env[61957]: DEBUG nova.network.neutron [None req-d92c7007-db9c-429d-aa03-41f7719669de tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 927af4c3-c7e8-404c-839e-d25af5e384a9] Updating instance_info_cache with network_info: [{"id": "38da1ba9-c5cc-4f76-b516-5ab8fdca9064", "address": "fa:16:3e:8e:54:ab", "network": {"id": "efe08823-51fe-4620-a849-5345e262cc71", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1035436819-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e16f6dce3f0e44fb96de516f17d4c6f5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "667a2e97-c1be-421d-9941-6b84c2629b43", "external-id": "nsx-vlan-transportzone-484", "segmentation_id": 484, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap38da1ba9-c5", "ovs_interfaceid": "38da1ba9-c5cc-4f76-b516-5ab8fdca9064", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1014.453412] env[61957]: DEBUG oslo_vmware.api [None req-0b16b4fe-c18a-4e4e-8833-61daba0fafce tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': task-1277966, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.496950] env[61957]: DEBUG oslo_vmware.api [None req-d6cf550f-30ee-4518-a1a7-00d815bdcf78 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': task-1277962, 'name': Destroy_Task, 'duration_secs': 0.596908} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.497266] env[61957]: INFO nova.virt.vmwareapi.vm_util [None req-d6cf550f-30ee-4518-a1a7-00d815bdcf78 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: e2107d5c-8453-4d24-b364-2b5c1d125062] Destroyed the VM [ 1014.497505] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-d6cf550f-30ee-4518-a1a7-00d815bdcf78 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: e2107d5c-8453-4d24-b364-2b5c1d125062] Deleting Snapshot of the VM instance {{(pid=61957) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1014.497766] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-3aca12da-02d8-43a3-9019-3fe96c2d9df9 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.504982] env[61957]: DEBUG oslo_vmware.api [None req-d6cf550f-30ee-4518-a1a7-00d815bdcf78 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Waiting for the task: (returnval){ [ 1014.504982] env[61957]: value = "task-1277967" [ 1014.504982] env[61957]: _type = "Task" [ 1014.504982] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.515196] env[61957]: DEBUG oslo_vmware.api [None req-d6cf550f-30ee-4518-a1a7-00d815bdcf78 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': task-1277967, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.541746] env[61957]: DEBUG nova.compute.manager [req-c7fca03f-3e62-4fd1-8348-e3202c1fd9f1 req-48ffe962-23ce-49ff-adf5-842fb4551858 service nova] [instance: 08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2] Received event network-changed-00c8220d-5633-4de7-9795-8a6a8543c4ec {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1014.542356] env[61957]: DEBUG nova.compute.manager [req-c7fca03f-3e62-4fd1-8348-e3202c1fd9f1 req-48ffe962-23ce-49ff-adf5-842fb4551858 service nova] [instance: 08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2] Refreshing instance network info cache due to event network-changed-00c8220d-5633-4de7-9795-8a6a8543c4ec. {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1014.542356] env[61957]: DEBUG oslo_concurrency.lockutils [req-c7fca03f-3e62-4fd1-8348-e3202c1fd9f1 req-48ffe962-23ce-49ff-adf5-842fb4551858 service nova] Acquiring lock "refresh_cache-08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1014.636147] env[61957]: INFO nova.compute.manager [None req-ad83dc8d-b942-4b5a-8561-778a1542e9de tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] [instance: 9b0053fe-420e-4b92-86a8-ed44f6a6ec49] Shelve offloading [ 1014.638506] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-ad83dc8d-b942-4b5a-8561-778a1542e9de tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] [instance: 9b0053fe-420e-4b92-86a8-ed44f6a6ec49] Powering off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1014.638506] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-97a2b912-3b0a-45a6-ba2b-9699cc44245e {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.647224] env[61957]: DEBUG oslo_vmware.api [None req-ad83dc8d-b942-4b5a-8561-778a1542e9de tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Waiting for the task: (returnval){ [ 1014.647224] env[61957]: value = "task-1277968" [ 1014.647224] env[61957]: _type = "Task" [ 1014.647224] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.648038] env[61957]: DEBUG oslo_concurrency.lockutils [None req-6a1be883-2588-47a0-91f4-274d9905657e tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.300s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1014.648755] env[61957]: DEBUG nova.compute.manager [None req-6a1be883-2588-47a0-91f4-274d9905657e tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 3f144012-c0cc-43ce-8c73-4422fe8c1fe6] Start building networks asynchronously for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1014.662298] env[61957]: DEBUG nova.network.neutron [None req-3409aaf8-ffe7-41b4-a265-d651c9473b75 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2] Updating instance_info_cache with network_info: [{"id": "f62a9f12-0b55-4138-a40e-b431855487fe", "address": "fa:16:3e:af:48:d3", "network": {"id": "364cabcb-046d-4f91-b1ee-5dd3adfae6d7", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1432396967-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.219", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8713aa35bcb24b86ad0b58ca9fc991ba", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b7a73c01-1bb9-4612-a1a7-16d71b732e81", "external-id": "nsx-vlan-transportzone-711", "segmentation_id": 711, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf62a9f12-0b", "ovs_interfaceid": "f62a9f12-0b55-4138-a40e-b431855487fe", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "6a70445e-1266-453e-a6ff-50e8cbdcfd08", "address": "fa:16:3e:bb:2f:b6", "network": {"id": "364cabcb-046d-4f91-b1ee-5dd3adfae6d7", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1432396967-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8713aa35bcb24b86ad0b58ca9fc991ba", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b7a73c01-1bb9-4612-a1a7-16d71b732e81", "external-id": "nsx-vlan-transportzone-711", "segmentation_id": 711, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6a70445e-12", "ovs_interfaceid": "6a70445e-1266-453e-a6ff-50e8cbdcfd08", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "00c8220d-5633-4de7-9795-8a6a8543c4ec", "address": "fa:16:3e:00:80:71", "network": {"id": "364cabcb-046d-4f91-b1ee-5dd3adfae6d7", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1432396967-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8713aa35bcb24b86ad0b58ca9fc991ba", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b7a73c01-1bb9-4612-a1a7-16d71b732e81", "external-id": "nsx-vlan-transportzone-711", "segmentation_id": 711, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap00c8220d-56", "ovs_interfaceid": "00c8220d-5633-4de7-9795-8a6a8543c4ec", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1014.663318] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-ad83dc8d-b942-4b5a-8561-778a1542e9de tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] [instance: 9b0053fe-420e-4b92-86a8-ed44f6a6ec49] VM already powered off {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 1014.663642] env[61957]: DEBUG nova.compute.manager [None req-ad83dc8d-b942-4b5a-8561-778a1542e9de tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] [instance: 9b0053fe-420e-4b92-86a8-ed44f6a6ec49] Checking state {{(pid=61957) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1014.664492] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfbbcce7-8a55-4bdd-81fb-0f55c16cfd81 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.671277] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ad83dc8d-b942-4b5a-8561-778a1542e9de tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Acquiring lock "refresh_cache-9b0053fe-420e-4b92-86a8-ed44f6a6ec49" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1014.671453] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ad83dc8d-b942-4b5a-8561-778a1542e9de tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Acquired lock "refresh_cache-9b0053fe-420e-4b92-86a8-ed44f6a6ec49" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1014.671629] env[61957]: DEBUG nova.network.neutron [None req-ad83dc8d-b942-4b5a-8561-778a1542e9de tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] [instance: 9b0053fe-420e-4b92-86a8-ed44f6a6ec49] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1014.948898] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d92c7007-db9c-429d-aa03-41f7719669de tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Releasing lock "refresh_cache-927af4c3-c7e8-404c-839e-d25af5e384a9" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1014.949299] env[61957]: DEBUG nova.compute.manager [None req-d92c7007-db9c-429d-aa03-41f7719669de tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 927af4c3-c7e8-404c-839e-d25af5e384a9] Instance network_info: |[{"id": "38da1ba9-c5cc-4f76-b516-5ab8fdca9064", "address": "fa:16:3e:8e:54:ab", "network": {"id": "efe08823-51fe-4620-a849-5345e262cc71", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1035436819-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e16f6dce3f0e44fb96de516f17d4c6f5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "667a2e97-c1be-421d-9941-6b84c2629b43", "external-id": "nsx-vlan-transportzone-484", "segmentation_id": 484, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap38da1ba9-c5", "ovs_interfaceid": "38da1ba9-c5cc-4f76-b516-5ab8fdca9064", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1014.949677] env[61957]: DEBUG oslo_vmware.api [None req-0b16b4fe-c18a-4e4e-8833-61daba0fafce tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': task-1277966, 'name': ReconfigVM_Task, 'duration_secs': 0.60721} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.950110] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-d92c7007-db9c-429d-aa03-41f7719669de tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 927af4c3-c7e8-404c-839e-d25af5e384a9] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:8e:54:ab', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '667a2e97-c1be-421d-9941-6b84c2629b43', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '38da1ba9-c5cc-4f76-b516-5ab8fdca9064', 'vif_model': 'vmxnet3'}] {{(pid=61957) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1014.957813] env[61957]: DEBUG oslo.service.loopingcall [None req-d92c7007-db9c-429d-aa03-41f7719669de tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1014.957929] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-0b16b4fe-c18a-4e4e-8833-61daba0fafce tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: 6509707e-afc6-4c4a-88f9-a471865d510c] Reconfigured VM instance instance-0000005d to attach disk [datastore1] 6509707e-afc6-4c4a-88f9-a471865d510c/6509707e-afc6-4c4a-88f9-a471865d510c.vmdk or device None with type thin {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1014.958141] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-0b16b4fe-c18a-4e4e-8833-61daba0fafce tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: 6509707e-afc6-4c4a-88f9-a471865d510c] Updating instance '6509707e-afc6-4c4a-88f9-a471865d510c' progress to 50 {{(pid=61957) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1014.961411] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 927af4c3-c7e8-404c-839e-d25af5e384a9] Creating VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1014.961950] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-cd5c7f37-a6a5-4d84-8843-f5b251eaf777 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.982052] env[61957]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1014.982052] env[61957]: value = "task-1277969" [ 1014.982052] env[61957]: _type = "Task" [ 1014.982052] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.990711] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277969, 'name': CreateVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.015144] env[61957]: DEBUG oslo_vmware.api [None req-d6cf550f-30ee-4518-a1a7-00d815bdcf78 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': task-1277967, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.156707] env[61957]: DEBUG nova.compute.utils [None req-6a1be883-2588-47a0-91f4-274d9905657e tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Using /dev/sd instead of None {{(pid=61957) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1015.158466] env[61957]: DEBUG nova.compute.manager [None req-6a1be883-2588-47a0-91f4-274d9905657e tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 3f144012-c0cc-43ce-8c73-4422fe8c1fe6] Allocating IP information in the background. {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1015.158657] env[61957]: DEBUG nova.network.neutron [None req-6a1be883-2588-47a0-91f4-274d9905657e tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 3f144012-c0cc-43ce-8c73-4422fe8c1fe6] allocate_for_instance() {{(pid=61957) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1015.164675] env[61957]: DEBUG oslo_concurrency.lockutils [None req-3409aaf8-ffe7-41b4-a265-d651c9473b75 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Releasing lock "refresh_cache-08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1015.165381] env[61957]: DEBUG oslo_concurrency.lockutils [None req-3409aaf8-ffe7-41b4-a265-d651c9473b75 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Acquiring lock "08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1015.165552] env[61957]: DEBUG oslo_concurrency.lockutils [None req-3409aaf8-ffe7-41b4-a265-d651c9473b75 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Acquired lock "08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1015.165834] env[61957]: DEBUG oslo_concurrency.lockutils [req-c7fca03f-3e62-4fd1-8348-e3202c1fd9f1 req-48ffe962-23ce-49ff-adf5-842fb4551858 service nova] Acquired lock "refresh_cache-08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1015.166061] env[61957]: DEBUG nova.network.neutron [req-c7fca03f-3e62-4fd1-8348-e3202c1fd9f1 req-48ffe962-23ce-49ff-adf5-842fb4551858 service nova] [instance: 08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2] Refreshing network info cache for port 00c8220d-5633-4de7-9795-8a6a8543c4ec {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1015.168293] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c0ef151-e7a1-4fa9-9e82-713af3242899 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.191956] env[61957]: DEBUG nova.virt.hardware [None req-3409aaf8-ffe7-41b4-a265-d651c9473b75 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T17:22:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1015.192364] env[61957]: DEBUG nova.virt.hardware [None req-3409aaf8-ffe7-41b4-a265-d651c9473b75 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Flavor limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1015.192522] env[61957]: DEBUG nova.virt.hardware [None req-3409aaf8-ffe7-41b4-a265-d651c9473b75 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Image limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1015.192676] env[61957]: DEBUG nova.virt.hardware [None req-3409aaf8-ffe7-41b4-a265-d651c9473b75 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Flavor pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1015.192756] env[61957]: DEBUG nova.virt.hardware [None req-3409aaf8-ffe7-41b4-a265-d651c9473b75 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Image pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1015.192881] env[61957]: DEBUG nova.virt.hardware [None req-3409aaf8-ffe7-41b4-a265-d651c9473b75 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1015.193113] env[61957]: DEBUG nova.virt.hardware [None req-3409aaf8-ffe7-41b4-a265-d651c9473b75 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1015.193265] env[61957]: DEBUG nova.virt.hardware [None req-3409aaf8-ffe7-41b4-a265-d651c9473b75 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1015.193435] env[61957]: DEBUG nova.virt.hardware [None req-3409aaf8-ffe7-41b4-a265-d651c9473b75 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Got 1 possible topologies {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1015.193599] env[61957]: DEBUG nova.virt.hardware [None req-3409aaf8-ffe7-41b4-a265-d651c9473b75 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1015.193837] env[61957]: DEBUG nova.virt.hardware [None req-3409aaf8-ffe7-41b4-a265-d651c9473b75 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1015.200053] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-3409aaf8-ffe7-41b4-a265-d651c9473b75 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2] Reconfiguring VM to attach interface {{(pid=61957) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1929}} [ 1015.201179] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4c1e35ae-a2a3-43f4-b6d7-09a2792f5ce6 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.214315] env[61957]: DEBUG nova.policy [None req-6a1be883-2588-47a0-91f4-274d9905657e tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e2f13689624b483d903ce12ef290db8b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '65fd71bcd1bf41238a9cc3a5d6dd4924', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61957) authorize /opt/stack/nova/nova/policy.py:203}} [ 1015.216631] env[61957]: DEBUG nova.network.neutron [-] [instance: 7793aa07-6aa2-459d-8a91-56c6b0412d3b] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1015.222377] env[61957]: DEBUG oslo_vmware.api [None req-3409aaf8-ffe7-41b4-a265-d651c9473b75 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Waiting for the task: (returnval){ [ 1015.222377] env[61957]: value = "task-1277970" [ 1015.222377] env[61957]: _type = "Task" [ 1015.222377] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.231552] env[61957]: DEBUG oslo_vmware.api [None req-3409aaf8-ffe7-41b4-a265-d651c9473b75 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': task-1277970, 'name': ReconfigVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.465194] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-637dee0f-a092-4d82-ab1a-ac0002955de3 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.490334] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fe1a837-8bb0-4e0a-8335-e642908a41d8 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.499864] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277969, 'name': CreateVM_Task, 'duration_secs': 0.457293} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.522742] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 927af4c3-c7e8-404c-839e-d25af5e384a9] Created VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1015.523390] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-0b16b4fe-c18a-4e4e-8833-61daba0fafce tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: 6509707e-afc6-4c4a-88f9-a471865d510c] Updating instance '6509707e-afc6-4c4a-88f9-a471865d510c' progress to 67 {{(pid=61957) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1015.527497] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d92c7007-db9c-429d-aa03-41f7719669de tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1015.528062] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d92c7007-db9c-429d-aa03-41f7719669de tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Acquired lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1015.528062] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d92c7007-db9c-429d-aa03-41f7719669de tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1015.533021] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-54415762-2033-4d41-b26b-ade302666467 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.537844] env[61957]: DEBUG oslo_vmware.api [None req-d92c7007-db9c-429d-aa03-41f7719669de tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Waiting for the task: (returnval){ [ 1015.537844] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52c22cdd-b63d-f32f-c8b2-cefc2e11c12a" [ 1015.537844] env[61957]: _type = "Task" [ 1015.537844] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.542090] env[61957]: DEBUG oslo_vmware.api [None req-d6cf550f-30ee-4518-a1a7-00d815bdcf78 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': task-1277967, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.556607] env[61957]: DEBUG oslo_vmware.api [None req-d92c7007-db9c-429d-aa03-41f7719669de tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52c22cdd-b63d-f32f-c8b2-cefc2e11c12a, 'name': SearchDatastore_Task, 'duration_secs': 0.01252} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.556607] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d92c7007-db9c-429d-aa03-41f7719669de tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Releasing lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1015.556607] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-d92c7007-db9c-429d-aa03-41f7719669de tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 927af4c3-c7e8-404c-839e-d25af5e384a9] Processing image 11c76a2c-f705-470a-ba9d-4657858bab38 {{(pid=61957) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1015.556607] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d92c7007-db9c-429d-aa03-41f7719669de tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1015.556607] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d92c7007-db9c-429d-aa03-41f7719669de tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Acquired lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1015.556607] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-d92c7007-db9c-429d-aa03-41f7719669de tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1015.556607] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0e3f3434-f004-4ca3-97e0-1c7373cde955 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.559817] env[61957]: DEBUG nova.network.neutron [None req-ad83dc8d-b942-4b5a-8561-778a1542e9de tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] [instance: 9b0053fe-420e-4b92-86a8-ed44f6a6ec49] Updating instance_info_cache with network_info: [{"id": "219230f0-0b14-4be9-81bb-67a765d8ced9", "address": "fa:16:3e:f6:9e:ce", "network": {"id": "cf378e9c-eaae-4168-bd60-9818fcfaf405", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1243326371-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f7c20ab3c822423fa837a03f066e4239", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8cb478a6-872c-4a90-a8db-526b374e82ce", "external-id": "nsx-vlan-transportzone-835", "segmentation_id": 835, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap219230f0-0b", "ovs_interfaceid": "219230f0-0b14-4be9-81bb-67a765d8ced9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1015.581842] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-d92c7007-db9c-429d-aa03-41f7719669de tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1015.582222] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-d92c7007-db9c-429d-aa03-41f7719669de tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61957) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1015.583946] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-63a9fd0f-c2a1-4d4f-a572-9ea76735efea {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.591934] env[61957]: DEBUG oslo_vmware.api [None req-d92c7007-db9c-429d-aa03-41f7719669de tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Waiting for the task: (returnval){ [ 1015.591934] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52dee0a5-f1d3-05a1-e9d8-01336fefa64a" [ 1015.591934] env[61957]: _type = "Task" [ 1015.591934] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.603453] env[61957]: DEBUG oslo_vmware.api [None req-d92c7007-db9c-429d-aa03-41f7719669de tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52dee0a5-f1d3-05a1-e9d8-01336fefa64a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.648308] env[61957]: DEBUG nova.network.neutron [None req-6a1be883-2588-47a0-91f4-274d9905657e tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 3f144012-c0cc-43ce-8c73-4422fe8c1fe6] Successfully created port: 921a0283-f0c8-4cd9-8548-5e1089a438c3 {{(pid=61957) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1015.663197] env[61957]: DEBUG nova.compute.manager [None req-6a1be883-2588-47a0-91f4-274d9905657e tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 3f144012-c0cc-43ce-8c73-4422fe8c1fe6] Start building block device mappings for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1015.718686] env[61957]: INFO nova.compute.manager [-] [instance: 7793aa07-6aa2-459d-8a91-56c6b0412d3b] Took 1.49 seconds to deallocate network for instance. [ 1015.738696] env[61957]: DEBUG oslo_vmware.api [None req-3409aaf8-ffe7-41b4-a265-d651c9473b75 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': task-1277970, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.018277] env[61957]: DEBUG nova.network.neutron [req-c7fca03f-3e62-4fd1-8348-e3202c1fd9f1 req-48ffe962-23ce-49ff-adf5-842fb4551858 service nova] [instance: 08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2] Updated VIF entry in instance network info cache for port 00c8220d-5633-4de7-9795-8a6a8543c4ec. {{(pid=61957) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1016.018774] env[61957]: DEBUG nova.network.neutron [req-c7fca03f-3e62-4fd1-8348-e3202c1fd9f1 req-48ffe962-23ce-49ff-adf5-842fb4551858 service nova] [instance: 08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2] Updating instance_info_cache with network_info: [{"id": "f62a9f12-0b55-4138-a40e-b431855487fe", "address": "fa:16:3e:af:48:d3", "network": {"id": "364cabcb-046d-4f91-b1ee-5dd3adfae6d7", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1432396967-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.219", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8713aa35bcb24b86ad0b58ca9fc991ba", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b7a73c01-1bb9-4612-a1a7-16d71b732e81", "external-id": "nsx-vlan-transportzone-711", "segmentation_id": 711, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf62a9f12-0b", "ovs_interfaceid": "f62a9f12-0b55-4138-a40e-b431855487fe", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "6a70445e-1266-453e-a6ff-50e8cbdcfd08", "address": "fa:16:3e:bb:2f:b6", "network": {"id": "364cabcb-046d-4f91-b1ee-5dd3adfae6d7", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1432396967-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8713aa35bcb24b86ad0b58ca9fc991ba", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b7a73c01-1bb9-4612-a1a7-16d71b732e81", "external-id": "nsx-vlan-transportzone-711", "segmentation_id": 711, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6a70445e-12", "ovs_interfaceid": "6a70445e-1266-453e-a6ff-50e8cbdcfd08", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "00c8220d-5633-4de7-9795-8a6a8543c4ec", "address": "fa:16:3e:00:80:71", "network": {"id": "364cabcb-046d-4f91-b1ee-5dd3adfae6d7", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1432396967-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8713aa35bcb24b86ad0b58ca9fc991ba", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b7a73c01-1bb9-4612-a1a7-16d71b732e81", "external-id": "nsx-vlan-transportzone-711", "segmentation_id": 711, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap00c8220d-56", "ovs_interfaceid": "00c8220d-5633-4de7-9795-8a6a8543c4ec", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1016.037081] env[61957]: DEBUG oslo_vmware.api [None req-d6cf550f-30ee-4518-a1a7-00d815bdcf78 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': task-1277967, 'name': RemoveSnapshot_Task, 'duration_secs': 1.043822} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.037739] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-d6cf550f-30ee-4518-a1a7-00d815bdcf78 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: e2107d5c-8453-4d24-b364-2b5c1d125062] Deleted Snapshot of the VM instance {{(pid=61957) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1016.062707] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ad83dc8d-b942-4b5a-8561-778a1542e9de tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Releasing lock "refresh_cache-9b0053fe-420e-4b92-86a8-ed44f6a6ec49" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1016.107252] env[61957]: DEBUG oslo_vmware.api [None req-d92c7007-db9c-429d-aa03-41f7719669de tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52dee0a5-f1d3-05a1-e9d8-01336fefa64a, 'name': SearchDatastore_Task, 'duration_secs': 0.011942} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.107440] env[61957]: DEBUG nova.network.neutron [None req-0b16b4fe-c18a-4e4e-8833-61daba0fafce tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: 6509707e-afc6-4c4a-88f9-a471865d510c] Port 7df2736b-10d0-4f9d-83aa-e2bbdf1ef45b binding to destination host cpu-1 is already ACTIVE {{(pid=61957) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 1016.109789] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a1bc30a3-7cb6-4255-a4db-17ffc08556f2 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.115736] env[61957]: DEBUG oslo_vmware.api [None req-d92c7007-db9c-429d-aa03-41f7719669de tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Waiting for the task: (returnval){ [ 1016.115736] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52bb775f-a06f-ce2e-c06e-e2cbfd82de67" [ 1016.115736] env[61957]: _type = "Task" [ 1016.115736] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.125737] env[61957]: DEBUG oslo_vmware.api [None req-d92c7007-db9c-429d-aa03-41f7719669de tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52bb775f-a06f-ce2e-c06e-e2cbfd82de67, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.227423] env[61957]: DEBUG oslo_concurrency.lockutils [None req-1238a3ba-00b0-4ef8-9842-11dbb65e047b tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1016.227696] env[61957]: DEBUG oslo_concurrency.lockutils [None req-1238a3ba-00b0-4ef8-9842-11dbb65e047b tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1016.227917] env[61957]: DEBUG nova.objects.instance [None req-1238a3ba-00b0-4ef8-9842-11dbb65e047b tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Lazy-loading 'resources' on Instance uuid 7793aa07-6aa2-459d-8a91-56c6b0412d3b {{(pid=61957) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1016.236361] env[61957]: DEBUG oslo_vmware.api [None req-3409aaf8-ffe7-41b4-a265-d651c9473b75 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': task-1277970, 'name': ReconfigVM_Task, 'duration_secs': 0.639195} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.237411] env[61957]: DEBUG oslo_concurrency.lockutils [None req-3409aaf8-ffe7-41b4-a265-d651c9473b75 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Releasing lock "08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1016.237635] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-3409aaf8-ffe7-41b4-a265-d651c9473b75 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2] Reconfigured VM to attach interface {{(pid=61957) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1943}} [ 1016.351561] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-ad83dc8d-b942-4b5a-8561-778a1542e9de tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] [instance: 9b0053fe-420e-4b92-86a8-ed44f6a6ec49] Destroying instance {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1016.352463] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-781bca4f-9eec-46bd-9509-bbe3476bf778 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.361384] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-ad83dc8d-b942-4b5a-8561-778a1542e9de tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] [instance: 9b0053fe-420e-4b92-86a8-ed44f6a6ec49] Unregistering the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1016.361667] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ffbddd6f-a60e-4399-94f1-cc31b11b83cc {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.436255] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-ad83dc8d-b942-4b5a-8561-778a1542e9de tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] [instance: 9b0053fe-420e-4b92-86a8-ed44f6a6ec49] Unregistered the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1016.436445] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-ad83dc8d-b942-4b5a-8561-778a1542e9de tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] [instance: 9b0053fe-420e-4b92-86a8-ed44f6a6ec49] Deleting contents of the VM from datastore datastore1 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1016.436617] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-ad83dc8d-b942-4b5a-8561-778a1542e9de tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Deleting the datastore file [datastore1] 9b0053fe-420e-4b92-86a8-ed44f6a6ec49 {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1016.436909] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-86b6992b-c46d-42f2-90c5-d7c5836319bb {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.446427] env[61957]: DEBUG oslo_vmware.api [None req-ad83dc8d-b942-4b5a-8561-778a1542e9de tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Waiting for the task: (returnval){ [ 1016.446427] env[61957]: value = "task-1277972" [ 1016.446427] env[61957]: _type = "Task" [ 1016.446427] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.454927] env[61957]: DEBUG oslo_vmware.api [None req-ad83dc8d-b942-4b5a-8561-778a1542e9de tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Task: {'id': task-1277972, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.523358] env[61957]: DEBUG oslo_concurrency.lockutils [req-c7fca03f-3e62-4fd1-8348-e3202c1fd9f1 req-48ffe962-23ce-49ff-adf5-842fb4551858 service nova] Releasing lock "refresh_cache-08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1016.523672] env[61957]: DEBUG nova.compute.manager [req-c7fca03f-3e62-4fd1-8348-e3202c1fd9f1 req-48ffe962-23ce-49ff-adf5-842fb4551858 service nova] [instance: 927af4c3-c7e8-404c-839e-d25af5e384a9] Received event network-vif-plugged-38da1ba9-c5cc-4f76-b516-5ab8fdca9064 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1016.523913] env[61957]: DEBUG oslo_concurrency.lockutils [req-c7fca03f-3e62-4fd1-8348-e3202c1fd9f1 req-48ffe962-23ce-49ff-adf5-842fb4551858 service nova] Acquiring lock "927af4c3-c7e8-404c-839e-d25af5e384a9-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1016.524176] env[61957]: DEBUG oslo_concurrency.lockutils [req-c7fca03f-3e62-4fd1-8348-e3202c1fd9f1 req-48ffe962-23ce-49ff-adf5-842fb4551858 service nova] Lock "927af4c3-c7e8-404c-839e-d25af5e384a9-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1016.524509] env[61957]: DEBUG oslo_concurrency.lockutils [req-c7fca03f-3e62-4fd1-8348-e3202c1fd9f1 req-48ffe962-23ce-49ff-adf5-842fb4551858 service nova] Lock "927af4c3-c7e8-404c-839e-d25af5e384a9-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1016.524597] env[61957]: DEBUG nova.compute.manager [req-c7fca03f-3e62-4fd1-8348-e3202c1fd9f1 req-48ffe962-23ce-49ff-adf5-842fb4551858 service nova] [instance: 927af4c3-c7e8-404c-839e-d25af5e384a9] No waiting events found dispatching network-vif-plugged-38da1ba9-c5cc-4f76-b516-5ab8fdca9064 {{(pid=61957) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1016.524771] env[61957]: WARNING nova.compute.manager [req-c7fca03f-3e62-4fd1-8348-e3202c1fd9f1 req-48ffe962-23ce-49ff-adf5-842fb4551858 service nova] [instance: 927af4c3-c7e8-404c-839e-d25af5e384a9] Received unexpected event network-vif-plugged-38da1ba9-c5cc-4f76-b516-5ab8fdca9064 for instance with vm_state building and task_state spawning. [ 1016.524936] env[61957]: DEBUG nova.compute.manager [req-c7fca03f-3e62-4fd1-8348-e3202c1fd9f1 req-48ffe962-23ce-49ff-adf5-842fb4551858 service nova] [instance: 927af4c3-c7e8-404c-839e-d25af5e384a9] Received event network-changed-38da1ba9-c5cc-4f76-b516-5ab8fdca9064 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1016.525118] env[61957]: DEBUG nova.compute.manager [req-c7fca03f-3e62-4fd1-8348-e3202c1fd9f1 req-48ffe962-23ce-49ff-adf5-842fb4551858 service nova] [instance: 927af4c3-c7e8-404c-839e-d25af5e384a9] Refreshing instance network info cache due to event network-changed-38da1ba9-c5cc-4f76-b516-5ab8fdca9064. {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1016.525320] env[61957]: DEBUG oslo_concurrency.lockutils [req-c7fca03f-3e62-4fd1-8348-e3202c1fd9f1 req-48ffe962-23ce-49ff-adf5-842fb4551858 service nova] Acquiring lock "refresh_cache-927af4c3-c7e8-404c-839e-d25af5e384a9" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1016.525461] env[61957]: DEBUG oslo_concurrency.lockutils [req-c7fca03f-3e62-4fd1-8348-e3202c1fd9f1 req-48ffe962-23ce-49ff-adf5-842fb4551858 service nova] Acquired lock "refresh_cache-927af4c3-c7e8-404c-839e-d25af5e384a9" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1016.525618] env[61957]: DEBUG nova.network.neutron [req-c7fca03f-3e62-4fd1-8348-e3202c1fd9f1 req-48ffe962-23ce-49ff-adf5-842fb4551858 service nova] [instance: 927af4c3-c7e8-404c-839e-d25af5e384a9] Refreshing network info cache for port 38da1ba9-c5cc-4f76-b516-5ab8fdca9064 {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1016.544912] env[61957]: WARNING nova.compute.manager [None req-d6cf550f-30ee-4518-a1a7-00d815bdcf78 tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: e2107d5c-8453-4d24-b364-2b5c1d125062] Image not found during snapshot: nova.exception.ImageNotFound: Image 594a9221-8714-4a94-8ce1-b7ac0dead682 could not be found. [ 1016.625214] env[61957]: DEBUG nova.compute.manager [req-16debee1-fc20-418c-8716-e7cb5152f756 req-dad78efb-9b50-4e74-a3a3-49109722847e service nova] [instance: 7793aa07-6aa2-459d-8a91-56c6b0412d3b] Received event network-vif-deleted-dfb387b2-224a-48bc-a581-17c9f58b213a {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1016.625860] env[61957]: DEBUG nova.compute.manager [req-16debee1-fc20-418c-8716-e7cb5152f756 req-dad78efb-9b50-4e74-a3a3-49109722847e service nova] [instance: 9b0053fe-420e-4b92-86a8-ed44f6a6ec49] Received event network-vif-unplugged-219230f0-0b14-4be9-81bb-67a765d8ced9 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1016.626286] env[61957]: DEBUG oslo_concurrency.lockutils [req-16debee1-fc20-418c-8716-e7cb5152f756 req-dad78efb-9b50-4e74-a3a3-49109722847e service nova] Acquiring lock "9b0053fe-420e-4b92-86a8-ed44f6a6ec49-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1016.626633] env[61957]: DEBUG oslo_concurrency.lockutils [req-16debee1-fc20-418c-8716-e7cb5152f756 req-dad78efb-9b50-4e74-a3a3-49109722847e service nova] Lock "9b0053fe-420e-4b92-86a8-ed44f6a6ec49-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1016.626911] env[61957]: DEBUG oslo_concurrency.lockutils [req-16debee1-fc20-418c-8716-e7cb5152f756 req-dad78efb-9b50-4e74-a3a3-49109722847e service nova] Lock "9b0053fe-420e-4b92-86a8-ed44f6a6ec49-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1016.627186] env[61957]: DEBUG nova.compute.manager [req-16debee1-fc20-418c-8716-e7cb5152f756 req-dad78efb-9b50-4e74-a3a3-49109722847e service nova] [instance: 9b0053fe-420e-4b92-86a8-ed44f6a6ec49] No waiting events found dispatching network-vif-unplugged-219230f0-0b14-4be9-81bb-67a765d8ced9 {{(pid=61957) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1016.627452] env[61957]: WARNING nova.compute.manager [req-16debee1-fc20-418c-8716-e7cb5152f756 req-dad78efb-9b50-4e74-a3a3-49109722847e service nova] [instance: 9b0053fe-420e-4b92-86a8-ed44f6a6ec49] Received unexpected event network-vif-unplugged-219230f0-0b14-4be9-81bb-67a765d8ced9 for instance with vm_state shelved and task_state shelving_offloading. [ 1016.627744] env[61957]: DEBUG nova.compute.manager [req-16debee1-fc20-418c-8716-e7cb5152f756 req-dad78efb-9b50-4e74-a3a3-49109722847e service nova] [instance: 9b0053fe-420e-4b92-86a8-ed44f6a6ec49] Received event network-changed-219230f0-0b14-4be9-81bb-67a765d8ced9 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1016.627971] env[61957]: DEBUG nova.compute.manager [req-16debee1-fc20-418c-8716-e7cb5152f756 req-dad78efb-9b50-4e74-a3a3-49109722847e service nova] [instance: 9b0053fe-420e-4b92-86a8-ed44f6a6ec49] Refreshing instance network info cache due to event network-changed-219230f0-0b14-4be9-81bb-67a765d8ced9. {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1016.628286] env[61957]: DEBUG oslo_concurrency.lockutils [req-16debee1-fc20-418c-8716-e7cb5152f756 req-dad78efb-9b50-4e74-a3a3-49109722847e service nova] Acquiring lock "refresh_cache-9b0053fe-420e-4b92-86a8-ed44f6a6ec49" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1016.628507] env[61957]: DEBUG oslo_concurrency.lockutils [req-16debee1-fc20-418c-8716-e7cb5152f756 req-dad78efb-9b50-4e74-a3a3-49109722847e service nova] Acquired lock "refresh_cache-9b0053fe-420e-4b92-86a8-ed44f6a6ec49" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1016.628744] env[61957]: DEBUG nova.network.neutron [req-16debee1-fc20-418c-8716-e7cb5152f756 req-dad78efb-9b50-4e74-a3a3-49109722847e service nova] [instance: 9b0053fe-420e-4b92-86a8-ed44f6a6ec49] Refreshing network info cache for port 219230f0-0b14-4be9-81bb-67a765d8ced9 {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1016.637282] env[61957]: DEBUG oslo_vmware.api [None req-d92c7007-db9c-429d-aa03-41f7719669de tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52bb775f-a06f-ce2e-c06e-e2cbfd82de67, 'name': SearchDatastore_Task, 'duration_secs': 0.043154} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.637955] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d92c7007-db9c-429d-aa03-41f7719669de tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Releasing lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1016.638342] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-d92c7007-db9c-429d-aa03-41f7719669de tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore1] 927af4c3-c7e8-404c-839e-d25af5e384a9/927af4c3-c7e8-404c-839e-d25af5e384a9.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1016.640272] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2f5a5b2b-e6b1-46c4-94b0-bc4d846995d0 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.650310] env[61957]: DEBUG oslo_vmware.api [None req-d92c7007-db9c-429d-aa03-41f7719669de tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Waiting for the task: (returnval){ [ 1016.650310] env[61957]: value = "task-1277973" [ 1016.650310] env[61957]: _type = "Task" [ 1016.650310] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.662893] env[61957]: DEBUG oslo_vmware.api [None req-d92c7007-db9c-429d-aa03-41f7719669de tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1277973, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.674497] env[61957]: DEBUG nova.compute.manager [None req-6a1be883-2588-47a0-91f4-274d9905657e tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 3f144012-c0cc-43ce-8c73-4422fe8c1fe6] Start spawning the instance on the hypervisor. {{(pid=61957) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1016.681035] env[61957]: DEBUG oslo_concurrency.lockutils [None req-20123269-72b2-4ce0-b74a-dd98ef589cf3 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Acquiring lock "964c8c9d-0721-4ad3-a146-fce1fd1d6513" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1016.681287] env[61957]: DEBUG oslo_concurrency.lockutils [None req-20123269-72b2-4ce0-b74a-dd98ef589cf3 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Lock "964c8c9d-0721-4ad3-a146-fce1fd1d6513" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1016.702433] env[61957]: DEBUG nova.virt.hardware [None req-6a1be883-2588-47a0-91f4-274d9905657e tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T17:22:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T17:21:50Z,direct_url=,disk_format='vmdk',id=11c76a2c-f705-470a-ba9d-4657858bab38,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='81d3096d0f094373913fc3dc8f5c3c6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T17:21:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1016.702716] env[61957]: DEBUG nova.virt.hardware [None req-6a1be883-2588-47a0-91f4-274d9905657e tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Flavor limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1016.702877] env[61957]: DEBUG nova.virt.hardware [None req-6a1be883-2588-47a0-91f4-274d9905657e tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Image limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1016.703077] env[61957]: DEBUG nova.virt.hardware [None req-6a1be883-2588-47a0-91f4-274d9905657e tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Flavor pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1016.703275] env[61957]: DEBUG nova.virt.hardware [None req-6a1be883-2588-47a0-91f4-274d9905657e tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Image pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1016.703457] env[61957]: DEBUG nova.virt.hardware [None req-6a1be883-2588-47a0-91f4-274d9905657e tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1016.703692] env[61957]: DEBUG nova.virt.hardware [None req-6a1be883-2588-47a0-91f4-274d9905657e tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1016.703864] env[61957]: DEBUG nova.virt.hardware [None req-6a1be883-2588-47a0-91f4-274d9905657e tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1016.704044] env[61957]: DEBUG nova.virt.hardware [None req-6a1be883-2588-47a0-91f4-274d9905657e tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Got 1 possible topologies {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1016.704241] env[61957]: DEBUG nova.virt.hardware [None req-6a1be883-2588-47a0-91f4-274d9905657e tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1016.704467] env[61957]: DEBUG nova.virt.hardware [None req-6a1be883-2588-47a0-91f4-274d9905657e tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1016.705548] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b256700-4282-4441-bc23-1edaec69c78e {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.714831] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f0cad08-55a3-460a-bd13-d38bfb43f264 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.744034] env[61957]: DEBUG oslo_concurrency.lockutils [None req-3409aaf8-ffe7-41b4-a265-d651c9473b75 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Lock "interface-08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2-00c8220d-5633-4de7-9795-8a6a8543c4ec" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 7.251s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1016.958542] env[61957]: DEBUG oslo_vmware.api [None req-ad83dc8d-b942-4b5a-8561-778a1542e9de tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Task: {'id': task-1277972, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.376518} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.961338] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-ad83dc8d-b942-4b5a-8561-778a1542e9de tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Deleted the datastore file {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1016.961539] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-ad83dc8d-b942-4b5a-8561-778a1542e9de tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] [instance: 9b0053fe-420e-4b92-86a8-ed44f6a6ec49] Deleted contents of the VM from datastore datastore1 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1016.961725] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-ad83dc8d-b942-4b5a-8561-778a1542e9de tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] [instance: 9b0053fe-420e-4b92-86a8-ed44f6a6ec49] Instance destroyed {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1016.981350] env[61957]: INFO nova.scheduler.client.report [None req-ad83dc8d-b942-4b5a-8561-778a1542e9de tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Deleted allocations for instance 9b0053fe-420e-4b92-86a8-ed44f6a6ec49 [ 1017.002362] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5628042-897b-42a4-a783-018b6f9987bf {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.010978] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f268311f-a28c-4803-be84-5e7388912fd0 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.046244] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b158d6e-b8aa-4427-b27b-53ad83fc5672 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.055413] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d018be3-5c19-47f9-aee2-cbdc2f7c7d8f {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.074133] env[61957]: DEBUG nova.compute.provider_tree [None req-1238a3ba-00b0-4ef8-9842-11dbb65e047b tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1017.130207] env[61957]: DEBUG oslo_concurrency.lockutils [None req-0b16b4fe-c18a-4e4e-8833-61daba0fafce tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Acquiring lock "6509707e-afc6-4c4a-88f9-a471865d510c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1017.130615] env[61957]: DEBUG oslo_concurrency.lockutils [None req-0b16b4fe-c18a-4e4e-8833-61daba0fafce tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Lock "6509707e-afc6-4c4a-88f9-a471865d510c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1017.130734] env[61957]: DEBUG oslo_concurrency.lockutils [None req-0b16b4fe-c18a-4e4e-8833-61daba0fafce tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Lock "6509707e-afc6-4c4a-88f9-a471865d510c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1017.161428] env[61957]: DEBUG oslo_vmware.api [None req-d92c7007-db9c-429d-aa03-41f7719669de tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1277973, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.506837} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.161712] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-d92c7007-db9c-429d-aa03-41f7719669de tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore1] 927af4c3-c7e8-404c-839e-d25af5e384a9/927af4c3-c7e8-404c-839e-d25af5e384a9.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1017.161929] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-d92c7007-db9c-429d-aa03-41f7719669de tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 927af4c3-c7e8-404c-839e-d25af5e384a9] Extending root virtual disk to 1048576 {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1017.162234] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-24a46779-6266-4483-b871-392e4a429140 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.168974] env[61957]: DEBUG oslo_vmware.api [None req-d92c7007-db9c-429d-aa03-41f7719669de tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Waiting for the task: (returnval){ [ 1017.168974] env[61957]: value = "task-1277974" [ 1017.168974] env[61957]: _type = "Task" [ 1017.168974] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.177901] env[61957]: DEBUG oslo_vmware.api [None req-d92c7007-db9c-429d-aa03-41f7719669de tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1277974, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.184536] env[61957]: DEBUG nova.compute.utils [None req-20123269-72b2-4ce0-b74a-dd98ef589cf3 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Using /dev/sd instead of None {{(pid=61957) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1017.486592] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ad83dc8d-b942-4b5a-8561-778a1542e9de tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1017.505394] env[61957]: DEBUG nova.network.neutron [req-c7fca03f-3e62-4fd1-8348-e3202c1fd9f1 req-48ffe962-23ce-49ff-adf5-842fb4551858 service nova] [instance: 927af4c3-c7e8-404c-839e-d25af5e384a9] Updated VIF entry in instance network info cache for port 38da1ba9-c5cc-4f76-b516-5ab8fdca9064. {{(pid=61957) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1017.505555] env[61957]: DEBUG nova.network.neutron [req-c7fca03f-3e62-4fd1-8348-e3202c1fd9f1 req-48ffe962-23ce-49ff-adf5-842fb4551858 service nova] [instance: 927af4c3-c7e8-404c-839e-d25af5e384a9] Updating instance_info_cache with network_info: [{"id": "38da1ba9-c5cc-4f76-b516-5ab8fdca9064", "address": "fa:16:3e:8e:54:ab", "network": {"id": "efe08823-51fe-4620-a849-5345e262cc71", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1035436819-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e16f6dce3f0e44fb96de516f17d4c6f5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "667a2e97-c1be-421d-9941-6b84c2629b43", "external-id": "nsx-vlan-transportzone-484", "segmentation_id": 484, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap38da1ba9-c5", "ovs_interfaceid": "38da1ba9-c5cc-4f76-b516-5ab8fdca9064", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1017.542753] env[61957]: DEBUG nova.network.neutron [req-16debee1-fc20-418c-8716-e7cb5152f756 req-dad78efb-9b50-4e74-a3a3-49109722847e service nova] [instance: 9b0053fe-420e-4b92-86a8-ed44f6a6ec49] Updated VIF entry in instance network info cache for port 219230f0-0b14-4be9-81bb-67a765d8ced9. {{(pid=61957) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1017.543070] env[61957]: DEBUG nova.network.neutron [req-16debee1-fc20-418c-8716-e7cb5152f756 req-dad78efb-9b50-4e74-a3a3-49109722847e service nova] [instance: 9b0053fe-420e-4b92-86a8-ed44f6a6ec49] Updating instance_info_cache with network_info: [{"id": "219230f0-0b14-4be9-81bb-67a765d8ced9", "address": "fa:16:3e:f6:9e:ce", "network": {"id": "cf378e9c-eaae-4168-bd60-9818fcfaf405", "bridge": null, "label": "tempest-ServersNegativeTestJSON-1243326371-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f7c20ab3c822423fa837a03f066e4239", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap219230f0-0b", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1017.577053] env[61957]: DEBUG nova.scheduler.client.report [None req-1238a3ba-00b0-4ef8-9842-11dbb65e047b tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1017.680029] env[61957]: DEBUG oslo_vmware.api [None req-d92c7007-db9c-429d-aa03-41f7719669de tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1277974, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.061621} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.680682] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-d92c7007-db9c-429d-aa03-41f7719669de tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 927af4c3-c7e8-404c-839e-d25af5e384a9] Extended root virtual disk {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1017.681158] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a0e3609-ff47-4692-a73a-f9fbad0756a4 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.686943] env[61957]: DEBUG oslo_concurrency.lockutils [None req-20123269-72b2-4ce0-b74a-dd98ef589cf3 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Lock "964c8c9d-0721-4ad3-a146-fce1fd1d6513" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.005s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1017.713261] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-d92c7007-db9c-429d-aa03-41f7719669de tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 927af4c3-c7e8-404c-839e-d25af5e384a9] Reconfiguring VM instance instance-00000060 to attach disk [datastore1] 927af4c3-c7e8-404c-839e-d25af5e384a9/927af4c3-c7e8-404c-839e-d25af5e384a9.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1017.713887] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d6c54dc0-e5cb-461e-a884-9700d3f5328c {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.734844] env[61957]: DEBUG oslo_vmware.api [None req-d92c7007-db9c-429d-aa03-41f7719669de tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Waiting for the task: (returnval){ [ 1017.734844] env[61957]: value = "task-1277975" [ 1017.734844] env[61957]: _type = "Task" [ 1017.734844] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.743592] env[61957]: DEBUG oslo_vmware.api [None req-d92c7007-db9c-429d-aa03-41f7719669de tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1277975, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.007830] env[61957]: DEBUG nova.compute.manager [req-2eee7345-1d69-47ca-8578-d0fe9f0acede req-9f6bf425-1765-44d4-8096-d302b4e00735 service nova] [instance: 3f144012-c0cc-43ce-8c73-4422fe8c1fe6] Received event network-vif-plugged-921a0283-f0c8-4cd9-8548-5e1089a438c3 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1018.008241] env[61957]: DEBUG oslo_concurrency.lockutils [req-2eee7345-1d69-47ca-8578-d0fe9f0acede req-9f6bf425-1765-44d4-8096-d302b4e00735 service nova] Acquiring lock "3f144012-c0cc-43ce-8c73-4422fe8c1fe6-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1018.008597] env[61957]: DEBUG oslo_concurrency.lockutils [req-2eee7345-1d69-47ca-8578-d0fe9f0acede req-9f6bf425-1765-44d4-8096-d302b4e00735 service nova] Lock "3f144012-c0cc-43ce-8c73-4422fe8c1fe6-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1018.008920] env[61957]: DEBUG oslo_concurrency.lockutils [req-2eee7345-1d69-47ca-8578-d0fe9f0acede req-9f6bf425-1765-44d4-8096-d302b4e00735 service nova] Lock "3f144012-c0cc-43ce-8c73-4422fe8c1fe6-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1018.009284] env[61957]: DEBUG nova.compute.manager [req-2eee7345-1d69-47ca-8578-d0fe9f0acede req-9f6bf425-1765-44d4-8096-d302b4e00735 service nova] [instance: 3f144012-c0cc-43ce-8c73-4422fe8c1fe6] No waiting events found dispatching network-vif-plugged-921a0283-f0c8-4cd9-8548-5e1089a438c3 {{(pid=61957) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1018.009596] env[61957]: WARNING nova.compute.manager [req-2eee7345-1d69-47ca-8578-d0fe9f0acede req-9f6bf425-1765-44d4-8096-d302b4e00735 service nova] [instance: 3f144012-c0cc-43ce-8c73-4422fe8c1fe6] Received unexpected event network-vif-plugged-921a0283-f0c8-4cd9-8548-5e1089a438c3 for instance with vm_state building and task_state spawning. [ 1018.010705] env[61957]: DEBUG oslo_concurrency.lockutils [req-c7fca03f-3e62-4fd1-8348-e3202c1fd9f1 req-48ffe962-23ce-49ff-adf5-842fb4551858 service nova] Releasing lock "refresh_cache-927af4c3-c7e8-404c-839e-d25af5e384a9" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1018.033489] env[61957]: DEBUG nova.network.neutron [None req-6a1be883-2588-47a0-91f4-274d9905657e tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 3f144012-c0cc-43ce-8c73-4422fe8c1fe6] Successfully updated port: 921a0283-f0c8-4cd9-8548-5e1089a438c3 {{(pid=61957) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1018.045265] env[61957]: DEBUG oslo_concurrency.lockutils [req-16debee1-fc20-418c-8716-e7cb5152f756 req-dad78efb-9b50-4e74-a3a3-49109722847e service nova] Releasing lock "refresh_cache-9b0053fe-420e-4b92-86a8-ed44f6a6ec49" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1018.082358] env[61957]: DEBUG oslo_concurrency.lockutils [None req-1238a3ba-00b0-4ef8-9842-11dbb65e047b tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.854s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1018.089501] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ad83dc8d-b942-4b5a-8561-778a1542e9de tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.603s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1018.089999] env[61957]: DEBUG nova.objects.instance [None req-ad83dc8d-b942-4b5a-8561-778a1542e9de tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Lazy-loading 'resources' on Instance uuid 9b0053fe-420e-4b92-86a8-ed44f6a6ec49 {{(pid=61957) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1018.108826] env[61957]: INFO nova.scheduler.client.report [None req-1238a3ba-00b0-4ef8-9842-11dbb65e047b tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Deleted allocations for instance 7793aa07-6aa2-459d-8a91-56c6b0412d3b [ 1018.167788] env[61957]: DEBUG oslo_concurrency.lockutils [None req-0b16b4fe-c18a-4e4e-8833-61daba0fafce tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Acquiring lock "refresh_cache-6509707e-afc6-4c4a-88f9-a471865d510c" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1018.167788] env[61957]: DEBUG oslo_concurrency.lockutils [None req-0b16b4fe-c18a-4e4e-8833-61daba0fafce tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Acquired lock "refresh_cache-6509707e-afc6-4c4a-88f9-a471865d510c" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1018.167788] env[61957]: DEBUG nova.network.neutron [None req-0b16b4fe-c18a-4e4e-8833-61daba0fafce tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: 6509707e-afc6-4c4a-88f9-a471865d510c] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1018.245803] env[61957]: DEBUG oslo_vmware.api [None req-d92c7007-db9c-429d-aa03-41f7719669de tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1277975, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.536077] env[61957]: DEBUG oslo_concurrency.lockutils [None req-6a1be883-2588-47a0-91f4-274d9905657e tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Acquiring lock "refresh_cache-3f144012-c0cc-43ce-8c73-4422fe8c1fe6" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1018.536245] env[61957]: DEBUG oslo_concurrency.lockutils [None req-6a1be883-2588-47a0-91f4-274d9905657e tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Acquired lock "refresh_cache-3f144012-c0cc-43ce-8c73-4422fe8c1fe6" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1018.536404] env[61957]: DEBUG nova.network.neutron [None req-6a1be883-2588-47a0-91f4-274d9905657e tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 3f144012-c0cc-43ce-8c73-4422fe8c1fe6] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1018.593463] env[61957]: DEBUG nova.objects.instance [None req-ad83dc8d-b942-4b5a-8561-778a1542e9de tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Lazy-loading 'numa_topology' on Instance uuid 9b0053fe-420e-4b92-86a8-ed44f6a6ec49 {{(pid=61957) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1018.615875] env[61957]: DEBUG oslo_concurrency.lockutils [None req-1238a3ba-00b0-4ef8-9842-11dbb65e047b tempest-ServerRescueTestJSON-1247544336 tempest-ServerRescueTestJSON-1247544336-project-member] Lock "7793aa07-6aa2-459d-8a91-56c6b0412d3b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.528s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1018.748274] env[61957]: DEBUG oslo_vmware.api [None req-d92c7007-db9c-429d-aa03-41f7719669de tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1277975, 'name': ReconfigVM_Task, 'duration_secs': 0.524262} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.748582] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-d92c7007-db9c-429d-aa03-41f7719669de tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 927af4c3-c7e8-404c-839e-d25af5e384a9] Reconfigured VM instance instance-00000060 to attach disk [datastore1] 927af4c3-c7e8-404c-839e-d25af5e384a9/927af4c3-c7e8-404c-839e-d25af5e384a9.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1018.749251] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a37e051a-d13a-473a-9593-09a5eb3df38a {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.758175] env[61957]: DEBUG oslo_concurrency.lockutils [None req-20123269-72b2-4ce0-b74a-dd98ef589cf3 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Acquiring lock "964c8c9d-0721-4ad3-a146-fce1fd1d6513" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1018.758175] env[61957]: DEBUG oslo_concurrency.lockutils [None req-20123269-72b2-4ce0-b74a-dd98ef589cf3 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Lock "964c8c9d-0721-4ad3-a146-fce1fd1d6513" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1018.758175] env[61957]: INFO nova.compute.manager [None req-20123269-72b2-4ce0-b74a-dd98ef589cf3 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] [instance: 964c8c9d-0721-4ad3-a146-fce1fd1d6513] Attaching volume 33adb5ac-f7e3-4840-a689-022217fc5da8 to /dev/sdb [ 1018.759663] env[61957]: DEBUG oslo_vmware.api [None req-d92c7007-db9c-429d-aa03-41f7719669de tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Waiting for the task: (returnval){ [ 1018.759663] env[61957]: value = "task-1277976" [ 1018.759663] env[61957]: _type = "Task" [ 1018.759663] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.774272] env[61957]: DEBUG nova.compute.manager [req-7cab942b-e658-46f6-a65e-01ba475564f7 req-c334358c-9181-4081-a7a7-8a2acd22dacc service nova] [instance: 3f144012-c0cc-43ce-8c73-4422fe8c1fe6] Received event network-changed-921a0283-f0c8-4cd9-8548-5e1089a438c3 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1018.774272] env[61957]: DEBUG nova.compute.manager [req-7cab942b-e658-46f6-a65e-01ba475564f7 req-c334358c-9181-4081-a7a7-8a2acd22dacc service nova] [instance: 3f144012-c0cc-43ce-8c73-4422fe8c1fe6] Refreshing instance network info cache due to event network-changed-921a0283-f0c8-4cd9-8548-5e1089a438c3. {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1018.774272] env[61957]: DEBUG oslo_concurrency.lockutils [req-7cab942b-e658-46f6-a65e-01ba475564f7 req-c334358c-9181-4081-a7a7-8a2acd22dacc service nova] Acquiring lock "refresh_cache-3f144012-c0cc-43ce-8c73-4422fe8c1fe6" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1018.781394] env[61957]: DEBUG oslo_vmware.api [None req-d92c7007-db9c-429d-aa03-41f7719669de tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1277976, 'name': Rename_Task} progress is 10%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.805757] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3163200-d6d4-4f5c-9185-df8a0e9f95e2 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.815333] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48d100cc-d6d6-4a86-bac6-3d6e333efc57 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.819700] env[61957]: DEBUG oslo_concurrency.lockutils [None req-437e09d9-7c87-4b34-90f6-4026cbd0876c tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Acquiring lock "e2107d5c-8453-4d24-b364-2b5c1d125062" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1018.819942] env[61957]: DEBUG oslo_concurrency.lockutils [None req-437e09d9-7c87-4b34-90f6-4026cbd0876c tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Lock "e2107d5c-8453-4d24-b364-2b5c1d125062" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1018.820221] env[61957]: DEBUG oslo_concurrency.lockutils [None req-437e09d9-7c87-4b34-90f6-4026cbd0876c tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Acquiring lock "e2107d5c-8453-4d24-b364-2b5c1d125062-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1018.820493] env[61957]: DEBUG oslo_concurrency.lockutils [None req-437e09d9-7c87-4b34-90f6-4026cbd0876c tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Lock "e2107d5c-8453-4d24-b364-2b5c1d125062-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1018.820677] env[61957]: DEBUG oslo_concurrency.lockutils [None req-437e09d9-7c87-4b34-90f6-4026cbd0876c tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Lock "e2107d5c-8453-4d24-b364-2b5c1d125062-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1018.822869] env[61957]: INFO nova.compute.manager [None req-437e09d9-7c87-4b34-90f6-4026cbd0876c tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: e2107d5c-8453-4d24-b364-2b5c1d125062] Terminating instance [ 1018.824706] env[61957]: DEBUG nova.compute.manager [None req-437e09d9-7c87-4b34-90f6-4026cbd0876c tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: e2107d5c-8453-4d24-b364-2b5c1d125062] Start destroying the instance on the hypervisor. {{(pid=61957) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1018.824899] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-437e09d9-7c87-4b34-90f6-4026cbd0876c tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: e2107d5c-8453-4d24-b364-2b5c1d125062] Destroying instance {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1018.825676] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa7e1abe-ddb2-447d-a0f5-bae628047965 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.837414] env[61957]: DEBUG nova.virt.block_device [None req-20123269-72b2-4ce0-b74a-dd98ef589cf3 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] [instance: 964c8c9d-0721-4ad3-a146-fce1fd1d6513] Updating existing volume attachment record: 723eeb75-90f1-4f13-98b5-915d6b6b0a57 {{(pid=61957) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1018.845332] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-437e09d9-7c87-4b34-90f6-4026cbd0876c tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: e2107d5c-8453-4d24-b364-2b5c1d125062] Powering off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1018.846411] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-667da977-f29b-4bbf-849b-17da228ae346 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.855607] env[61957]: DEBUG oslo_vmware.api [None req-437e09d9-7c87-4b34-90f6-4026cbd0876c tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Waiting for the task: (returnval){ [ 1018.855607] env[61957]: value = "task-1277977" [ 1018.855607] env[61957]: _type = "Task" [ 1018.855607] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.855607] env[61957]: DEBUG oslo_concurrency.lockutils [None req-4ca8b060-369e-429d-bf7a-53fb2d00bd30 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Acquiring lock "526ff179-62a6-4763-ab25-797617c4ed57" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1018.855607] env[61957]: DEBUG oslo_concurrency.lockutils [None req-4ca8b060-369e-429d-bf7a-53fb2d00bd30 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Lock "526ff179-62a6-4763-ab25-797617c4ed57" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1018.855607] env[61957]: DEBUG oslo_concurrency.lockutils [None req-4ca8b060-369e-429d-bf7a-53fb2d00bd30 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Acquiring lock "526ff179-62a6-4763-ab25-797617c4ed57-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1018.855607] env[61957]: DEBUG oslo_concurrency.lockutils [None req-4ca8b060-369e-429d-bf7a-53fb2d00bd30 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Lock "526ff179-62a6-4763-ab25-797617c4ed57-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1018.855607] env[61957]: DEBUG oslo_concurrency.lockutils [None req-4ca8b060-369e-429d-bf7a-53fb2d00bd30 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Lock "526ff179-62a6-4763-ab25-797617c4ed57-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1018.861395] env[61957]: INFO nova.compute.manager [None req-4ca8b060-369e-429d-bf7a-53fb2d00bd30 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] [instance: 526ff179-62a6-4763-ab25-797617c4ed57] Terminating instance [ 1018.864706] env[61957]: DEBUG nova.compute.manager [None req-4ca8b060-369e-429d-bf7a-53fb2d00bd30 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] [instance: 526ff179-62a6-4763-ab25-797617c4ed57] Start destroying the instance on the hypervisor. {{(pid=61957) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1018.864935] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-4ca8b060-369e-429d-bf7a-53fb2d00bd30 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] [instance: 526ff179-62a6-4763-ab25-797617c4ed57] Destroying instance {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1018.865918] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60e74ae2-ffa8-4603-8647-ef47fbc63c25 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.872927] env[61957]: DEBUG oslo_vmware.api [None req-437e09d9-7c87-4b34-90f6-4026cbd0876c tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': task-1277977, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.878577] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-4ca8b060-369e-429d-bf7a-53fb2d00bd30 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] [instance: 526ff179-62a6-4763-ab25-797617c4ed57] Powering off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1018.881140] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b1dcbc39-8e2b-441d-a7ed-478e2b46c4e7 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.893535] env[61957]: DEBUG oslo_vmware.api [None req-4ca8b060-369e-429d-bf7a-53fb2d00bd30 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Waiting for the task: (returnval){ [ 1018.893535] env[61957]: value = "task-1277978" [ 1018.893535] env[61957]: _type = "Task" [ 1018.893535] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.906250] env[61957]: DEBUG oslo_vmware.api [None req-4ca8b060-369e-429d-bf7a-53fb2d00bd30 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Task: {'id': task-1277978, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.921512] env[61957]: DEBUG nova.network.neutron [None req-0b16b4fe-c18a-4e4e-8833-61daba0fafce tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: 6509707e-afc6-4c4a-88f9-a471865d510c] Updating instance_info_cache with network_info: [{"id": "7df2736b-10d0-4f9d-83aa-e2bbdf1ef45b", "address": "fa:16:3e:2e:da:a9", "network": {"id": "6fbccc17-6674-4cd5-8f1b-aa73570baac7", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-394544811-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1dbdc699741a48af93c52068d88a5357", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "20e3f794-c7a3-4696-9488-ecf34c570ef9", "external-id": "nsx-vlan-transportzone-509", "segmentation_id": 509, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7df2736b-10", "ovs_interfaceid": "7df2736b-10d0-4f9d-83aa-e2bbdf1ef45b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1019.095140] env[61957]: DEBUG nova.objects.base [None req-ad83dc8d-b942-4b5a-8561-778a1542e9de tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Object Instance<9b0053fe-420e-4b92-86a8-ed44f6a6ec49> lazy-loaded attributes: resources,numa_topology {{(pid=61957) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1019.106736] env[61957]: DEBUG nova.network.neutron [None req-6a1be883-2588-47a0-91f4-274d9905657e tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 3f144012-c0cc-43ce-8c73-4422fe8c1fe6] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1019.264021] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e7afaa4-32a0-4640-bbf5-a21815973866 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.273626] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b42b68a-2d11-444f-90c2-50fceae654dd {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.281460] env[61957]: DEBUG oslo_vmware.api [None req-d92c7007-db9c-429d-aa03-41f7719669de tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1277976, 'name': Rename_Task, 'duration_secs': 0.163493} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.282252] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-d92c7007-db9c-429d-aa03-41f7719669de tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 927af4c3-c7e8-404c-839e-d25af5e384a9] Powering on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1019.282539] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-74382b6a-284a-4107-bb74-82d0ccd1a1f3 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.316585] env[61957]: DEBUG oslo_concurrency.lockutils [None req-c6a91fc7-15fa-404b-a88f-c93ede8d0713 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Acquiring lock "interface-08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2-6a70445e-1266-453e-a6ff-50e8cbdcfd08" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1019.317032] env[61957]: DEBUG oslo_concurrency.lockutils [None req-c6a91fc7-15fa-404b-a88f-c93ede8d0713 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Lock "interface-08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2-6a70445e-1266-453e-a6ff-50e8cbdcfd08" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.001s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1019.323593] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3641799e-7c92-4eb5-92f4-5c791dd79494 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.327194] env[61957]: DEBUG oslo_vmware.api [None req-d92c7007-db9c-429d-aa03-41f7719669de tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Waiting for the task: (returnval){ [ 1019.327194] env[61957]: value = "task-1277980" [ 1019.327194] env[61957]: _type = "Task" [ 1019.327194] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.337400] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37f10728-cb43-47d6-9fc6-2327cc45c823 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.346699] env[61957]: DEBUG oslo_vmware.api [None req-d92c7007-db9c-429d-aa03-41f7719669de tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1277980, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.359785] env[61957]: DEBUG nova.compute.provider_tree [None req-ad83dc8d-b942-4b5a-8561-778a1542e9de tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1019.372248] env[61957]: DEBUG oslo_vmware.api [None req-437e09d9-7c87-4b34-90f6-4026cbd0876c tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': task-1277977, 'name': PowerOffVM_Task, 'duration_secs': 0.26358} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.372619] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-437e09d9-7c87-4b34-90f6-4026cbd0876c tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: e2107d5c-8453-4d24-b364-2b5c1d125062] Powered off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1019.372855] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-437e09d9-7c87-4b34-90f6-4026cbd0876c tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: e2107d5c-8453-4d24-b364-2b5c1d125062] Unregistering the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1019.373262] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-bb39a113-786d-4eb6-b89b-eac04225666c {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.410983] env[61957]: DEBUG oslo_vmware.api [None req-4ca8b060-369e-429d-bf7a-53fb2d00bd30 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Task: {'id': task-1277978, 'name': PowerOffVM_Task, 'duration_secs': 0.330339} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.410983] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-4ca8b060-369e-429d-bf7a-53fb2d00bd30 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] [instance: 526ff179-62a6-4763-ab25-797617c4ed57] Powered off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1019.410983] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-4ca8b060-369e-429d-bf7a-53fb2d00bd30 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] [instance: 526ff179-62a6-4763-ab25-797617c4ed57] Unregistering the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1019.410983] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9090ace7-f359-420d-af33-f76888661895 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.410983] env[61957]: DEBUG nova.network.neutron [None req-6a1be883-2588-47a0-91f4-274d9905657e tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 3f144012-c0cc-43ce-8c73-4422fe8c1fe6] Updating instance_info_cache with network_info: [{"id": "921a0283-f0c8-4cd9-8548-5e1089a438c3", "address": "fa:16:3e:ee:de:e6", "network": {"id": "4f0e57be-8906-48b5-865f-59ff12250a8b", "bridge": "br-int", "label": "tempest-ServersTestJSON-1916182687-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "65fd71bcd1bf41238a9cc3a5d6dd4924", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "939c05b6-8f31-4f3a-95ac-6297e0bd243e", "external-id": "nsx-vlan-transportzone-825", "segmentation_id": 825, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap921a0283-f0", "ovs_interfaceid": "921a0283-f0c8-4cd9-8548-5e1089a438c3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1019.424391] env[61957]: DEBUG oslo_concurrency.lockutils [None req-0b16b4fe-c18a-4e4e-8833-61daba0fafce tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Releasing lock "refresh_cache-6509707e-afc6-4c4a-88f9-a471865d510c" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1019.451492] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-437e09d9-7c87-4b34-90f6-4026cbd0876c tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: e2107d5c-8453-4d24-b364-2b5c1d125062] Unregistered the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1019.451884] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-437e09d9-7c87-4b34-90f6-4026cbd0876c tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: e2107d5c-8453-4d24-b364-2b5c1d125062] Deleting contents of the VM from datastore datastore2 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1019.452598] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-437e09d9-7c87-4b34-90f6-4026cbd0876c tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Deleting the datastore file [datastore2] e2107d5c-8453-4d24-b364-2b5c1d125062 {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1019.453109] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a66d9ab8-5867-4815-8457-8139570b6a96 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.464408] env[61957]: DEBUG oslo_vmware.api [None req-437e09d9-7c87-4b34-90f6-4026cbd0876c tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Waiting for the task: (returnval){ [ 1019.464408] env[61957]: value = "task-1277983" [ 1019.464408] env[61957]: _type = "Task" [ 1019.464408] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.484036] env[61957]: DEBUG oslo_vmware.api [None req-437e09d9-7c87-4b34-90f6-4026cbd0876c tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': task-1277983, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.487328] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-4ca8b060-369e-429d-bf7a-53fb2d00bd30 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] [instance: 526ff179-62a6-4763-ab25-797617c4ed57] Unregistered the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1019.487645] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-4ca8b060-369e-429d-bf7a-53fb2d00bd30 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] [instance: 526ff179-62a6-4763-ab25-797617c4ed57] Deleting contents of the VM from datastore datastore2 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1019.487921] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-4ca8b060-369e-429d-bf7a-53fb2d00bd30 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Deleting the datastore file [datastore2] 526ff179-62a6-4763-ab25-797617c4ed57 {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1019.488621] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f1bac968-4f55-43b1-a363-5e416b5f92c5 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.499648] env[61957]: DEBUG oslo_vmware.api [None req-4ca8b060-369e-429d-bf7a-53fb2d00bd30 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Waiting for the task: (returnval){ [ 1019.499648] env[61957]: value = "task-1277984" [ 1019.499648] env[61957]: _type = "Task" [ 1019.499648] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.511756] env[61957]: DEBUG oslo_vmware.api [None req-4ca8b060-369e-429d-bf7a-53fb2d00bd30 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Task: {'id': task-1277984, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.820544] env[61957]: DEBUG oslo_concurrency.lockutils [None req-c6a91fc7-15fa-404b-a88f-c93ede8d0713 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Acquiring lock "08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1019.820980] env[61957]: DEBUG oslo_concurrency.lockutils [None req-c6a91fc7-15fa-404b-a88f-c93ede8d0713 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Acquired lock "08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1019.821731] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-690e2fd7-48fa-4b19-9265-0b8ece60ca16 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.849742] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77ffe736-3314-4659-8a88-5cfe314c6bdb {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.887641] env[61957]: DEBUG nova.scheduler.client.report [None req-ad83dc8d-b942-4b5a-8561-778a1542e9de tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1019.900546] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-c6a91fc7-15fa-404b-a88f-c93ede8d0713 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2] Reconfiguring VM to detach interface {{(pid=61957) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1974}} [ 1019.903840] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ff3a2d97-0c68-4999-a09c-18930800ec03 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.918541] env[61957]: DEBUG oslo_vmware.api [None req-d92c7007-db9c-429d-aa03-41f7719669de tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1277980, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.919860] env[61957]: DEBUG oslo_concurrency.lockutils [None req-6a1be883-2588-47a0-91f4-274d9905657e tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Releasing lock "refresh_cache-3f144012-c0cc-43ce-8c73-4422fe8c1fe6" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1019.920209] env[61957]: DEBUG nova.compute.manager [None req-6a1be883-2588-47a0-91f4-274d9905657e tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 3f144012-c0cc-43ce-8c73-4422fe8c1fe6] Instance network_info: |[{"id": "921a0283-f0c8-4cd9-8548-5e1089a438c3", "address": "fa:16:3e:ee:de:e6", "network": {"id": "4f0e57be-8906-48b5-865f-59ff12250a8b", "bridge": "br-int", "label": "tempest-ServersTestJSON-1916182687-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "65fd71bcd1bf41238a9cc3a5d6dd4924", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "939c05b6-8f31-4f3a-95ac-6297e0bd243e", "external-id": "nsx-vlan-transportzone-825", "segmentation_id": 825, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap921a0283-f0", "ovs_interfaceid": "921a0283-f0c8-4cd9-8548-5e1089a438c3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1019.920520] env[61957]: DEBUG oslo_concurrency.lockutils [req-7cab942b-e658-46f6-a65e-01ba475564f7 req-c334358c-9181-4081-a7a7-8a2acd22dacc service nova] Acquired lock "refresh_cache-3f144012-c0cc-43ce-8c73-4422fe8c1fe6" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1019.920699] env[61957]: DEBUG nova.network.neutron [req-7cab942b-e658-46f6-a65e-01ba475564f7 req-c334358c-9181-4081-a7a7-8a2acd22dacc service nova] [instance: 3f144012-c0cc-43ce-8c73-4422fe8c1fe6] Refreshing network info cache for port 921a0283-f0c8-4cd9-8548-5e1089a438c3 {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1019.925335] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-6a1be883-2588-47a0-91f4-274d9905657e tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 3f144012-c0cc-43ce-8c73-4422fe8c1fe6] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ee:de:e6', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '939c05b6-8f31-4f3a-95ac-6297e0bd243e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '921a0283-f0c8-4cd9-8548-5e1089a438c3', 'vif_model': 'vmxnet3'}] {{(pid=61957) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1019.932776] env[61957]: DEBUG oslo.service.loopingcall [None req-6a1be883-2588-47a0-91f4-274d9905657e tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1019.937255] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3f144012-c0cc-43ce-8c73-4422fe8c1fe6] Creating VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1019.940187] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a7b08684-0213-4240-b20f-a8d978f08d07 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.957121] env[61957]: DEBUG oslo_vmware.api [None req-c6a91fc7-15fa-404b-a88f-c93ede8d0713 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Waiting for the task: (returnval){ [ 1019.957121] env[61957]: value = "task-1277985" [ 1019.957121] env[61957]: _type = "Task" [ 1019.957121] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.964096] env[61957]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1019.964096] env[61957]: value = "task-1277986" [ 1019.964096] env[61957]: _type = "Task" [ 1019.964096] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.970835] env[61957]: DEBUG oslo_vmware.api [None req-c6a91fc7-15fa-404b-a88f-c93ede8d0713 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': task-1277985, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.976260] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1544eed6-8734-4431-8612-c1de40629c25 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.982657] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277986, 'name': CreateVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.989107] env[61957]: DEBUG oslo_vmware.api [None req-437e09d9-7c87-4b34-90f6-4026cbd0876c tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Task: {'id': task-1277983, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.168921} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.003424] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-437e09d9-7c87-4b34-90f6-4026cbd0876c tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Deleted the datastore file {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1020.003636] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-437e09d9-7c87-4b34-90f6-4026cbd0876c tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: e2107d5c-8453-4d24-b364-2b5c1d125062] Deleted contents of the VM from datastore datastore2 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1020.003820] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-437e09d9-7c87-4b34-90f6-4026cbd0876c tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: e2107d5c-8453-4d24-b364-2b5c1d125062] Instance destroyed {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1020.003999] env[61957]: INFO nova.compute.manager [None req-437e09d9-7c87-4b34-90f6-4026cbd0876c tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] [instance: e2107d5c-8453-4d24-b364-2b5c1d125062] Took 1.18 seconds to destroy the instance on the hypervisor. [ 1020.004315] env[61957]: DEBUG oslo.service.loopingcall [None req-437e09d9-7c87-4b34-90f6-4026cbd0876c tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1020.004766] env[61957]: DEBUG nova.compute.manager [-] [instance: e2107d5c-8453-4d24-b364-2b5c1d125062] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1020.004859] env[61957]: DEBUG nova.network.neutron [-] [instance: e2107d5c-8453-4d24-b364-2b5c1d125062] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1020.010409] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92d0532e-aa77-4975-8f45-aab4ad7f9e2f {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.019223] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-0b16b4fe-c18a-4e4e-8833-61daba0fafce tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: 6509707e-afc6-4c4a-88f9-a471865d510c] Updating instance '6509707e-afc6-4c4a-88f9-a471865d510c' progress to 83 {{(pid=61957) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1020.026501] env[61957]: DEBUG oslo_vmware.api [None req-4ca8b060-369e-429d-bf7a-53fb2d00bd30 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Task: {'id': task-1277984, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.181778} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.027066] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-4ca8b060-369e-429d-bf7a-53fb2d00bd30 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Deleted the datastore file {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1020.027339] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-4ca8b060-369e-429d-bf7a-53fb2d00bd30 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] [instance: 526ff179-62a6-4763-ab25-797617c4ed57] Deleted contents of the VM from datastore datastore2 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1020.027452] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-4ca8b060-369e-429d-bf7a-53fb2d00bd30 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] [instance: 526ff179-62a6-4763-ab25-797617c4ed57] Instance destroyed {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1020.027626] env[61957]: INFO nova.compute.manager [None req-4ca8b060-369e-429d-bf7a-53fb2d00bd30 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] [instance: 526ff179-62a6-4763-ab25-797617c4ed57] Took 1.16 seconds to destroy the instance on the hypervisor. [ 1020.028773] env[61957]: DEBUG oslo.service.loopingcall [None req-4ca8b060-369e-429d-bf7a-53fb2d00bd30 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1020.028773] env[61957]: DEBUG nova.compute.manager [-] [instance: 526ff179-62a6-4763-ab25-797617c4ed57] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1020.028773] env[61957]: DEBUG nova.network.neutron [-] [instance: 526ff179-62a6-4763-ab25-797617c4ed57] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1020.153446] env[61957]: DEBUG oslo_concurrency.lockutils [None req-3ec44fbf-1f22-4ad1-8b26-d0253bbe3176 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Acquiring lock "9b0053fe-420e-4b92-86a8-ed44f6a6ec49" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1020.355416] env[61957]: DEBUG oslo_vmware.api [None req-d92c7007-db9c-429d-aa03-41f7719669de tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1277980, 'name': PowerOnVM_Task, 'duration_secs': 0.592124} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.355731] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-d92c7007-db9c-429d-aa03-41f7719669de tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 927af4c3-c7e8-404c-839e-d25af5e384a9] Powered on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1020.355954] env[61957]: INFO nova.compute.manager [None req-d92c7007-db9c-429d-aa03-41f7719669de tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 927af4c3-c7e8-404c-839e-d25af5e384a9] Took 8.34 seconds to spawn the instance on the hypervisor. [ 1020.356172] env[61957]: DEBUG nova.compute.manager [None req-d92c7007-db9c-429d-aa03-41f7719669de tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 927af4c3-c7e8-404c-839e-d25af5e384a9] Checking state {{(pid=61957) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1020.357054] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62956361-048b-4773-8721-4ad91ca87082 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.401726] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ad83dc8d-b942-4b5a-8561-778a1542e9de tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.312s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1021.246445] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b16b4fe-c18a-4e4e-8833-61daba0fafce tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: 6509707e-afc6-4c4a-88f9-a471865d510c] Powering on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1021.254741] env[61957]: DEBUG nova.compute.manager [req-5aaaf87d-1333-47c1-8309-140c6f9e6e18 req-261b415f-473d-496e-8c79-6faeba12b049 service nova] [instance: e2107d5c-8453-4d24-b364-2b5c1d125062] Received event network-vif-deleted-f17ed646-b16b-40e9-a98f-a1324721caa1 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1021.254955] env[61957]: INFO nova.compute.manager [req-5aaaf87d-1333-47c1-8309-140c6f9e6e18 req-261b415f-473d-496e-8c79-6faeba12b049 service nova] [instance: e2107d5c-8453-4d24-b364-2b5c1d125062] Neutron deleted interface f17ed646-b16b-40e9-a98f-a1324721caa1; detaching it from the instance and deleting it from the info cache [ 1021.259019] env[61957]: DEBUG nova.network.neutron [req-5aaaf87d-1333-47c1-8309-140c6f9e6e18 req-261b415f-473d-496e-8c79-6faeba12b049 service nova] [instance: e2107d5c-8453-4d24-b364-2b5c1d125062] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1021.259019] env[61957]: DEBUG oslo_vmware.api [None req-c6a91fc7-15fa-404b-a88f-c93ede8d0713 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': task-1277985, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.259019] env[61957]: WARNING oslo_vmware.common.loopingcall [None req-c6a91fc7-15fa-404b-a88f-c93ede8d0713 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] task run outlasted interval by 0.29948600000000003 sec [ 1021.259621] env[61957]: INFO nova.compute.manager [None req-d92c7007-db9c-429d-aa03-41f7719669de tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 927af4c3-c7e8-404c-839e-d25af5e384a9] Took 13.58 seconds to build instance. [ 1021.263727] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ad83dc8d-b942-4b5a-8561-778a1542e9de tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Lock "9b0053fe-420e-4b92-86a8-ed44f6a6ec49" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 23.187s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1021.265064] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f3fe1af1-08ce-4983-9e79-721029cde8d7 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.269570] env[61957]: DEBUG oslo_concurrency.lockutils [None req-3ec44fbf-1f22-4ad1-8b26-d0253bbe3176 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Lock "9b0053fe-420e-4b92-86a8-ed44f6a6ec49" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 1.116s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1021.269887] env[61957]: INFO nova.compute.manager [None req-3ec44fbf-1f22-4ad1-8b26-d0253bbe3176 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] [instance: 9b0053fe-420e-4b92-86a8-ed44f6a6ec49] Unshelving [ 1021.279797] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1277986, 'name': CreateVM_Task, 'duration_secs': 0.624606} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.280868] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3f144012-c0cc-43ce-8c73-4422fe8c1fe6] Created VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1021.282824] env[61957]: DEBUG oslo_concurrency.lockutils [None req-6a1be883-2588-47a0-91f4-274d9905657e tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1021.283090] env[61957]: DEBUG oslo_concurrency.lockutils [None req-6a1be883-2588-47a0-91f4-274d9905657e tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Acquired lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1021.283522] env[61957]: DEBUG oslo_concurrency.lockutils [None req-6a1be883-2588-47a0-91f4-274d9905657e tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1021.284645] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0d50b4a7-7fdf-43bc-b486-fae9d3458cb8 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.297171] env[61957]: DEBUG oslo_vmware.api [None req-c6a91fc7-15fa-404b-a88f-c93ede8d0713 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': task-1277985, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.297171] env[61957]: DEBUG oslo_vmware.api [None req-0b16b4fe-c18a-4e4e-8833-61daba0fafce tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Waiting for the task: (returnval){ [ 1021.297171] env[61957]: value = "task-1277988" [ 1021.297171] env[61957]: _type = "Task" [ 1021.297171] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.301355] env[61957]: DEBUG oslo_vmware.api [None req-6a1be883-2588-47a0-91f4-274d9905657e tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Waiting for the task: (returnval){ [ 1021.301355] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52f3ba74-8a83-61ed-b9ce-56e4aca02759" [ 1021.301355] env[61957]: _type = "Task" [ 1021.301355] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.313619] env[61957]: DEBUG oslo_vmware.api [None req-0b16b4fe-c18a-4e4e-8833-61daba0fafce tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': task-1277988, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.325214] env[61957]: DEBUG oslo_vmware.api [None req-6a1be883-2588-47a0-91f4-274d9905657e tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52f3ba74-8a83-61ed-b9ce-56e4aca02759, 'name': SearchDatastore_Task, 'duration_secs': 0.011162} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.325648] env[61957]: DEBUG oslo_concurrency.lockutils [None req-6a1be883-2588-47a0-91f4-274d9905657e tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Releasing lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1021.326010] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-6a1be883-2588-47a0-91f4-274d9905657e tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 3f144012-c0cc-43ce-8c73-4422fe8c1fe6] Processing image 11c76a2c-f705-470a-ba9d-4657858bab38 {{(pid=61957) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1021.328784] env[61957]: DEBUG oslo_concurrency.lockutils [None req-6a1be883-2588-47a0-91f4-274d9905657e tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1021.328784] env[61957]: DEBUG oslo_concurrency.lockutils [None req-6a1be883-2588-47a0-91f4-274d9905657e tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Acquired lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1021.328784] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-6a1be883-2588-47a0-91f4-274d9905657e tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1021.328784] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-866e3cdc-493f-4747-87e9-e13f22ce68f7 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.337133] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-6a1be883-2588-47a0-91f4-274d9905657e tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1021.337508] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-6a1be883-2588-47a0-91f4-274d9905657e tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61957) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1021.338260] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d6209358-9ecb-4606-b9f6-3aa9f681fa4a {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.346599] env[61957]: DEBUG oslo_vmware.api [None req-6a1be883-2588-47a0-91f4-274d9905657e tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Waiting for the task: (returnval){ [ 1021.346599] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]5268408a-37e0-4f03-23ae-b4bb8f7b1f62" [ 1021.346599] env[61957]: _type = "Task" [ 1021.346599] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.354958] env[61957]: DEBUG oslo_vmware.api [None req-6a1be883-2588-47a0-91f4-274d9905657e tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]5268408a-37e0-4f03-23ae-b4bb8f7b1f62, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.496900] env[61957]: DEBUG nova.network.neutron [req-7cab942b-e658-46f6-a65e-01ba475564f7 req-c334358c-9181-4081-a7a7-8a2acd22dacc service nova] [instance: 3f144012-c0cc-43ce-8c73-4422fe8c1fe6] Updated VIF entry in instance network info cache for port 921a0283-f0c8-4cd9-8548-5e1089a438c3. {{(pid=61957) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1021.497378] env[61957]: DEBUG nova.network.neutron [req-7cab942b-e658-46f6-a65e-01ba475564f7 req-c334358c-9181-4081-a7a7-8a2acd22dacc service nova] [instance: 3f144012-c0cc-43ce-8c73-4422fe8c1fe6] Updating instance_info_cache with network_info: [{"id": "921a0283-f0c8-4cd9-8548-5e1089a438c3", "address": "fa:16:3e:ee:de:e6", "network": {"id": "4f0e57be-8906-48b5-865f-59ff12250a8b", "bridge": "br-int", "label": "tempest-ServersTestJSON-1916182687-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "65fd71bcd1bf41238a9cc3a5d6dd4924", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "939c05b6-8f31-4f3a-95ac-6297e0bd243e", "external-id": "nsx-vlan-transportzone-825", "segmentation_id": 825, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap921a0283-f0", "ovs_interfaceid": "921a0283-f0c8-4cd9-8548-5e1089a438c3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1021.737695] env[61957]: DEBUG nova.network.neutron [-] [instance: e2107d5c-8453-4d24-b364-2b5c1d125062] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1021.744960] env[61957]: DEBUG nova.network.neutron [-] [instance: 526ff179-62a6-4763-ab25-797617c4ed57] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1021.758793] env[61957]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1e331153-8b89-4f90-bb62-2c5e58604e86 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.764264] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d92c7007-db9c-429d-aa03-41f7719669de tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Lock "927af4c3-c7e8-404c-839e-d25af5e384a9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.096s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1021.777853] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8aabdda8-35f8-478a-9c12-74a52dd8c567 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.802956] env[61957]: DEBUG oslo_vmware.api [None req-c6a91fc7-15fa-404b-a88f-c93ede8d0713 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': task-1277985, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.816761] env[61957]: DEBUG nova.compute.manager [req-5aaaf87d-1333-47c1-8309-140c6f9e6e18 req-261b415f-473d-496e-8c79-6faeba12b049 service nova] [instance: e2107d5c-8453-4d24-b364-2b5c1d125062] Detach interface failed, port_id=f17ed646-b16b-40e9-a98f-a1324721caa1, reason: Instance e2107d5c-8453-4d24-b364-2b5c1d125062 could not be found. {{(pid=61957) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1021.821107] env[61957]: DEBUG oslo_vmware.api [None req-0b16b4fe-c18a-4e4e-8833-61daba0fafce tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': task-1277988, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.859339] env[61957]: DEBUG oslo_vmware.api [None req-6a1be883-2588-47a0-91f4-274d9905657e tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]5268408a-37e0-4f03-23ae-b4bb8f7b1f62, 'name': SearchDatastore_Task, 'duration_secs': 0.008325} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.860602] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-150722ef-0304-4b07-9c96-a68428c6d297 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.867907] env[61957]: DEBUG oslo_vmware.api [None req-6a1be883-2588-47a0-91f4-274d9905657e tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Waiting for the task: (returnval){ [ 1021.867907] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]523f9274-1fb5-c880-9060-bcc640c0668e" [ 1021.867907] env[61957]: _type = "Task" [ 1021.867907] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.877496] env[61957]: DEBUG oslo_vmware.api [None req-6a1be883-2588-47a0-91f4-274d9905657e tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]523f9274-1fb5-c880-9060-bcc640c0668e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.000743] env[61957]: DEBUG oslo_concurrency.lockutils [req-7cab942b-e658-46f6-a65e-01ba475564f7 req-c334358c-9181-4081-a7a7-8a2acd22dacc service nova] Releasing lock "refresh_cache-3f144012-c0cc-43ce-8c73-4422fe8c1fe6" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1022.129317] env[61957]: DEBUG oslo_concurrency.lockutils [None req-987070d8-27b8-45f7-833c-32d3bb141bbb tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Acquiring lock "927af4c3-c7e8-404c-839e-d25af5e384a9" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1022.129598] env[61957]: DEBUG oslo_concurrency.lockutils [None req-987070d8-27b8-45f7-833c-32d3bb141bbb tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Lock "927af4c3-c7e8-404c-839e-d25af5e384a9" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1022.129787] env[61957]: DEBUG nova.compute.manager [None req-987070d8-27b8-45f7-833c-32d3bb141bbb tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 927af4c3-c7e8-404c-839e-d25af5e384a9] Checking state {{(pid=61957) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1022.130807] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a191fd3b-46fb-485e-bfe2-827a3373a981 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.137976] env[61957]: DEBUG nova.compute.manager [None req-987070d8-27b8-45f7-833c-32d3bb141bbb tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 927af4c3-c7e8-404c-839e-d25af5e384a9] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61957) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1022.138536] env[61957]: DEBUG nova.objects.instance [None req-987070d8-27b8-45f7-833c-32d3bb141bbb tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Lazy-loading 'flavor' on Instance uuid 927af4c3-c7e8-404c-839e-d25af5e384a9 {{(pid=61957) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1022.240931] env[61957]: INFO nova.compute.manager [-] [instance: e2107d5c-8453-4d24-b364-2b5c1d125062] Took 2.24 seconds to deallocate network for instance. [ 1022.247580] env[61957]: INFO nova.compute.manager [-] [instance: 526ff179-62a6-4763-ab25-797617c4ed57] Took 2.22 seconds to deallocate network for instance. [ 1022.282996] env[61957]: DEBUG oslo_vmware.api [None req-c6a91fc7-15fa-404b-a88f-c93ede8d0713 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': task-1277985, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.315679] env[61957]: DEBUG oslo_vmware.api [None req-0b16b4fe-c18a-4e4e-8833-61daba0fafce tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': task-1277988, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.323598] env[61957]: DEBUG oslo_concurrency.lockutils [None req-3ec44fbf-1f22-4ad1-8b26-d0253bbe3176 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1022.323749] env[61957]: DEBUG oslo_concurrency.lockutils [None req-3ec44fbf-1f22-4ad1-8b26-d0253bbe3176 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1022.324077] env[61957]: DEBUG nova.objects.instance [None req-3ec44fbf-1f22-4ad1-8b26-d0253bbe3176 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Lazy-loading 'pci_requests' on Instance uuid 9b0053fe-420e-4b92-86a8-ed44f6a6ec49 {{(pid=61957) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1022.380221] env[61957]: DEBUG oslo_vmware.api [None req-6a1be883-2588-47a0-91f4-274d9905657e tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]523f9274-1fb5-c880-9060-bcc640c0668e, 'name': SearchDatastore_Task, 'duration_secs': 0.009593} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.380368] env[61957]: DEBUG oslo_concurrency.lockutils [None req-6a1be883-2588-47a0-91f4-274d9905657e tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Releasing lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1022.380648] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-6a1be883-2588-47a0-91f4-274d9905657e tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore1] 3f144012-c0cc-43ce-8c73-4422fe8c1fe6/3f144012-c0cc-43ce-8c73-4422fe8c1fe6.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1022.381087] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d73161ae-957a-4db8-a6f3-330c8af88e87 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.388894] env[61957]: DEBUG oslo_vmware.api [None req-6a1be883-2588-47a0-91f4-274d9905657e tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Waiting for the task: (returnval){ [ 1022.388894] env[61957]: value = "task-1277989" [ 1022.388894] env[61957]: _type = "Task" [ 1022.388894] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.398607] env[61957]: DEBUG oslo_vmware.api [None req-6a1be883-2588-47a0-91f4-274d9905657e tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': task-1277989, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.645065] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-987070d8-27b8-45f7-833c-32d3bb141bbb tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 927af4c3-c7e8-404c-839e-d25af5e384a9] Powering off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1022.645253] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3dc24c6a-eca4-4fbc-b4d3-0be9401eba92 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.654808] env[61957]: DEBUG oslo_vmware.api [None req-987070d8-27b8-45f7-833c-32d3bb141bbb tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Waiting for the task: (returnval){ [ 1022.654808] env[61957]: value = "task-1277990" [ 1022.654808] env[61957]: _type = "Task" [ 1022.654808] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.667487] env[61957]: DEBUG oslo_vmware.api [None req-987070d8-27b8-45f7-833c-32d3bb141bbb tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1277990, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.748123] env[61957]: DEBUG oslo_concurrency.lockutils [None req-437e09d9-7c87-4b34-90f6-4026cbd0876c tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1022.755564] env[61957]: DEBUG oslo_concurrency.lockutils [None req-4ca8b060-369e-429d-bf7a-53fb2d00bd30 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1022.782450] env[61957]: DEBUG oslo_vmware.api [None req-c6a91fc7-15fa-404b-a88f-c93ede8d0713 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': task-1277985, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.816344] env[61957]: DEBUG oslo_vmware.api [None req-0b16b4fe-c18a-4e4e-8833-61daba0fafce tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': task-1277988, 'name': PowerOnVM_Task, 'duration_secs': 1.483404} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.816697] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b16b4fe-c18a-4e4e-8833-61daba0fafce tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: 6509707e-afc6-4c4a-88f9-a471865d510c] Powered on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1022.816958] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-0b16b4fe-c18a-4e4e-8833-61daba0fafce tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: 6509707e-afc6-4c4a-88f9-a471865d510c] Updating instance '6509707e-afc6-4c4a-88f9-a471865d510c' progress to 100 {{(pid=61957) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1022.828186] env[61957]: DEBUG nova.objects.instance [None req-3ec44fbf-1f22-4ad1-8b26-d0253bbe3176 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Lazy-loading 'numa_topology' on Instance uuid 9b0053fe-420e-4b92-86a8-ed44f6a6ec49 {{(pid=61957) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1022.900571] env[61957]: DEBUG oslo_vmware.api [None req-6a1be883-2588-47a0-91f4-274d9905657e tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': task-1277989, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.498069} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.900866] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-6a1be883-2588-47a0-91f4-274d9905657e tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore1] 3f144012-c0cc-43ce-8c73-4422fe8c1fe6/3f144012-c0cc-43ce-8c73-4422fe8c1fe6.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1022.901119] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-6a1be883-2588-47a0-91f4-274d9905657e tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 3f144012-c0cc-43ce-8c73-4422fe8c1fe6] Extending root virtual disk to 1048576 {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1022.901400] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-cc4c38b6-4a96-403e-b723-897982c6189d {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.909093] env[61957]: DEBUG oslo_vmware.api [None req-6a1be883-2588-47a0-91f4-274d9905657e tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Waiting for the task: (returnval){ [ 1022.909093] env[61957]: value = "task-1277991" [ 1022.909093] env[61957]: _type = "Task" [ 1022.909093] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.918652] env[61957]: DEBUG oslo_vmware.api [None req-6a1be883-2588-47a0-91f4-274d9905657e tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': task-1277991, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.984944] env[61957]: DEBUG nova.compute.manager [req-53c85fe3-4658-4832-ad00-e84bbe19622b req-c6348b5d-2c17-4bbf-92fe-19fc6ea0e7ff service nova] [instance: 526ff179-62a6-4763-ab25-797617c4ed57] Received event network-vif-deleted-25abeea6-dd92-49e0-b014-d35dd1c0f8ee {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1023.164661] env[61957]: DEBUG oslo_vmware.api [None req-987070d8-27b8-45f7-833c-32d3bb141bbb tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1277990, 'name': PowerOffVM_Task, 'duration_secs': 0.240631} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.164938] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-987070d8-27b8-45f7-833c-32d3bb141bbb tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 927af4c3-c7e8-404c-839e-d25af5e384a9] Powered off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1023.165139] env[61957]: DEBUG nova.compute.manager [None req-987070d8-27b8-45f7-833c-32d3bb141bbb tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 927af4c3-c7e8-404c-839e-d25af5e384a9] Checking state {{(pid=61957) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1023.165879] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7221f47-e0cb-4fbc-b923-ae8a0d62d513 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.282032] env[61957]: DEBUG oslo_vmware.api [None req-c6a91fc7-15fa-404b-a88f-c93ede8d0713 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': task-1277985, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.331052] env[61957]: INFO nova.compute.claims [None req-3ec44fbf-1f22-4ad1-8b26-d0253bbe3176 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] [instance: 9b0053fe-420e-4b92-86a8-ed44f6a6ec49] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1023.391566] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-20123269-72b2-4ce0-b74a-dd98ef589cf3 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] [instance: 964c8c9d-0721-4ad3-a146-fce1fd1d6513] Volume attach. Driver type: vmdk {{(pid=61957) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1023.391815] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-20123269-72b2-4ce0-b74a-dd98ef589cf3 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] [instance: 964c8c9d-0721-4ad3-a146-fce1fd1d6513] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-274596', 'volume_id': '33adb5ac-f7e3-4840-a689-022217fc5da8', 'name': 'volume-33adb5ac-f7e3-4840-a689-022217fc5da8', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '964c8c9d-0721-4ad3-a146-fce1fd1d6513', 'attached_at': '', 'detached_at': '', 'volume_id': '33adb5ac-f7e3-4840-a689-022217fc5da8', 'serial': '33adb5ac-f7e3-4840-a689-022217fc5da8'} {{(pid=61957) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1023.392740] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc0c6fe0-1c7f-464a-ae7e-6e6de359b266 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.409530] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-385b32f9-90a4-4fbb-a6b4-3f55d9bff94e {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.419628] env[61957]: DEBUG oslo_vmware.api [None req-6a1be883-2588-47a0-91f4-274d9905657e tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': task-1277991, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.388033} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.432021] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-6a1be883-2588-47a0-91f4-274d9905657e tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 3f144012-c0cc-43ce-8c73-4422fe8c1fe6] Extended root virtual disk {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1023.439816] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-20123269-72b2-4ce0-b74a-dd98ef589cf3 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] [instance: 964c8c9d-0721-4ad3-a146-fce1fd1d6513] Reconfiguring VM instance instance-00000058 to attach disk [datastore1] volume-33adb5ac-f7e3-4840-a689-022217fc5da8/volume-33adb5ac-f7e3-4840-a689-022217fc5da8.vmdk or device None with type thin {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1023.440533] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9c16d2f-28e5-4703-a381-4bdfa966500e {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.442928] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-472154b5-06e0-4906-ae43-e915873693c0 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.475937] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-6a1be883-2588-47a0-91f4-274d9905657e tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 3f144012-c0cc-43ce-8c73-4422fe8c1fe6] Reconfiguring VM instance instance-00000061 to attach disk [datastore1] 3f144012-c0cc-43ce-8c73-4422fe8c1fe6/3f144012-c0cc-43ce-8c73-4422fe8c1fe6.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1023.477388] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-73b98591-7823-43ce-a479-60729f6c495c {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.493576] env[61957]: DEBUG oslo_vmware.api [None req-20123269-72b2-4ce0-b74a-dd98ef589cf3 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Waiting for the task: (returnval){ [ 1023.493576] env[61957]: value = "task-1277992" [ 1023.493576] env[61957]: _type = "Task" [ 1023.493576] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.500123] env[61957]: DEBUG oslo_vmware.api [None req-6a1be883-2588-47a0-91f4-274d9905657e tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Waiting for the task: (returnval){ [ 1023.500123] env[61957]: value = "task-1277993" [ 1023.500123] env[61957]: _type = "Task" [ 1023.500123] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.503364] env[61957]: DEBUG oslo_vmware.api [None req-20123269-72b2-4ce0-b74a-dd98ef589cf3 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Task: {'id': task-1277992, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.511769] env[61957]: DEBUG oslo_vmware.api [None req-6a1be883-2588-47a0-91f4-274d9905657e tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': task-1277993, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.677525] env[61957]: DEBUG oslo_concurrency.lockutils [None req-987070d8-27b8-45f7-833c-32d3bb141bbb tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Lock "927af4c3-c7e8-404c-839e-d25af5e384a9" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.548s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1023.783806] env[61957]: DEBUG oslo_vmware.api [None req-c6a91fc7-15fa-404b-a88f-c93ede8d0713 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': task-1277985, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.005676] env[61957]: DEBUG oslo_vmware.api [None req-20123269-72b2-4ce0-b74a-dd98ef589cf3 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Task: {'id': task-1277992, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.016386] env[61957]: DEBUG oslo_vmware.api [None req-6a1be883-2588-47a0-91f4-274d9905657e tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': task-1277993, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.283592] env[61957]: DEBUG oslo_vmware.api [None req-c6a91fc7-15fa-404b-a88f-c93ede8d0713 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': task-1277985, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.480942] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7b0536c-add8-4a42-8d0d-0d84b9ad35b4 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.490914] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3cf70c3e-c964-4381-8afe-41e25e406094 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.525112] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bc334fd-cb71-4c7e-9a51-cf6c05aa77d9 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.530204] env[61957]: DEBUG oslo_vmware.api [None req-20123269-72b2-4ce0-b74a-dd98ef589cf3 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Task: {'id': task-1277992, 'name': ReconfigVM_Task, 'duration_secs': 0.718929} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.531063] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-20123269-72b2-4ce0-b74a-dd98ef589cf3 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] [instance: 964c8c9d-0721-4ad3-a146-fce1fd1d6513] Reconfigured VM instance instance-00000058 to attach disk [datastore1] volume-33adb5ac-f7e3-4840-a689-022217fc5da8/volume-33adb5ac-f7e3-4840-a689-022217fc5da8.vmdk or device None with type thin {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1024.539851] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f9bce809-af3f-4d8f-aa20-22622e7b8d8e {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.549458] env[61957]: DEBUG oslo_vmware.api [None req-6a1be883-2588-47a0-91f4-274d9905657e tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': task-1277993, 'name': ReconfigVM_Task, 'duration_secs': 0.627914} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.550644] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba9d1485-87b2-4d01-9997-9008ebc90305 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.554405] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-6a1be883-2588-47a0-91f4-274d9905657e tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 3f144012-c0cc-43ce-8c73-4422fe8c1fe6] Reconfigured VM instance instance-00000061 to attach disk [datastore1] 3f144012-c0cc-43ce-8c73-4422fe8c1fe6/3f144012-c0cc-43ce-8c73-4422fe8c1fe6.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1024.555335] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c027eccd-faa0-42c6-a7cf-65f4dfab1047 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.567213] env[61957]: DEBUG nova.compute.provider_tree [None req-3ec44fbf-1f22-4ad1-8b26-d0253bbe3176 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1024.570619] env[61957]: DEBUG oslo_vmware.api [None req-6a1be883-2588-47a0-91f4-274d9905657e tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Waiting for the task: (returnval){ [ 1024.570619] env[61957]: value = "task-1277995" [ 1024.570619] env[61957]: _type = "Task" [ 1024.570619] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.570619] env[61957]: DEBUG oslo_vmware.api [None req-20123269-72b2-4ce0-b74a-dd98ef589cf3 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Waiting for the task: (returnval){ [ 1024.570619] env[61957]: value = "task-1277994" [ 1024.570619] env[61957]: _type = "Task" [ 1024.570619] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.582920] env[61957]: DEBUG oslo_vmware.api [None req-6a1be883-2588-47a0-91f4-274d9905657e tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': task-1277995, 'name': Rename_Task} progress is 10%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.586011] env[61957]: DEBUG oslo_vmware.api [None req-20123269-72b2-4ce0-b74a-dd98ef589cf3 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Task: {'id': task-1277994, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.674566] env[61957]: DEBUG oslo_concurrency.lockutils [None req-49c7a8aa-e585-4bf6-ae5e-46c63c7fde84 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Acquiring lock "6509707e-afc6-4c4a-88f9-a471865d510c" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1024.674789] env[61957]: DEBUG oslo_concurrency.lockutils [None req-49c7a8aa-e585-4bf6-ae5e-46c63c7fde84 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Lock "6509707e-afc6-4c4a-88f9-a471865d510c" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1024.675054] env[61957]: DEBUG nova.compute.manager [None req-49c7a8aa-e585-4bf6-ae5e-46c63c7fde84 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: 6509707e-afc6-4c4a-88f9-a471865d510c] Going to confirm migration 2 {{(pid=61957) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 1024.702743] env[61957]: DEBUG oslo_concurrency.lockutils [None req-9f803f6c-daf3-4957-bc49-ff16e02c7831 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Acquiring lock "927af4c3-c7e8-404c-839e-d25af5e384a9" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1024.702743] env[61957]: DEBUG oslo_concurrency.lockutils [None req-9f803f6c-daf3-4957-bc49-ff16e02c7831 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Lock "927af4c3-c7e8-404c-839e-d25af5e384a9" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1024.702955] env[61957]: DEBUG oslo_concurrency.lockutils [None req-9f803f6c-daf3-4957-bc49-ff16e02c7831 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Acquiring lock "927af4c3-c7e8-404c-839e-d25af5e384a9-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1024.703326] env[61957]: DEBUG oslo_concurrency.lockutils [None req-9f803f6c-daf3-4957-bc49-ff16e02c7831 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Lock "927af4c3-c7e8-404c-839e-d25af5e384a9-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1024.703428] env[61957]: DEBUG oslo_concurrency.lockutils [None req-9f803f6c-daf3-4957-bc49-ff16e02c7831 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Lock "927af4c3-c7e8-404c-839e-d25af5e384a9-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1024.708254] env[61957]: INFO nova.compute.manager [None req-9f803f6c-daf3-4957-bc49-ff16e02c7831 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 927af4c3-c7e8-404c-839e-d25af5e384a9] Terminating instance [ 1024.710272] env[61957]: DEBUG nova.compute.manager [None req-9f803f6c-daf3-4957-bc49-ff16e02c7831 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 927af4c3-c7e8-404c-839e-d25af5e384a9] Start destroying the instance on the hypervisor. {{(pid=61957) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1024.710474] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-9f803f6c-daf3-4957-bc49-ff16e02c7831 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 927af4c3-c7e8-404c-839e-d25af5e384a9] Destroying instance {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1024.711338] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09920784-b1ab-40fa-a8e8-ed987014f84f {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.720610] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-9f803f6c-daf3-4957-bc49-ff16e02c7831 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 927af4c3-c7e8-404c-839e-d25af5e384a9] Unregistering the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1024.720869] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c46f2407-527b-4307-b65f-2927c504bbb9 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.787639] env[61957]: DEBUG oslo_vmware.api [None req-c6a91fc7-15fa-404b-a88f-c93ede8d0713 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': task-1277985, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.791300] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-9f803f6c-daf3-4957-bc49-ff16e02c7831 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 927af4c3-c7e8-404c-839e-d25af5e384a9] Unregistered the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1024.791512] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-9f803f6c-daf3-4957-bc49-ff16e02c7831 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 927af4c3-c7e8-404c-839e-d25af5e384a9] Deleting contents of the VM from datastore datastore1 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1024.791694] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-9f803f6c-daf3-4957-bc49-ff16e02c7831 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Deleting the datastore file [datastore1] 927af4c3-c7e8-404c-839e-d25af5e384a9 {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1024.791956] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d2ff5d62-a02d-44a3-b8a8-602ab6359795 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.798671] env[61957]: DEBUG oslo_vmware.api [None req-9f803f6c-daf3-4957-bc49-ff16e02c7831 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Waiting for the task: (returnval){ [ 1024.798671] env[61957]: value = "task-1277997" [ 1024.798671] env[61957]: _type = "Task" [ 1024.798671] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.807928] env[61957]: DEBUG oslo_vmware.api [None req-9f803f6c-daf3-4957-bc49-ff16e02c7831 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1277997, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.072843] env[61957]: DEBUG nova.scheduler.client.report [None req-3ec44fbf-1f22-4ad1-8b26-d0253bbe3176 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1025.087505] env[61957]: DEBUG oslo_vmware.api [None req-20123269-72b2-4ce0-b74a-dd98ef589cf3 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Task: {'id': task-1277994, 'name': ReconfigVM_Task, 'duration_secs': 0.159734} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.091035] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-20123269-72b2-4ce0-b74a-dd98ef589cf3 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] [instance: 964c8c9d-0721-4ad3-a146-fce1fd1d6513] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-274596', 'volume_id': '33adb5ac-f7e3-4840-a689-022217fc5da8', 'name': 'volume-33adb5ac-f7e3-4840-a689-022217fc5da8', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '964c8c9d-0721-4ad3-a146-fce1fd1d6513', 'attached_at': '', 'detached_at': '', 'volume_id': '33adb5ac-f7e3-4840-a689-022217fc5da8', 'serial': '33adb5ac-f7e3-4840-a689-022217fc5da8'} {{(pid=61957) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1025.092678] env[61957]: DEBUG oslo_vmware.api [None req-6a1be883-2588-47a0-91f4-274d9905657e tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': task-1277995, 'name': Rename_Task, 'duration_secs': 0.140746} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.093276] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-6a1be883-2588-47a0-91f4-274d9905657e tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 3f144012-c0cc-43ce-8c73-4422fe8c1fe6] Powering on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1025.093616] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-cc911b13-2b33-4c44-9399-d94b35ba6ba9 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.100611] env[61957]: DEBUG oslo_vmware.api [None req-6a1be883-2588-47a0-91f4-274d9905657e tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Waiting for the task: (returnval){ [ 1025.100611] env[61957]: value = "task-1277998" [ 1025.100611] env[61957]: _type = "Task" [ 1025.100611] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.109714] env[61957]: DEBUG oslo_vmware.api [None req-6a1be883-2588-47a0-91f4-274d9905657e tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': task-1277998, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.248890] env[61957]: DEBUG oslo_concurrency.lockutils [None req-49c7a8aa-e585-4bf6-ae5e-46c63c7fde84 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Acquiring lock "refresh_cache-6509707e-afc6-4c4a-88f9-a471865d510c" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1025.249104] env[61957]: DEBUG oslo_concurrency.lockutils [None req-49c7a8aa-e585-4bf6-ae5e-46c63c7fde84 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Acquired lock "refresh_cache-6509707e-afc6-4c4a-88f9-a471865d510c" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1025.249338] env[61957]: DEBUG nova.network.neutron [None req-49c7a8aa-e585-4bf6-ae5e-46c63c7fde84 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: 6509707e-afc6-4c4a-88f9-a471865d510c] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1025.249542] env[61957]: DEBUG nova.objects.instance [None req-49c7a8aa-e585-4bf6-ae5e-46c63c7fde84 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Lazy-loading 'info_cache' on Instance uuid 6509707e-afc6-4c4a-88f9-a471865d510c {{(pid=61957) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1025.285100] env[61957]: DEBUG oslo_vmware.api [None req-c6a91fc7-15fa-404b-a88f-c93ede8d0713 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': task-1277985, 'name': ReconfigVM_Task} progress is 18%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.309703] env[61957]: DEBUG oslo_vmware.api [None req-9f803f6c-daf3-4957-bc49-ff16e02c7831 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1277997, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.136147} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.309978] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-9f803f6c-daf3-4957-bc49-ff16e02c7831 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Deleted the datastore file {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1025.310239] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-9f803f6c-daf3-4957-bc49-ff16e02c7831 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 927af4c3-c7e8-404c-839e-d25af5e384a9] Deleted contents of the VM from datastore datastore1 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1025.310437] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-9f803f6c-daf3-4957-bc49-ff16e02c7831 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 927af4c3-c7e8-404c-839e-d25af5e384a9] Instance destroyed {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1025.310613] env[61957]: INFO nova.compute.manager [None req-9f803f6c-daf3-4957-bc49-ff16e02c7831 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 927af4c3-c7e8-404c-839e-d25af5e384a9] Took 0.60 seconds to destroy the instance on the hypervisor. [ 1025.310862] env[61957]: DEBUG oslo.service.loopingcall [None req-9f803f6c-daf3-4957-bc49-ff16e02c7831 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1025.311071] env[61957]: DEBUG nova.compute.manager [-] [instance: 927af4c3-c7e8-404c-839e-d25af5e384a9] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1025.311169] env[61957]: DEBUG nova.network.neutron [-] [instance: 927af4c3-c7e8-404c-839e-d25af5e384a9] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1025.583837] env[61957]: DEBUG oslo_concurrency.lockutils [None req-3ec44fbf-1f22-4ad1-8b26-d0253bbe3176 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.259s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1025.587421] env[61957]: DEBUG oslo_concurrency.lockutils [None req-437e09d9-7c87-4b34-90f6-4026cbd0876c tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.839s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1025.587829] env[61957]: DEBUG nova.objects.instance [None req-437e09d9-7c87-4b34-90f6-4026cbd0876c tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Lazy-loading 'resources' on Instance uuid e2107d5c-8453-4d24-b364-2b5c1d125062 {{(pid=61957) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1025.600626] env[61957]: DEBUG nova.compute.manager [req-33b99a42-3a2d-4e0e-8552-ab2c1f148d49 req-52e95a35-d6dd-4407-9edf-e5e744da5cc2 service nova] [instance: 927af4c3-c7e8-404c-839e-d25af5e384a9] Received event network-vif-deleted-38da1ba9-c5cc-4f76-b516-5ab8fdca9064 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1025.600822] env[61957]: INFO nova.compute.manager [req-33b99a42-3a2d-4e0e-8552-ab2c1f148d49 req-52e95a35-d6dd-4407-9edf-e5e744da5cc2 service nova] [instance: 927af4c3-c7e8-404c-839e-d25af5e384a9] Neutron deleted interface 38da1ba9-c5cc-4f76-b516-5ab8fdca9064; detaching it from the instance and deleting it from the info cache [ 1025.601183] env[61957]: DEBUG nova.network.neutron [req-33b99a42-3a2d-4e0e-8552-ab2c1f148d49 req-52e95a35-d6dd-4407-9edf-e5e744da5cc2 service nova] [instance: 927af4c3-c7e8-404c-839e-d25af5e384a9] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1025.615102] env[61957]: DEBUG oslo_vmware.api [None req-6a1be883-2588-47a0-91f4-274d9905657e tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': task-1277998, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.629618] env[61957]: INFO nova.network.neutron [None req-3ec44fbf-1f22-4ad1-8b26-d0253bbe3176 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] [instance: 9b0053fe-420e-4b92-86a8-ed44f6a6ec49] Updating port 219230f0-0b14-4be9-81bb-67a765d8ced9 with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1025.784419] env[61957]: DEBUG oslo_vmware.api [None req-c6a91fc7-15fa-404b-a88f-c93ede8d0713 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': task-1277985, 'name': ReconfigVM_Task, 'duration_secs': 5.811434} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.784758] env[61957]: DEBUG oslo_concurrency.lockutils [None req-c6a91fc7-15fa-404b-a88f-c93ede8d0713 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Releasing lock "08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1025.785047] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-c6a91fc7-15fa-404b-a88f-c93ede8d0713 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2] Reconfigured VM to detach interface {{(pid=61957) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1984}} [ 1026.077031] env[61957]: DEBUG nova.network.neutron [-] [instance: 927af4c3-c7e8-404c-839e-d25af5e384a9] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1026.103821] env[61957]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3c8498b0-e413-4366-a40b-d906c79b8fb0 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.115562] env[61957]: DEBUG oslo_vmware.api [None req-6a1be883-2588-47a0-91f4-274d9905657e tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': task-1277998, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.123024] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01aef807-e67f-45c3-816d-1045ca2877d4 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.137439] env[61957]: DEBUG nova.objects.instance [None req-20123269-72b2-4ce0-b74a-dd98ef589cf3 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Lazy-loading 'flavor' on Instance uuid 964c8c9d-0721-4ad3-a146-fce1fd1d6513 {{(pid=61957) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1026.155332] env[61957]: DEBUG nova.compute.manager [req-33b99a42-3a2d-4e0e-8552-ab2c1f148d49 req-52e95a35-d6dd-4407-9edf-e5e744da5cc2 service nova] [instance: 927af4c3-c7e8-404c-839e-d25af5e384a9] Detach interface failed, port_id=38da1ba9-c5cc-4f76-b516-5ab8fdca9064, reason: Instance 927af4c3-c7e8-404c-839e-d25af5e384a9 could not be found. {{(pid=61957) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1026.268498] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-303fabd5-0c02-44b3-b18f-6f74c9578bb6 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.275949] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5840f59-87ff-44df-9666-3d64ead92605 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.307468] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6079c04-b815-457f-8cf7-ee1d15cc5aa9 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.315011] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da18ef14-b426-4eae-b386-4d219caff6c6 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.329741] env[61957]: DEBUG nova.compute.provider_tree [None req-437e09d9-7c87-4b34-90f6-4026cbd0876c tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1026.484106] env[61957]: DEBUG nova.network.neutron [None req-49c7a8aa-e585-4bf6-ae5e-46c63c7fde84 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: 6509707e-afc6-4c4a-88f9-a471865d510c] Updating instance_info_cache with network_info: [{"id": "7df2736b-10d0-4f9d-83aa-e2bbdf1ef45b", "address": "fa:16:3e:2e:da:a9", "network": {"id": "6fbccc17-6674-4cd5-8f1b-aa73570baac7", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-394544811-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1dbdc699741a48af93c52068d88a5357", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "20e3f794-c7a3-4696-9488-ecf34c570ef9", "external-id": "nsx-vlan-transportzone-509", "segmentation_id": 509, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7df2736b-10", "ovs_interfaceid": "7df2736b-10d0-4f9d-83aa-e2bbdf1ef45b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1026.581181] env[61957]: INFO nova.compute.manager [-] [instance: 927af4c3-c7e8-404c-839e-d25af5e384a9] Took 1.27 seconds to deallocate network for instance. [ 1026.617419] env[61957]: DEBUG oslo_vmware.api [None req-6a1be883-2588-47a0-91f4-274d9905657e tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': task-1277998, 'name': PowerOnVM_Task, 'duration_secs': 1.017781} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.618450] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-6a1be883-2588-47a0-91f4-274d9905657e tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 3f144012-c0cc-43ce-8c73-4422fe8c1fe6] Powered on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1026.618699] env[61957]: INFO nova.compute.manager [None req-6a1be883-2588-47a0-91f4-274d9905657e tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 3f144012-c0cc-43ce-8c73-4422fe8c1fe6] Took 9.94 seconds to spawn the instance on the hypervisor. [ 1026.618941] env[61957]: DEBUG nova.compute.manager [None req-6a1be883-2588-47a0-91f4-274d9905657e tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 3f144012-c0cc-43ce-8c73-4422fe8c1fe6] Checking state {{(pid=61957) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1026.619727] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-399f6c31-fc05-46a3-8576-e522a68b2adf {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.643358] env[61957]: DEBUG oslo_concurrency.lockutils [None req-20123269-72b2-4ce0-b74a-dd98ef589cf3 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Lock "964c8c9d-0721-4ad3-a146-fce1fd1d6513" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.885s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1026.832191] env[61957]: DEBUG nova.scheduler.client.report [None req-437e09d9-7c87-4b34-90f6-4026cbd0876c tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1026.987280] env[61957]: DEBUG oslo_concurrency.lockutils [None req-49c7a8aa-e585-4bf6-ae5e-46c63c7fde84 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Releasing lock "refresh_cache-6509707e-afc6-4c4a-88f9-a471865d510c" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1026.987555] env[61957]: DEBUG nova.objects.instance [None req-49c7a8aa-e585-4bf6-ae5e-46c63c7fde84 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Lazy-loading 'migration_context' on Instance uuid 6509707e-afc6-4c4a-88f9-a471865d510c {{(pid=61957) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1027.065170] env[61957]: DEBUG oslo_concurrency.lockutils [None req-c6a91fc7-15fa-404b-a88f-c93ede8d0713 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Acquiring lock "refresh_cache-08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1027.065534] env[61957]: DEBUG oslo_concurrency.lockutils [None req-c6a91fc7-15fa-404b-a88f-c93ede8d0713 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Acquired lock "refresh_cache-08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1027.065821] env[61957]: DEBUG nova.network.neutron [None req-c6a91fc7-15fa-404b-a88f-c93ede8d0713 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1027.087525] env[61957]: DEBUG oslo_concurrency.lockutils [None req-9f803f6c-daf3-4957-bc49-ff16e02c7831 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1027.092657] env[61957]: DEBUG nova.compute.manager [req-d0775d92-7d76-40e7-9688-796e165eac56 req-333d4f0f-cd6b-419c-acdd-56111ba4d89f service nova] [instance: 08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2] Received event network-vif-deleted-00c8220d-5633-4de7-9795-8a6a8543c4ec {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1027.092657] env[61957]: INFO nova.compute.manager [req-d0775d92-7d76-40e7-9688-796e165eac56 req-333d4f0f-cd6b-419c-acdd-56111ba4d89f service nova] [instance: 08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2] Neutron deleted interface 00c8220d-5633-4de7-9795-8a6a8543c4ec; detaching it from the instance and deleting it from the info cache [ 1027.092657] env[61957]: DEBUG nova.network.neutron [req-d0775d92-7d76-40e7-9688-796e165eac56 req-333d4f0f-cd6b-419c-acdd-56111ba4d89f service nova] [instance: 08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2] Updating instance_info_cache with network_info: [{"id": "f62a9f12-0b55-4138-a40e-b431855487fe", "address": "fa:16:3e:af:48:d3", "network": {"id": "364cabcb-046d-4f91-b1ee-5dd3adfae6d7", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1432396967-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.219", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8713aa35bcb24b86ad0b58ca9fc991ba", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b7a73c01-1bb9-4612-a1a7-16d71b732e81", "external-id": "nsx-vlan-transportzone-711", "segmentation_id": 711, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf62a9f12-0b", "ovs_interfaceid": "f62a9f12-0b55-4138-a40e-b431855487fe", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "6a70445e-1266-453e-a6ff-50e8cbdcfd08", "address": "fa:16:3e:bb:2f:b6", "network": {"id": "364cabcb-046d-4f91-b1ee-5dd3adfae6d7", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1432396967-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8713aa35bcb24b86ad0b58ca9fc991ba", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b7a73c01-1bb9-4612-a1a7-16d71b732e81", "external-id": "nsx-vlan-transportzone-711", "segmentation_id": 711, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6a70445e-12", "ovs_interfaceid": "6a70445e-1266-453e-a6ff-50e8cbdcfd08", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1027.139668] env[61957]: INFO nova.compute.manager [None req-6a1be883-2588-47a0-91f4-274d9905657e tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 3f144012-c0cc-43ce-8c73-4422fe8c1fe6] Took 14.81 seconds to build instance. [ 1027.223097] env[61957]: DEBUG oslo_concurrency.lockutils [None req-3ec44fbf-1f22-4ad1-8b26-d0253bbe3176 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Acquiring lock "refresh_cache-9b0053fe-420e-4b92-86a8-ed44f6a6ec49" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1027.223097] env[61957]: DEBUG oslo_concurrency.lockutils [None req-3ec44fbf-1f22-4ad1-8b26-d0253bbe3176 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Acquired lock "refresh_cache-9b0053fe-420e-4b92-86a8-ed44f6a6ec49" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1027.223097] env[61957]: DEBUG nova.network.neutron [None req-3ec44fbf-1f22-4ad1-8b26-d0253bbe3176 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] [instance: 9b0053fe-420e-4b92-86a8-ed44f6a6ec49] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1027.337222] env[61957]: DEBUG oslo_concurrency.lockutils [None req-9a7c3783-e05c-490e-9f05-e499ff43e1b0 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Acquiring lock "964c8c9d-0721-4ad3-a146-fce1fd1d6513" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1027.337578] env[61957]: DEBUG oslo_concurrency.lockutils [None req-9a7c3783-e05c-490e-9f05-e499ff43e1b0 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Lock "964c8c9d-0721-4ad3-a146-fce1fd1d6513" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1027.339677] env[61957]: DEBUG oslo_concurrency.lockutils [None req-437e09d9-7c87-4b34-90f6-4026cbd0876c tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.753s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1027.343405] env[61957]: DEBUG oslo_concurrency.lockutils [None req-4ca8b060-369e-429d-bf7a-53fb2d00bd30 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.587s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1027.343405] env[61957]: DEBUG nova.objects.instance [None req-4ca8b060-369e-429d-bf7a-53fb2d00bd30 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Lazy-loading 'resources' on Instance uuid 526ff179-62a6-4763-ab25-797617c4ed57 {{(pid=61957) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1027.374546] env[61957]: INFO nova.scheduler.client.report [None req-437e09d9-7c87-4b34-90f6-4026cbd0876c tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Deleted allocations for instance e2107d5c-8453-4d24-b364-2b5c1d125062 [ 1027.490438] env[61957]: DEBUG nova.objects.base [None req-49c7a8aa-e585-4bf6-ae5e-46c63c7fde84 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Object Instance<6509707e-afc6-4c4a-88f9-a471865d510c> lazy-loaded attributes: info_cache,migration_context {{(pid=61957) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1027.491474] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6df644c-00d9-45e8-9641-8ad20348b3cc {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.510919] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-987cc27f-fcae-4691-b5ff-a56a0e6d089e {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.516693] env[61957]: DEBUG oslo_vmware.api [None req-49c7a8aa-e585-4bf6-ae5e-46c63c7fde84 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Waiting for the task: (returnval){ [ 1027.516693] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52ff74a4-fe3b-a4a4-bab1-a2c9c6e764ea" [ 1027.516693] env[61957]: _type = "Task" [ 1027.516693] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1027.525154] env[61957]: DEBUG oslo_vmware.api [None req-49c7a8aa-e585-4bf6-ae5e-46c63c7fde84 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52ff74a4-fe3b-a4a4-bab1-a2c9c6e764ea, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.567875] env[61957]: DEBUG oslo_concurrency.lockutils [None req-507275cf-57ab-4489-9d8f-aa0676728f56 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Acquiring lock "08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1027.568161] env[61957]: DEBUG oslo_concurrency.lockutils [None req-507275cf-57ab-4489-9d8f-aa0676728f56 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Lock "08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1027.568380] env[61957]: DEBUG oslo_concurrency.lockutils [None req-507275cf-57ab-4489-9d8f-aa0676728f56 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Acquiring lock "08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1027.568590] env[61957]: DEBUG oslo_concurrency.lockutils [None req-507275cf-57ab-4489-9d8f-aa0676728f56 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Lock "08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1027.568739] env[61957]: DEBUG oslo_concurrency.lockutils [None req-507275cf-57ab-4489-9d8f-aa0676728f56 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Lock "08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1027.572438] env[61957]: INFO nova.compute.manager [None req-507275cf-57ab-4489-9d8f-aa0676728f56 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2] Terminating instance [ 1027.574231] env[61957]: DEBUG nova.compute.manager [None req-507275cf-57ab-4489-9d8f-aa0676728f56 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2] Start destroying the instance on the hypervisor. {{(pid=61957) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1027.574432] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-507275cf-57ab-4489-9d8f-aa0676728f56 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2] Destroying instance {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1027.575492] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00f79aa0-5d39-428a-afd7-3dca0f50eb01 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.582824] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-507275cf-57ab-4489-9d8f-aa0676728f56 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2] Powering off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1027.583062] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e9366b3c-0004-4b90-963d-ef84c442c8ab {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.589768] env[61957]: DEBUG oslo_vmware.api [None req-507275cf-57ab-4489-9d8f-aa0676728f56 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Waiting for the task: (returnval){ [ 1027.589768] env[61957]: value = "task-1277999" [ 1027.589768] env[61957]: _type = "Task" [ 1027.589768] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1027.593909] env[61957]: DEBUG oslo_concurrency.lockutils [req-d0775d92-7d76-40e7-9688-796e165eac56 req-333d4f0f-cd6b-419c-acdd-56111ba4d89f service nova] Acquiring lock "08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1027.596956] env[61957]: DEBUG oslo_vmware.api [None req-507275cf-57ab-4489-9d8f-aa0676728f56 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': task-1277999, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.636720] env[61957]: DEBUG nova.compute.manager [req-f87cfb1f-b2a2-4db6-880c-a3fa7f8c0df7 req-da22b3c4-4128-486f-9423-a8115e382e2f service nova] [instance: 08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2] Received event network-vif-deleted-6a70445e-1266-453e-a6ff-50e8cbdcfd08 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1027.636829] env[61957]: INFO nova.compute.manager [req-f87cfb1f-b2a2-4db6-880c-a3fa7f8c0df7 req-da22b3c4-4128-486f-9423-a8115e382e2f service nova] [instance: 08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2] Neutron deleted interface 6a70445e-1266-453e-a6ff-50e8cbdcfd08; detaching it from the instance and deleting it from the info cache [ 1027.637041] env[61957]: DEBUG nova.network.neutron [req-f87cfb1f-b2a2-4db6-880c-a3fa7f8c0df7 req-da22b3c4-4128-486f-9423-a8115e382e2f service nova] [instance: 08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2] Updating instance_info_cache with network_info: [{"id": "f62a9f12-0b55-4138-a40e-b431855487fe", "address": "fa:16:3e:af:48:d3", "network": {"id": "364cabcb-046d-4f91-b1ee-5dd3adfae6d7", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1432396967-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.219", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8713aa35bcb24b86ad0b58ca9fc991ba", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b7a73c01-1bb9-4612-a1a7-16d71b732e81", "external-id": "nsx-vlan-transportzone-711", "segmentation_id": 711, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf62a9f12-0b", "ovs_interfaceid": "f62a9f12-0b55-4138-a40e-b431855487fe", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1027.645416] env[61957]: DEBUG oslo_concurrency.lockutils [None req-6a1be883-2588-47a0-91f4-274d9905657e tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Lock "3f144012-c0cc-43ce-8c73-4422fe8c1fe6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.325s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1027.843891] env[61957]: INFO nova.compute.manager [None req-9a7c3783-e05c-490e-9f05-e499ff43e1b0 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] [instance: 964c8c9d-0721-4ad3-a146-fce1fd1d6513] Detaching volume 33adb5ac-f7e3-4840-a689-022217fc5da8 [ 1027.881915] env[61957]: INFO nova.network.neutron [None req-c6a91fc7-15fa-404b-a88f-c93ede8d0713 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2] Port 6a70445e-1266-453e-a6ff-50e8cbdcfd08 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 1027.882154] env[61957]: INFO nova.network.neutron [None req-c6a91fc7-15fa-404b-a88f-c93ede8d0713 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2] Port 00c8220d-5633-4de7-9795-8a6a8543c4ec from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 1027.882426] env[61957]: DEBUG nova.network.neutron [None req-c6a91fc7-15fa-404b-a88f-c93ede8d0713 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2] Updating instance_info_cache with network_info: [{"id": "f62a9f12-0b55-4138-a40e-b431855487fe", "address": "fa:16:3e:af:48:d3", "network": {"id": "364cabcb-046d-4f91-b1ee-5dd3adfae6d7", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1432396967-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.219", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8713aa35bcb24b86ad0b58ca9fc991ba", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b7a73c01-1bb9-4612-a1a7-16d71b732e81", "external-id": "nsx-vlan-transportzone-711", "segmentation_id": 711, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf62a9f12-0b", "ovs_interfaceid": "f62a9f12-0b55-4138-a40e-b431855487fe", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1027.887122] env[61957]: INFO nova.virt.block_device [None req-9a7c3783-e05c-490e-9f05-e499ff43e1b0 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] [instance: 964c8c9d-0721-4ad3-a146-fce1fd1d6513] Attempting to driver detach volume 33adb5ac-f7e3-4840-a689-022217fc5da8 from mountpoint /dev/sdb [ 1027.887122] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-9a7c3783-e05c-490e-9f05-e499ff43e1b0 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] [instance: 964c8c9d-0721-4ad3-a146-fce1fd1d6513] Volume detach. Driver type: vmdk {{(pid=61957) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1027.887122] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-9a7c3783-e05c-490e-9f05-e499ff43e1b0 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] [instance: 964c8c9d-0721-4ad3-a146-fce1fd1d6513] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-274596', 'volume_id': '33adb5ac-f7e3-4840-a689-022217fc5da8', 'name': 'volume-33adb5ac-f7e3-4840-a689-022217fc5da8', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '964c8c9d-0721-4ad3-a146-fce1fd1d6513', 'attached_at': '', 'detached_at': '', 'volume_id': '33adb5ac-f7e3-4840-a689-022217fc5da8', 'serial': '33adb5ac-f7e3-4840-a689-022217fc5da8'} {{(pid=61957) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1027.888138] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b30032e2-3b2c-4464-99ec-26ba90bf7525 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.892912] env[61957]: DEBUG oslo_concurrency.lockutils [None req-437e09d9-7c87-4b34-90f6-4026cbd0876c tempest-ImagesTestJSON-1774349710 tempest-ImagesTestJSON-1774349710-project-member] Lock "e2107d5c-8453-4d24-b364-2b5c1d125062" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.073s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1027.920141] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98381229-c6e7-4e49-97a6-206758d983f5 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.928449] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd7f3214-8a5c-48fb-a976-e273d00123bc {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.948421] env[61957]: DEBUG oslo_concurrency.lockutils [None req-fa4af5a4-f456-4969-8ee2-40fce4f1cd43 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Acquiring lock "3f144012-c0cc-43ce-8c73-4422fe8c1fe6" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1027.948713] env[61957]: DEBUG oslo_concurrency.lockutils [None req-fa4af5a4-f456-4969-8ee2-40fce4f1cd43 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Lock "3f144012-c0cc-43ce-8c73-4422fe8c1fe6" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1027.948950] env[61957]: DEBUG oslo_concurrency.lockutils [None req-fa4af5a4-f456-4969-8ee2-40fce4f1cd43 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Acquiring lock "3f144012-c0cc-43ce-8c73-4422fe8c1fe6-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1027.949156] env[61957]: DEBUG oslo_concurrency.lockutils [None req-fa4af5a4-f456-4969-8ee2-40fce4f1cd43 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Lock "3f144012-c0cc-43ce-8c73-4422fe8c1fe6-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1027.949337] env[61957]: DEBUG oslo_concurrency.lockutils [None req-fa4af5a4-f456-4969-8ee2-40fce4f1cd43 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Lock "3f144012-c0cc-43ce-8c73-4422fe8c1fe6-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1027.956047] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c384abde-6d9b-49b4-b478-a152533affc5 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.959014] env[61957]: INFO nova.compute.manager [None req-fa4af5a4-f456-4969-8ee2-40fce4f1cd43 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 3f144012-c0cc-43ce-8c73-4422fe8c1fe6] Terminating instance [ 1027.961130] env[61957]: DEBUG nova.compute.manager [None req-fa4af5a4-f456-4969-8ee2-40fce4f1cd43 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 3f144012-c0cc-43ce-8c73-4422fe8c1fe6] Start destroying the instance on the hypervisor. {{(pid=61957) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1027.961283] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-fa4af5a4-f456-4969-8ee2-40fce4f1cd43 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 3f144012-c0cc-43ce-8c73-4422fe8c1fe6] Destroying instance {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1027.962503] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4a3e595-38e6-4b22-a917-4acd59658141 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.975191] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-9a7c3783-e05c-490e-9f05-e499ff43e1b0 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] The volume has not been displaced from its original location: [datastore1] volume-33adb5ac-f7e3-4840-a689-022217fc5da8/volume-33adb5ac-f7e3-4840-a689-022217fc5da8.vmdk. No consolidation needed. {{(pid=61957) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1027.980504] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-9a7c3783-e05c-490e-9f05-e499ff43e1b0 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] [instance: 964c8c9d-0721-4ad3-a146-fce1fd1d6513] Reconfiguring VM instance instance-00000058 to detach disk 2001 {{(pid=61957) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1027.983373] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-614836cc-3684-4f8c-b74a-438d8c171826 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.000421] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-fa4af5a4-f456-4969-8ee2-40fce4f1cd43 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 3f144012-c0cc-43ce-8c73-4422fe8c1fe6] Powering off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1028.000924] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e4c55956-99f9-4431-992f-b98c3264272c {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.006809] env[61957]: DEBUG oslo_vmware.api [None req-9a7c3783-e05c-490e-9f05-e499ff43e1b0 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Waiting for the task: (returnval){ [ 1028.006809] env[61957]: value = "task-1278000" [ 1028.006809] env[61957]: _type = "Task" [ 1028.006809] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.010461] env[61957]: DEBUG oslo_vmware.api [None req-fa4af5a4-f456-4969-8ee2-40fce4f1cd43 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Waiting for the task: (returnval){ [ 1028.010461] env[61957]: value = "task-1278001" [ 1028.010461] env[61957]: _type = "Task" [ 1028.010461] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.028890] env[61957]: DEBUG oslo_vmware.api [None req-fa4af5a4-f456-4969-8ee2-40fce4f1cd43 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': task-1278001, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.029181] env[61957]: DEBUG oslo_vmware.api [None req-9a7c3783-e05c-490e-9f05-e499ff43e1b0 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Task: {'id': task-1278000, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.034641] env[61957]: DEBUG oslo_vmware.api [None req-49c7a8aa-e585-4bf6-ae5e-46c63c7fde84 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52ff74a4-fe3b-a4a4-bab1-a2c9c6e764ea, 'name': SearchDatastore_Task, 'duration_secs': 0.008096} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.035451] env[61957]: DEBUG oslo_concurrency.lockutils [None req-49c7a8aa-e585-4bf6-ae5e-46c63c7fde84 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1028.039076] env[61957]: DEBUG nova.network.neutron [None req-3ec44fbf-1f22-4ad1-8b26-d0253bbe3176 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] [instance: 9b0053fe-420e-4b92-86a8-ed44f6a6ec49] Updating instance_info_cache with network_info: [{"id": "219230f0-0b14-4be9-81bb-67a765d8ced9", "address": "fa:16:3e:f6:9e:ce", "network": {"id": "cf378e9c-eaae-4168-bd60-9818fcfaf405", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1243326371-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f7c20ab3c822423fa837a03f066e4239", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8cb478a6-872c-4a90-a8db-526b374e82ce", "external-id": "nsx-vlan-transportzone-835", "segmentation_id": 835, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap219230f0-0b", "ovs_interfaceid": "219230f0-0b14-4be9-81bb-67a765d8ced9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1028.096074] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05e5f038-22af-4459-b38b-e91678dbff91 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.106090] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04fcd9dc-82bd-4fdc-9c89-8abe2390c2b9 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.109524] env[61957]: DEBUG oslo_vmware.api [None req-507275cf-57ab-4489-9d8f-aa0676728f56 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': task-1277999, 'name': PowerOffVM_Task, 'duration_secs': 0.232033} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.109824] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-507275cf-57ab-4489-9d8f-aa0676728f56 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2] Powered off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1028.109992] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-507275cf-57ab-4489-9d8f-aa0676728f56 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2] Unregistering the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1028.110577] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7bffb56c-3531-4cf2-b254-59fa3141f64d {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.139333] env[61957]: DEBUG oslo_concurrency.lockutils [req-f87cfb1f-b2a2-4db6-880c-a3fa7f8c0df7 req-da22b3c4-4128-486f-9423-a8115e382e2f service nova] Acquiring lock "08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1028.140139] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7c558fb-bb8b-4fbb-83b5-869c33f6c24b {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.148054] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c663f01-8a45-4a46-b100-597b23136541 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.163574] env[61957]: DEBUG nova.compute.provider_tree [None req-4ca8b060-369e-429d-bf7a-53fb2d00bd30 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1028.385015] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-507275cf-57ab-4489-9d8f-aa0676728f56 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2] Unregistered the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1028.385412] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-507275cf-57ab-4489-9d8f-aa0676728f56 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2] Deleting contents of the VM from datastore datastore2 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1028.385856] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-507275cf-57ab-4489-9d8f-aa0676728f56 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Deleting the datastore file [datastore2] 08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2 {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1028.386532] env[61957]: DEBUG oslo_concurrency.lockutils [None req-c6a91fc7-15fa-404b-a88f-c93ede8d0713 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Releasing lock "refresh_cache-08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1028.389943] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5cedc896-7b56-48fe-af7b-e0d21e25b147 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.398214] env[61957]: DEBUG oslo_vmware.api [None req-507275cf-57ab-4489-9d8f-aa0676728f56 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Waiting for the task: (returnval){ [ 1028.398214] env[61957]: value = "task-1278003" [ 1028.398214] env[61957]: _type = "Task" [ 1028.398214] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.408143] env[61957]: DEBUG oslo_vmware.api [None req-507275cf-57ab-4489-9d8f-aa0676728f56 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': task-1278003, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.522577] env[61957]: DEBUG oslo_vmware.api [None req-9a7c3783-e05c-490e-9f05-e499ff43e1b0 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Task: {'id': task-1278000, 'name': ReconfigVM_Task, 'duration_secs': 0.243321} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.525579] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-9a7c3783-e05c-490e-9f05-e499ff43e1b0 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] [instance: 964c8c9d-0721-4ad3-a146-fce1fd1d6513] Reconfigured VM instance instance-00000058 to detach disk 2001 {{(pid=61957) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1028.530244] env[61957]: DEBUG oslo_vmware.api [None req-fa4af5a4-f456-4969-8ee2-40fce4f1cd43 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': task-1278001, 'name': PowerOffVM_Task, 'duration_secs': 0.159314} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.530487] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a32ef689-fda3-48c9-900f-01cb0cc8153a {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.540966] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-fa4af5a4-f456-4969-8ee2-40fce4f1cd43 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 3f144012-c0cc-43ce-8c73-4422fe8c1fe6] Powered off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1028.541159] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-fa4af5a4-f456-4969-8ee2-40fce4f1cd43 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 3f144012-c0cc-43ce-8c73-4422fe8c1fe6] Unregistering the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1028.541764] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f49e415c-4edf-48ab-be9c-a4ab028b8805 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.544079] env[61957]: DEBUG oslo_concurrency.lockutils [None req-3ec44fbf-1f22-4ad1-8b26-d0253bbe3176 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Releasing lock "refresh_cache-9b0053fe-420e-4b92-86a8-ed44f6a6ec49" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1028.550975] env[61957]: DEBUG oslo_vmware.api [None req-9a7c3783-e05c-490e-9f05-e499ff43e1b0 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Waiting for the task: (returnval){ [ 1028.550975] env[61957]: value = "task-1278004" [ 1028.550975] env[61957]: _type = "Task" [ 1028.550975] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.558594] env[61957]: DEBUG oslo_vmware.api [None req-9a7c3783-e05c-490e-9f05-e499ff43e1b0 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Task: {'id': task-1278004, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.573179] env[61957]: DEBUG nova.virt.hardware [None req-3ec44fbf-1f22-4ad1-8b26-d0253bbe3176 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T17:22:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='ff774270126cd60f831863bcf4396ca5',container_format='bare',created_at=2024-10-12T17:31:03Z,direct_url=,disk_format='vmdk',id=f04c1641-6ff1-4b2c-9748-9cf176307db0,min_disk=1,min_ram=0,name='tempest-ServersNegativeTestJSON-server-1024982246-shelved',owner='f7c20ab3c822423fa837a03f066e4239',properties=ImageMetaProps,protected=,size=31667200,status='active',tags=,updated_at=2024-10-12T17:31:18Z,virtual_size=,visibility=), allow threads: False {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1028.573394] env[61957]: DEBUG nova.virt.hardware [None req-3ec44fbf-1f22-4ad1-8b26-d0253bbe3176 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Flavor limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1028.573557] env[61957]: DEBUG nova.virt.hardware [None req-3ec44fbf-1f22-4ad1-8b26-d0253bbe3176 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Image limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1028.573748] env[61957]: DEBUG nova.virt.hardware [None req-3ec44fbf-1f22-4ad1-8b26-d0253bbe3176 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Flavor pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1028.573900] env[61957]: DEBUG nova.virt.hardware [None req-3ec44fbf-1f22-4ad1-8b26-d0253bbe3176 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Image pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1028.574065] env[61957]: DEBUG nova.virt.hardware [None req-3ec44fbf-1f22-4ad1-8b26-d0253bbe3176 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1028.574286] env[61957]: DEBUG nova.virt.hardware [None req-3ec44fbf-1f22-4ad1-8b26-d0253bbe3176 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1028.574464] env[61957]: DEBUG nova.virt.hardware [None req-3ec44fbf-1f22-4ad1-8b26-d0253bbe3176 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1028.574633] env[61957]: DEBUG nova.virt.hardware [None req-3ec44fbf-1f22-4ad1-8b26-d0253bbe3176 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Got 1 possible topologies {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1028.574785] env[61957]: DEBUG nova.virt.hardware [None req-3ec44fbf-1f22-4ad1-8b26-d0253bbe3176 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1028.574956] env[61957]: DEBUG nova.virt.hardware [None req-3ec44fbf-1f22-4ad1-8b26-d0253bbe3176 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1028.575837] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38857420-08e7-4968-8aad-2c7dc1dafece {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.583315] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d813efb-1d36-4aed-9c78-0beeb78e9eb5 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.597060] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-3ec44fbf-1f22-4ad1-8b26-d0253bbe3176 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] [instance: 9b0053fe-420e-4b92-86a8-ed44f6a6ec49] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f6:9e:ce', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8cb478a6-872c-4a90-a8db-526b374e82ce', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '219230f0-0b14-4be9-81bb-67a765d8ced9', 'vif_model': 'vmxnet3'}] {{(pid=61957) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1028.604515] env[61957]: DEBUG oslo.service.loopingcall [None req-3ec44fbf-1f22-4ad1-8b26-d0253bbe3176 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1028.604877] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9b0053fe-420e-4b92-86a8-ed44f6a6ec49] Creating VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1028.605181] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3d563351-0550-409d-b745-7bd3e313eed0 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.620604] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-fa4af5a4-f456-4969-8ee2-40fce4f1cd43 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 3f144012-c0cc-43ce-8c73-4422fe8c1fe6] Unregistered the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1028.620816] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-fa4af5a4-f456-4969-8ee2-40fce4f1cd43 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 3f144012-c0cc-43ce-8c73-4422fe8c1fe6] Deleting contents of the VM from datastore datastore1 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1028.621041] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-fa4af5a4-f456-4969-8ee2-40fce4f1cd43 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Deleting the datastore file [datastore1] 3f144012-c0cc-43ce-8c73-4422fe8c1fe6 {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1028.621676] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f1aca00a-0f96-40eb-bfb6-e862a2ff8fdb {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.625834] env[61957]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1028.625834] env[61957]: value = "task-1278006" [ 1028.625834] env[61957]: _type = "Task" [ 1028.625834] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.633876] env[61957]: DEBUG oslo_vmware.api [None req-fa4af5a4-f456-4969-8ee2-40fce4f1cd43 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Waiting for the task: (returnval){ [ 1028.633876] env[61957]: value = "task-1278007" [ 1028.633876] env[61957]: _type = "Task" [ 1028.633876] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.636836] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1278006, 'name': CreateVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.644438] env[61957]: DEBUG oslo_vmware.api [None req-fa4af5a4-f456-4969-8ee2-40fce4f1cd43 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': task-1278007, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.668674] env[61957]: DEBUG nova.scheduler.client.report [None req-4ca8b060-369e-429d-bf7a-53fb2d00bd30 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1028.893838] env[61957]: DEBUG oslo_concurrency.lockutils [None req-c6a91fc7-15fa-404b-a88f-c93ede8d0713 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Lock "interface-08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2-6a70445e-1266-453e-a6ff-50e8cbdcfd08" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 9.577s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1028.910169] env[61957]: DEBUG oslo_vmware.api [None req-507275cf-57ab-4489-9d8f-aa0676728f56 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': task-1278003, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.165152} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.910596] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-507275cf-57ab-4489-9d8f-aa0676728f56 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Deleted the datastore file {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1028.910702] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-507275cf-57ab-4489-9d8f-aa0676728f56 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2] Deleted contents of the VM from datastore datastore2 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1028.911124] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-507275cf-57ab-4489-9d8f-aa0676728f56 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2] Instance destroyed {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1028.911255] env[61957]: INFO nova.compute.manager [None req-507275cf-57ab-4489-9d8f-aa0676728f56 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2] Took 1.34 seconds to destroy the instance on the hypervisor. [ 1028.911589] env[61957]: DEBUG oslo.service.loopingcall [None req-507275cf-57ab-4489-9d8f-aa0676728f56 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1028.912586] env[61957]: DEBUG nova.compute.manager [-] [instance: 08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1028.912713] env[61957]: DEBUG nova.network.neutron [-] [instance: 08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1028.959539] env[61957]: DEBUG neutronclient.v2_0.client [-] Error message: {"NeutronError": {"type": "PortNotFound", "message": "Port 00c8220d-5633-4de7-9795-8a6a8543c4ec could not be found.", "detail": ""}} {{(pid=61957) _handle_fault_response /opt/stack/data/venv/lib/python3.10/site-packages/neutronclient/v2_0/client.py:262}} [ 1028.960023] env[61957]: DEBUG nova.network.neutron [-] Unable to show port 00c8220d-5633-4de7-9795-8a6a8543c4ec as it no longer exists. {{(pid=61957) _unbind_ports /opt/stack/nova/nova/network/neutron.py:666}} [ 1029.060572] env[61957]: DEBUG oslo_vmware.api [None req-9a7c3783-e05c-490e-9f05-e499ff43e1b0 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Task: {'id': task-1278004, 'name': ReconfigVM_Task, 'duration_secs': 0.278825} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.060872] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-9a7c3783-e05c-490e-9f05-e499ff43e1b0 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] [instance: 964c8c9d-0721-4ad3-a146-fce1fd1d6513] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-274596', 'volume_id': '33adb5ac-f7e3-4840-a689-022217fc5da8', 'name': 'volume-33adb5ac-f7e3-4840-a689-022217fc5da8', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '964c8c9d-0721-4ad3-a146-fce1fd1d6513', 'attached_at': '', 'detached_at': '', 'volume_id': '33adb5ac-f7e3-4840-a689-022217fc5da8', 'serial': '33adb5ac-f7e3-4840-a689-022217fc5da8'} {{(pid=61957) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1029.136596] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1278006, 'name': CreateVM_Task, 'duration_secs': 0.316556} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.136864] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9b0053fe-420e-4b92-86a8-ed44f6a6ec49] Created VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1029.140275] env[61957]: DEBUG oslo_concurrency.lockutils [None req-3ec44fbf-1f22-4ad1-8b26-d0253bbe3176 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/f04c1641-6ff1-4b2c-9748-9cf176307db0" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1029.140443] env[61957]: DEBUG oslo_concurrency.lockutils [None req-3ec44fbf-1f22-4ad1-8b26-d0253bbe3176 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Acquired lock "[datastore2] devstack-image-cache_base/f04c1641-6ff1-4b2c-9748-9cf176307db0" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1029.140832] env[61957]: DEBUG oslo_concurrency.lockutils [None req-3ec44fbf-1f22-4ad1-8b26-d0253bbe3176 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/f04c1641-6ff1-4b2c-9748-9cf176307db0" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1029.141123] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e6adec2e-fbbb-4fdb-9852-aa4202cec597 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.149307] env[61957]: DEBUG oslo_vmware.api [None req-fa4af5a4-f456-4969-8ee2-40fce4f1cd43 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': task-1278007, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.175407} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.150581] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-fa4af5a4-f456-4969-8ee2-40fce4f1cd43 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Deleted the datastore file {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1029.150781] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-fa4af5a4-f456-4969-8ee2-40fce4f1cd43 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 3f144012-c0cc-43ce-8c73-4422fe8c1fe6] Deleted contents of the VM from datastore datastore1 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1029.150979] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-fa4af5a4-f456-4969-8ee2-40fce4f1cd43 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 3f144012-c0cc-43ce-8c73-4422fe8c1fe6] Instance destroyed {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1029.151452] env[61957]: INFO nova.compute.manager [None req-fa4af5a4-f456-4969-8ee2-40fce4f1cd43 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 3f144012-c0cc-43ce-8c73-4422fe8c1fe6] Took 1.19 seconds to destroy the instance on the hypervisor. [ 1029.151720] env[61957]: DEBUG oslo.service.loopingcall [None req-fa4af5a4-f456-4969-8ee2-40fce4f1cd43 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1029.152016] env[61957]: DEBUG oslo_vmware.api [None req-3ec44fbf-1f22-4ad1-8b26-d0253bbe3176 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Waiting for the task: (returnval){ [ 1029.152016] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]520ba29d-a51e-9be0-eaec-800b5f1858a7" [ 1029.152016] env[61957]: _type = "Task" [ 1029.152016] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.152237] env[61957]: DEBUG nova.compute.manager [-] [instance: 3f144012-c0cc-43ce-8c73-4422fe8c1fe6] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1029.152346] env[61957]: DEBUG nova.network.neutron [-] [instance: 3f144012-c0cc-43ce-8c73-4422fe8c1fe6] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1029.164466] env[61957]: DEBUG oslo_concurrency.lockutils [None req-3ec44fbf-1f22-4ad1-8b26-d0253bbe3176 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Releasing lock "[datastore2] devstack-image-cache_base/f04c1641-6ff1-4b2c-9748-9cf176307db0" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1029.164729] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-3ec44fbf-1f22-4ad1-8b26-d0253bbe3176 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] [instance: 9b0053fe-420e-4b92-86a8-ed44f6a6ec49] Processing image f04c1641-6ff1-4b2c-9748-9cf176307db0 {{(pid=61957) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1029.164969] env[61957]: DEBUG oslo_concurrency.lockutils [None req-3ec44fbf-1f22-4ad1-8b26-d0253bbe3176 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/f04c1641-6ff1-4b2c-9748-9cf176307db0/f04c1641-6ff1-4b2c-9748-9cf176307db0.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1029.165134] env[61957]: DEBUG oslo_concurrency.lockutils [None req-3ec44fbf-1f22-4ad1-8b26-d0253bbe3176 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Acquired lock "[datastore2] devstack-image-cache_base/f04c1641-6ff1-4b2c-9748-9cf176307db0/f04c1641-6ff1-4b2c-9748-9cf176307db0.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1029.165319] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-3ec44fbf-1f22-4ad1-8b26-d0253bbe3176 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1029.165587] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-33ccbd31-6260-411c-aa9f-69817825dff4 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.175077] env[61957]: DEBUG oslo_concurrency.lockutils [None req-4ca8b060-369e-429d-bf7a-53fb2d00bd30 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.832s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1029.180390] env[61957]: DEBUG oslo_concurrency.lockutils [None req-9f803f6c-daf3-4957-bc49-ff16e02c7831 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.093s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1029.181028] env[61957]: DEBUG nova.objects.instance [None req-9f803f6c-daf3-4957-bc49-ff16e02c7831 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Lazy-loading 'resources' on Instance uuid 927af4c3-c7e8-404c-839e-d25af5e384a9 {{(pid=61957) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1029.182019] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-3ec44fbf-1f22-4ad1-8b26-d0253bbe3176 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1029.182433] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-3ec44fbf-1f22-4ad1-8b26-d0253bbe3176 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61957) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1029.183408] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-11d976d8-93e7-4448-917d-f4b337d60a7d {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.189277] env[61957]: DEBUG oslo_vmware.api [None req-3ec44fbf-1f22-4ad1-8b26-d0253bbe3176 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Waiting for the task: (returnval){ [ 1029.189277] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52ef2a6b-60c5-2f3f-f9b9-2da131c03a68" [ 1029.189277] env[61957]: _type = "Task" [ 1029.189277] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.200419] env[61957]: DEBUG oslo_vmware.api [None req-3ec44fbf-1f22-4ad1-8b26-d0253bbe3176 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52ef2a6b-60c5-2f3f-f9b9-2da131c03a68, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.212895] env[61957]: INFO nova.scheduler.client.report [None req-4ca8b060-369e-429d-bf7a-53fb2d00bd30 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Deleted allocations for instance 526ff179-62a6-4763-ab25-797617c4ed57 [ 1029.620289] env[61957]: DEBUG nova.objects.instance [None req-9a7c3783-e05c-490e-9f05-e499ff43e1b0 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Lazy-loading 'flavor' on Instance uuid 964c8c9d-0721-4ad3-a146-fce1fd1d6513 {{(pid=61957) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1029.673402] env[61957]: DEBUG nova.compute.manager [req-7e35d1e5-14d7-442f-8073-55e14fd5fdf2 req-94563c81-3bf2-46f1-b8aa-98d2cae759a8 service nova] [instance: 08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2] Received event network-vif-deleted-f62a9f12-0b55-4138-a40e-b431855487fe {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1029.673576] env[61957]: INFO nova.compute.manager [req-7e35d1e5-14d7-442f-8073-55e14fd5fdf2 req-94563c81-3bf2-46f1-b8aa-98d2cae759a8 service nova] [instance: 08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2] Neutron deleted interface f62a9f12-0b55-4138-a40e-b431855487fe; detaching it from the instance and deleting it from the info cache [ 1029.673824] env[61957]: DEBUG nova.network.neutron [req-7e35d1e5-14d7-442f-8073-55e14fd5fdf2 req-94563c81-3bf2-46f1-b8aa-98d2cae759a8 service nova] [instance: 08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1029.701265] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-3ec44fbf-1f22-4ad1-8b26-d0253bbe3176 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] [instance: 9b0053fe-420e-4b92-86a8-ed44f6a6ec49] Preparing fetch location {{(pid=61957) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1029.701537] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-3ec44fbf-1f22-4ad1-8b26-d0253bbe3176 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] [instance: 9b0053fe-420e-4b92-86a8-ed44f6a6ec49] Fetch image to [datastore2] OSTACK_IMG_e5927446-8842-4986-883f-0c074bfdd8b3/OSTACK_IMG_e5927446-8842-4986-883f-0c074bfdd8b3.vmdk {{(pid=61957) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1029.701721] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-3ec44fbf-1f22-4ad1-8b26-d0253bbe3176 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] [instance: 9b0053fe-420e-4b92-86a8-ed44f6a6ec49] Downloading stream optimized image f04c1641-6ff1-4b2c-9748-9cf176307db0 to [datastore2] OSTACK_IMG_e5927446-8842-4986-883f-0c074bfdd8b3/OSTACK_IMG_e5927446-8842-4986-883f-0c074bfdd8b3.vmdk on the data store datastore2 as vApp {{(pid=61957) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1029.701956] env[61957]: DEBUG nova.virt.vmwareapi.images [None req-3ec44fbf-1f22-4ad1-8b26-d0253bbe3176 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] [instance: 9b0053fe-420e-4b92-86a8-ed44f6a6ec49] Downloading image file data f04c1641-6ff1-4b2c-9748-9cf176307db0 to the ESX as VM named 'OSTACK_IMG_e5927446-8842-4986-883f-0c074bfdd8b3' {{(pid=61957) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1029.729083] env[61957]: DEBUG oslo_concurrency.lockutils [None req-4ca8b060-369e-429d-bf7a-53fb2d00bd30 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Lock "526ff179-62a6-4763-ab25-797617c4ed57" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.874s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1029.814717] env[61957]: DEBUG oslo_vmware.rw_handles [None req-3ec44fbf-1f22-4ad1-8b26-d0253bbe3176 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1029.814717] env[61957]: value = "resgroup-9" [ 1029.814717] env[61957]: _type = "ResourcePool" [ 1029.814717] env[61957]: }. {{(pid=61957) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1029.815037] env[61957]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-9fda6a5d-846f-4238-8fe2-c979a436ddf4 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.841219] env[61957]: DEBUG oslo_vmware.rw_handles [None req-3ec44fbf-1f22-4ad1-8b26-d0253bbe3176 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Lease: (returnval){ [ 1029.841219] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52bfe838-795a-956e-d59a-b4865e0dc87e" [ 1029.841219] env[61957]: _type = "HttpNfcLease" [ 1029.841219] env[61957]: } obtained for vApp import into resource pool (val){ [ 1029.841219] env[61957]: value = "resgroup-9" [ 1029.841219] env[61957]: _type = "ResourcePool" [ 1029.841219] env[61957]: }. {{(pid=61957) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1029.841219] env[61957]: DEBUG oslo_vmware.api [None req-3ec44fbf-1f22-4ad1-8b26-d0253bbe3176 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Waiting for the lease: (returnval){ [ 1029.841219] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52bfe838-795a-956e-d59a-b4865e0dc87e" [ 1029.841219] env[61957]: _type = "HttpNfcLease" [ 1029.841219] env[61957]: } to be ready. {{(pid=61957) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1029.851686] env[61957]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1029.851686] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52bfe838-795a-956e-d59a-b4865e0dc87e" [ 1029.851686] env[61957]: _type = "HttpNfcLease" [ 1029.851686] env[61957]: } is initializing. {{(pid=61957) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1029.883873] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d991d2f-82de-4f2b-950b-64f9ef4f45a4 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.891449] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb037b04-d662-4f59-930b-0393b15b0115 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.925692] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e822dad-7ad9-402e-ae5d-cdb6388bf64f {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.933636] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-550b4dde-aef9-4767-bd94-22596ec3a8c1 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.947733] env[61957]: DEBUG nova.compute.provider_tree [None req-9f803f6c-daf3-4957-bc49-ff16e02c7831 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Updating inventory in ProviderTree for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1030.002711] env[61957]: DEBUG nova.network.neutron [-] [instance: 08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1030.052250] env[61957]: DEBUG nova.network.neutron [-] [instance: 3f144012-c0cc-43ce-8c73-4422fe8c1fe6] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1030.178885] env[61957]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-45f7392b-7805-48fe-8199-66a0eb35f6a7 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.188635] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91e469e4-4ab1-4a40-8d53-16071d3f03f8 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.223640] env[61957]: DEBUG nova.compute.manager [req-7e35d1e5-14d7-442f-8073-55e14fd5fdf2 req-94563c81-3bf2-46f1-b8aa-98d2cae759a8 service nova] [instance: 08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2] Detach interface failed, port_id=f62a9f12-0b55-4138-a40e-b431855487fe, reason: Instance 08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2 could not be found. {{(pid=61957) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1030.224233] env[61957]: DEBUG nova.compute.manager [req-7e35d1e5-14d7-442f-8073-55e14fd5fdf2 req-94563c81-3bf2-46f1-b8aa-98d2cae759a8 service nova] [instance: 3f144012-c0cc-43ce-8c73-4422fe8c1fe6] Received event network-vif-deleted-921a0283-f0c8-4cd9-8548-5e1089a438c3 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1030.224233] env[61957]: INFO nova.compute.manager [req-7e35d1e5-14d7-442f-8073-55e14fd5fdf2 req-94563c81-3bf2-46f1-b8aa-98d2cae759a8 service nova] [instance: 3f144012-c0cc-43ce-8c73-4422fe8c1fe6] Neutron deleted interface 921a0283-f0c8-4cd9-8548-5e1089a438c3; detaching it from the instance and deleting it from the info cache [ 1030.224233] env[61957]: DEBUG nova.network.neutron [req-7e35d1e5-14d7-442f-8073-55e14fd5fdf2 req-94563c81-3bf2-46f1-b8aa-98d2cae759a8 service nova] [instance: 3f144012-c0cc-43ce-8c73-4422fe8c1fe6] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1030.350845] env[61957]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1030.350845] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52bfe838-795a-956e-d59a-b4865e0dc87e" [ 1030.350845] env[61957]: _type = "HttpNfcLease" [ 1030.350845] env[61957]: } is initializing. {{(pid=61957) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1030.469618] env[61957]: ERROR nova.scheduler.client.report [None req-9f803f6c-daf3-4957-bc49-ff16e02c7831 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [req-ff8af5d1-e922-4756-bd1d-559094a82fa0] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 0ceb6c9e-61c6-496d-8579-9d32627e96da. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-ff8af5d1-e922-4756-bd1d-559094a82fa0"}]} [ 1030.489027] env[61957]: DEBUG nova.scheduler.client.report [None req-9f803f6c-daf3-4957-bc49-ff16e02c7831 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Refreshing inventories for resource provider 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 1030.503852] env[61957]: DEBUG nova.scheduler.client.report [None req-9f803f6c-daf3-4957-bc49-ff16e02c7831 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Updating ProviderTree inventory for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 1030.504223] env[61957]: DEBUG nova.compute.provider_tree [None req-9f803f6c-daf3-4957-bc49-ff16e02c7831 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Updating inventory in ProviderTree for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1030.506815] env[61957]: INFO nova.compute.manager [-] [instance: 08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2] Took 1.59 seconds to deallocate network for instance. [ 1030.518117] env[61957]: DEBUG nova.scheduler.client.report [None req-9f803f6c-daf3-4957-bc49-ff16e02c7831 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Refreshing aggregate associations for resource provider 0ceb6c9e-61c6-496d-8579-9d32627e96da, aggregates: None {{(pid=61957) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 1030.536813] env[61957]: DEBUG nova.scheduler.client.report [None req-9f803f6c-daf3-4957-bc49-ff16e02c7831 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Refreshing trait associations for resource provider 0ceb6c9e-61c6-496d-8579-9d32627e96da, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO,HW_ARCH_X86_64,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=61957) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 1030.560113] env[61957]: INFO nova.compute.manager [-] [instance: 3f144012-c0cc-43ce-8c73-4422fe8c1fe6] Took 1.41 seconds to deallocate network for instance. [ 1030.631872] env[61957]: DEBUG oslo_concurrency.lockutils [None req-9a7c3783-e05c-490e-9f05-e499ff43e1b0 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Lock "964c8c9d-0721-4ad3-a146-fce1fd1d6513" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.294s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1030.697688] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d80b2ba5-e013-4ca1-85ab-4e58eccb3f07 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.705673] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f19b6e4-6472-492d-ac8a-6ce30482c704 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.737729] env[61957]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-cfa38d14-754a-43bf-bbf5-44bc16ded933 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.740261] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7f03314-2af5-4233-acb0-0b92e6239e78 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.748996] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-312052c9-afaf-43eb-af1e-9afe6deab857 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.755776] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0285db6b-7b13-491a-8844-3c4c7a157b8a {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.776982] env[61957]: DEBUG nova.compute.provider_tree [None req-9f803f6c-daf3-4957-bc49-ff16e02c7831 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Updating inventory in ProviderTree for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1030.785217] env[61957]: DEBUG nova.compute.manager [req-7e35d1e5-14d7-442f-8073-55e14fd5fdf2 req-94563c81-3bf2-46f1-b8aa-98d2cae759a8 service nova] [instance: 3f144012-c0cc-43ce-8c73-4422fe8c1fe6] Detach interface failed, port_id=921a0283-f0c8-4cd9-8548-5e1089a438c3, reason: Instance 3f144012-c0cc-43ce-8c73-4422fe8c1fe6 could not be found. {{(pid=61957) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1030.814457] env[61957]: DEBUG nova.scheduler.client.report [None req-9f803f6c-daf3-4957-bc49-ff16e02c7831 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Updated inventory for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da with generation 115 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 1030.814457] env[61957]: DEBUG nova.compute.provider_tree [None req-9f803f6c-daf3-4957-bc49-ff16e02c7831 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Updating resource provider 0ceb6c9e-61c6-496d-8579-9d32627e96da generation from 115 to 116 during operation: update_inventory {{(pid=61957) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1030.814457] env[61957]: DEBUG nova.compute.provider_tree [None req-9f803f6c-daf3-4957-bc49-ff16e02c7831 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Updating inventory in ProviderTree for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1030.850247] env[61957]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1030.850247] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52bfe838-795a-956e-d59a-b4865e0dc87e" [ 1030.850247] env[61957]: _type = "HttpNfcLease" [ 1030.850247] env[61957]: } is ready. {{(pid=61957) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1030.850554] env[61957]: DEBUG oslo_vmware.rw_handles [None req-3ec44fbf-1f22-4ad1-8b26-d0253bbe3176 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1030.850554] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52bfe838-795a-956e-d59a-b4865e0dc87e" [ 1030.850554] env[61957]: _type = "HttpNfcLease" [ 1030.850554] env[61957]: }. {{(pid=61957) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1030.851279] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a64d14a9-bce1-44a6-9e93-3583017daa96 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.860197] env[61957]: DEBUG oslo_vmware.rw_handles [None req-3ec44fbf-1f22-4ad1-8b26-d0253bbe3176 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/527b8b7e-bb12-a1ad-8bbc-c4bb048205ba/disk-0.vmdk from lease info. {{(pid=61957) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1030.860197] env[61957]: DEBUG oslo_vmware.rw_handles [None req-3ec44fbf-1f22-4ad1-8b26-d0253bbe3176 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Creating HTTP connection to write to file with size = 31667200 and URL = https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/527b8b7e-bb12-a1ad-8bbc-c4bb048205ba/disk-0.vmdk. {{(pid=61957) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1030.924066] env[61957]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-adf039c2-3eae-4984-b3dc-4a064728d8ea {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.016182] env[61957]: DEBUG oslo_concurrency.lockutils [None req-507275cf-57ab-4489-9d8f-aa0676728f56 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1031.067614] env[61957]: DEBUG oslo_concurrency.lockutils [None req-fa4af5a4-f456-4969-8ee2-40fce4f1cd43 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1031.319478] env[61957]: DEBUG oslo_concurrency.lockutils [None req-9f803f6c-daf3-4957-bc49-ff16e02c7831 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.139s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1031.326370] env[61957]: DEBUG oslo_concurrency.lockutils [None req-49c7a8aa-e585-4bf6-ae5e-46c63c7fde84 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 3.289s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1031.355182] env[61957]: INFO nova.scheduler.client.report [None req-9f803f6c-daf3-4957-bc49-ff16e02c7831 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Deleted allocations for instance 927af4c3-c7e8-404c-839e-d25af5e384a9 [ 1031.814580] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d07fdeef-3a24-45b1-a008-52046b4459b1 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Acquiring lock "964c8c9d-0721-4ad3-a146-fce1fd1d6513" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1031.814885] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d07fdeef-3a24-45b1-a008-52046b4459b1 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Lock "964c8c9d-0721-4ad3-a146-fce1fd1d6513" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1031.815108] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d07fdeef-3a24-45b1-a008-52046b4459b1 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Acquiring lock "964c8c9d-0721-4ad3-a146-fce1fd1d6513-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1031.815300] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d07fdeef-3a24-45b1-a008-52046b4459b1 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Lock "964c8c9d-0721-4ad3-a146-fce1fd1d6513-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1031.815477] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d07fdeef-3a24-45b1-a008-52046b4459b1 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Lock "964c8c9d-0721-4ad3-a146-fce1fd1d6513-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1031.818494] env[61957]: INFO nova.compute.manager [None req-d07fdeef-3a24-45b1-a008-52046b4459b1 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] [instance: 964c8c9d-0721-4ad3-a146-fce1fd1d6513] Terminating instance [ 1031.820384] env[61957]: DEBUG nova.compute.manager [None req-d07fdeef-3a24-45b1-a008-52046b4459b1 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] [instance: 964c8c9d-0721-4ad3-a146-fce1fd1d6513] Start destroying the instance on the hypervisor. {{(pid=61957) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1031.820580] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-d07fdeef-3a24-45b1-a008-52046b4459b1 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] [instance: 964c8c9d-0721-4ad3-a146-fce1fd1d6513] Destroying instance {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1031.821466] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8264b23c-b567-440b-8a4d-2975ddb40e61 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.831786] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-d07fdeef-3a24-45b1-a008-52046b4459b1 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] [instance: 964c8c9d-0721-4ad3-a146-fce1fd1d6513] Powering off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1031.834035] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b5ecc022-3aaf-4336-a4f4-58a50b3dd163 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.841887] env[61957]: DEBUG oslo_vmware.api [None req-d07fdeef-3a24-45b1-a008-52046b4459b1 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Waiting for the task: (returnval){ [ 1031.841887] env[61957]: value = "task-1278009" [ 1031.841887] env[61957]: _type = "Task" [ 1031.841887] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.854414] env[61957]: DEBUG oslo_vmware.api [None req-d07fdeef-3a24-45b1-a008-52046b4459b1 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Task: {'id': task-1278009, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.866034] env[61957]: DEBUG oslo_concurrency.lockutils [None req-9f803f6c-daf3-4957-bc49-ff16e02c7831 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Lock "927af4c3-c7e8-404c-839e-d25af5e384a9" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.163s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1031.986386] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62d323c2-c43b-4e95-b5f9-066b58119036 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.997112] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-715ab925-b032-42e7-bd03-2f8f4b3978b5 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.039261] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b29926f-7f8c-4161-8f1b-5327dc22f2c9 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.050962] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b2df902-5355-428b-86c6-66f2e54e9403 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.068474] env[61957]: DEBUG nova.compute.provider_tree [None req-49c7a8aa-e585-4bf6-ae5e-46c63c7fde84 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1032.111223] env[61957]: DEBUG oslo_concurrency.lockutils [None req-de4650d3-4738-443f-8147-36e90c68fb19 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Acquiring lock "039a6b01-a745-478c-9ff7-c8c37afe2ce4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1032.111472] env[61957]: DEBUG oslo_concurrency.lockutils [None req-de4650d3-4738-443f-8147-36e90c68fb19 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Lock "039a6b01-a745-478c-9ff7-c8c37afe2ce4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1032.203148] env[61957]: DEBUG oslo_vmware.rw_handles [None req-3ec44fbf-1f22-4ad1-8b26-d0253bbe3176 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Completed reading data from the image iterator. {{(pid=61957) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1032.203441] env[61957]: DEBUG oslo_vmware.rw_handles [None req-3ec44fbf-1f22-4ad1-8b26-d0253bbe3176 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/527b8b7e-bb12-a1ad-8bbc-c4bb048205ba/disk-0.vmdk. {{(pid=61957) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1032.204352] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3816df94-cf1c-46a2-869d-54365544f400 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.211290] env[61957]: DEBUG oslo_vmware.rw_handles [None req-3ec44fbf-1f22-4ad1-8b26-d0253bbe3176 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/527b8b7e-bb12-a1ad-8bbc-c4bb048205ba/disk-0.vmdk is in state: ready. {{(pid=61957) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1032.211435] env[61957]: DEBUG oslo_vmware.rw_handles [None req-3ec44fbf-1f22-4ad1-8b26-d0253bbe3176 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Releasing lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/527b8b7e-bb12-a1ad-8bbc-c4bb048205ba/disk-0.vmdk. {{(pid=61957) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1032.211563] env[61957]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-1e420d17-080a-487b-8eaf-25cda081acb0 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.351550] env[61957]: DEBUG oslo_vmware.api [None req-d07fdeef-3a24-45b1-a008-52046b4459b1 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Task: {'id': task-1278009, 'name': PowerOffVM_Task, 'duration_secs': 0.282478} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.351871] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-d07fdeef-3a24-45b1-a008-52046b4459b1 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] [instance: 964c8c9d-0721-4ad3-a146-fce1fd1d6513] Powered off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1032.352015] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-d07fdeef-3a24-45b1-a008-52046b4459b1 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] [instance: 964c8c9d-0721-4ad3-a146-fce1fd1d6513] Unregistering the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1032.352347] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7c227d4e-4395-4e57-b3fb-c9e297248ed4 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.412949] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-d07fdeef-3a24-45b1-a008-52046b4459b1 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] [instance: 964c8c9d-0721-4ad3-a146-fce1fd1d6513] Unregistered the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1032.413234] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-d07fdeef-3a24-45b1-a008-52046b4459b1 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] [instance: 964c8c9d-0721-4ad3-a146-fce1fd1d6513] Deleting contents of the VM from datastore datastore1 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1032.413449] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-d07fdeef-3a24-45b1-a008-52046b4459b1 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Deleting the datastore file [datastore1] 964c8c9d-0721-4ad3-a146-fce1fd1d6513 {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1032.413722] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1fef7023-cb49-4551-9f70-2b22f904784e {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.422096] env[61957]: DEBUG oslo_vmware.api [None req-d07fdeef-3a24-45b1-a008-52046b4459b1 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Waiting for the task: (returnval){ [ 1032.422096] env[61957]: value = "task-1278011" [ 1032.422096] env[61957]: _type = "Task" [ 1032.422096] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.429913] env[61957]: DEBUG oslo_vmware.api [None req-d07fdeef-3a24-45b1-a008-52046b4459b1 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Task: {'id': task-1278011, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.468797] env[61957]: DEBUG oslo_vmware.rw_handles [None req-3ec44fbf-1f22-4ad1-8b26-d0253bbe3176 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Closed VMDK write handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/527b8b7e-bb12-a1ad-8bbc-c4bb048205ba/disk-0.vmdk. {{(pid=61957) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1032.469092] env[61957]: INFO nova.virt.vmwareapi.images [None req-3ec44fbf-1f22-4ad1-8b26-d0253bbe3176 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] [instance: 9b0053fe-420e-4b92-86a8-ed44f6a6ec49] Downloaded image file data f04c1641-6ff1-4b2c-9748-9cf176307db0 [ 1032.469926] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f99ca38c-abf8-40b8-86f8-7e2fb146734b {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.488080] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d9e43fdc-6e08-43a8-9105-8839a1580d67 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.517539] env[61957]: INFO nova.virt.vmwareapi.images [None req-3ec44fbf-1f22-4ad1-8b26-d0253bbe3176 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] [instance: 9b0053fe-420e-4b92-86a8-ed44f6a6ec49] The imported VM was unregistered [ 1032.520527] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-3ec44fbf-1f22-4ad1-8b26-d0253bbe3176 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] [instance: 9b0053fe-420e-4b92-86a8-ed44f6a6ec49] Caching image {{(pid=61957) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1032.520527] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-3ec44fbf-1f22-4ad1-8b26-d0253bbe3176 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Creating directory with path [datastore2] devstack-image-cache_base/f04c1641-6ff1-4b2c-9748-9cf176307db0 {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1032.520934] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-59199e8c-7915-44e0-a626-694aa60fbc1b {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.531664] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-3ec44fbf-1f22-4ad1-8b26-d0253bbe3176 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Created directory with path [datastore2] devstack-image-cache_base/f04c1641-6ff1-4b2c-9748-9cf176307db0 {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1032.531938] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-3ec44fbf-1f22-4ad1-8b26-d0253bbe3176 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Moving virtual disk from [datastore2] OSTACK_IMG_e5927446-8842-4986-883f-0c074bfdd8b3/OSTACK_IMG_e5927446-8842-4986-883f-0c074bfdd8b3.vmdk to [datastore2] devstack-image-cache_base/f04c1641-6ff1-4b2c-9748-9cf176307db0/f04c1641-6ff1-4b2c-9748-9cf176307db0.vmdk. {{(pid=61957) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1032.532298] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-e9e41f91-00cb-434c-b206-4cb5d666ecd3 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.539129] env[61957]: DEBUG oslo_vmware.api [None req-3ec44fbf-1f22-4ad1-8b26-d0253bbe3176 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Waiting for the task: (returnval){ [ 1032.539129] env[61957]: value = "task-1278013" [ 1032.539129] env[61957]: _type = "Task" [ 1032.539129] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.547636] env[61957]: DEBUG oslo_vmware.api [None req-3ec44fbf-1f22-4ad1-8b26-d0253bbe3176 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Task: {'id': task-1278013, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.571828] env[61957]: DEBUG nova.scheduler.client.report [None req-49c7a8aa-e585-4bf6-ae5e-46c63c7fde84 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1032.615657] env[61957]: DEBUG nova.compute.manager [None req-de4650d3-4738-443f-8147-36e90c68fb19 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] [instance: 039a6b01-a745-478c-9ff7-c8c37afe2ce4] Starting instance... {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1032.931760] env[61957]: DEBUG oslo_vmware.api [None req-d07fdeef-3a24-45b1-a008-52046b4459b1 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Task: {'id': task-1278011, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.162929} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.932327] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-d07fdeef-3a24-45b1-a008-52046b4459b1 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Deleted the datastore file {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1032.932327] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-d07fdeef-3a24-45b1-a008-52046b4459b1 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] [instance: 964c8c9d-0721-4ad3-a146-fce1fd1d6513] Deleted contents of the VM from datastore datastore1 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1032.932509] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-d07fdeef-3a24-45b1-a008-52046b4459b1 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] [instance: 964c8c9d-0721-4ad3-a146-fce1fd1d6513] Instance destroyed {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1032.932702] env[61957]: INFO nova.compute.manager [None req-d07fdeef-3a24-45b1-a008-52046b4459b1 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] [instance: 964c8c9d-0721-4ad3-a146-fce1fd1d6513] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1032.933026] env[61957]: DEBUG oslo.service.loopingcall [None req-d07fdeef-3a24-45b1-a008-52046b4459b1 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1032.933234] env[61957]: DEBUG nova.compute.manager [-] [instance: 964c8c9d-0721-4ad3-a146-fce1fd1d6513] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1032.933333] env[61957]: DEBUG nova.network.neutron [-] [instance: 964c8c9d-0721-4ad3-a146-fce1fd1d6513] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1033.011746] env[61957]: DEBUG oslo_concurrency.lockutils [None req-0d170cdb-983c-43d4-9998-8676a1404b75 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Acquiring lock "42860944-c700-43be-b291-970803fc7955" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1033.011972] env[61957]: DEBUG oslo_concurrency.lockutils [None req-0d170cdb-983c-43d4-9998-8676a1404b75 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Lock "42860944-c700-43be-b291-970803fc7955" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1033.049852] env[61957]: DEBUG oslo_vmware.api [None req-3ec44fbf-1f22-4ad1-8b26-d0253bbe3176 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Task: {'id': task-1278013, 'name': MoveVirtualDisk_Task} progress is 24%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.136724] env[61957]: DEBUG oslo_concurrency.lockutils [None req-de4650d3-4738-443f-8147-36e90c68fb19 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1033.467672] env[61957]: DEBUG nova.compute.manager [req-babec671-ad52-452b-bd6f-34a2e163bd3d req-2c6bbe92-dc61-4cfd-8edb-9366699935a7 service nova] [instance: 964c8c9d-0721-4ad3-a146-fce1fd1d6513] Received event network-vif-deleted-57483094-30a7-4f77-8c6e-25dbd1db61c7 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1033.467769] env[61957]: INFO nova.compute.manager [req-babec671-ad52-452b-bd6f-34a2e163bd3d req-2c6bbe92-dc61-4cfd-8edb-9366699935a7 service nova] [instance: 964c8c9d-0721-4ad3-a146-fce1fd1d6513] Neutron deleted interface 57483094-30a7-4f77-8c6e-25dbd1db61c7; detaching it from the instance and deleting it from the info cache [ 1033.467996] env[61957]: DEBUG nova.network.neutron [req-babec671-ad52-452b-bd6f-34a2e163bd3d req-2c6bbe92-dc61-4cfd-8edb-9366699935a7 service nova] [instance: 964c8c9d-0721-4ad3-a146-fce1fd1d6513] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1033.514573] env[61957]: DEBUG nova.compute.manager [None req-0d170cdb-983c-43d4-9998-8676a1404b75 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 42860944-c700-43be-b291-970803fc7955] Starting instance... {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1033.550474] env[61957]: DEBUG oslo_vmware.api [None req-3ec44fbf-1f22-4ad1-8b26-d0253bbe3176 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Task: {'id': task-1278013, 'name': MoveVirtualDisk_Task} progress is 46%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.581849] env[61957]: DEBUG oslo_concurrency.lockutils [None req-49c7a8aa-e585-4bf6-ae5e-46c63c7fde84 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.257s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1033.585307] env[61957]: DEBUG oslo_concurrency.lockutils [None req-507275cf-57ab-4489-9d8f-aa0676728f56 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.569s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1033.585568] env[61957]: DEBUG nova.objects.instance [None req-507275cf-57ab-4489-9d8f-aa0676728f56 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Lazy-loading 'resources' on Instance uuid 08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2 {{(pid=61957) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1033.930847] env[61957]: DEBUG nova.network.neutron [-] [instance: 964c8c9d-0721-4ad3-a146-fce1fd1d6513] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1033.971341] env[61957]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2fa5ba7e-2d9a-4929-af79-e70e7d0ecdb0 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.981753] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee810b15-5177-4f81-a8f6-00c98bbfb6af {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.008901] env[61957]: DEBUG nova.compute.manager [req-babec671-ad52-452b-bd6f-34a2e163bd3d req-2c6bbe92-dc61-4cfd-8edb-9366699935a7 service nova] [instance: 964c8c9d-0721-4ad3-a146-fce1fd1d6513] Detach interface failed, port_id=57483094-30a7-4f77-8c6e-25dbd1db61c7, reason: Instance 964c8c9d-0721-4ad3-a146-fce1fd1d6513 could not be found. {{(pid=61957) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1034.039353] env[61957]: DEBUG oslo_concurrency.lockutils [None req-0d170cdb-983c-43d4-9998-8676a1404b75 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1034.049830] env[61957]: DEBUG oslo_vmware.api [None req-3ec44fbf-1f22-4ad1-8b26-d0253bbe3176 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Task: {'id': task-1278013, 'name': MoveVirtualDisk_Task} progress is 66%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.157512] env[61957]: INFO nova.scheduler.client.report [None req-49c7a8aa-e585-4bf6-ae5e-46c63c7fde84 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Deleted allocation for migration c0538d2d-e6c4-4a67-b30e-dadb1e5d8f8e [ 1034.272995] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09640f9e-e714-46aa-acfc-6c02af01948e {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.281464] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a262e26-e6a3-45e6-89f2-47b38c671f75 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.314430] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86b5ab18-498a-42ec-9915-9f7d10c8deb5 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.322985] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c918031-5c62-4e63-b5f3-77d44b53aa4a {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.339047] env[61957]: DEBUG nova.compute.provider_tree [None req-507275cf-57ab-4489-9d8f-aa0676728f56 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1034.433298] env[61957]: INFO nova.compute.manager [-] [instance: 964c8c9d-0721-4ad3-a146-fce1fd1d6513] Took 1.50 seconds to deallocate network for instance. [ 1034.552451] env[61957]: DEBUG oslo_vmware.api [None req-3ec44fbf-1f22-4ad1-8b26-d0253bbe3176 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Task: {'id': task-1278013, 'name': MoveVirtualDisk_Task} progress is 88%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.664653] env[61957]: DEBUG oslo_concurrency.lockutils [None req-49c7a8aa-e585-4bf6-ae5e-46c63c7fde84 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Lock "6509707e-afc6-4c4a-88f9-a471865d510c" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 9.990s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1034.748824] env[61957]: DEBUG oslo_concurrency.lockutils [None req-5a8aaf0c-6884-47c6-aa4a-9ccf495b4481 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Acquiring lock "6509707e-afc6-4c4a-88f9-a471865d510c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1034.749148] env[61957]: DEBUG oslo_concurrency.lockutils [None req-5a8aaf0c-6884-47c6-aa4a-9ccf495b4481 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Lock "6509707e-afc6-4c4a-88f9-a471865d510c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1034.749361] env[61957]: DEBUG oslo_concurrency.lockutils [None req-5a8aaf0c-6884-47c6-aa4a-9ccf495b4481 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Acquiring lock "6509707e-afc6-4c4a-88f9-a471865d510c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1034.749550] env[61957]: DEBUG oslo_concurrency.lockutils [None req-5a8aaf0c-6884-47c6-aa4a-9ccf495b4481 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Lock "6509707e-afc6-4c4a-88f9-a471865d510c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1034.749721] env[61957]: DEBUG oslo_concurrency.lockutils [None req-5a8aaf0c-6884-47c6-aa4a-9ccf495b4481 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Lock "6509707e-afc6-4c4a-88f9-a471865d510c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1034.751951] env[61957]: INFO nova.compute.manager [None req-5a8aaf0c-6884-47c6-aa4a-9ccf495b4481 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: 6509707e-afc6-4c4a-88f9-a471865d510c] Terminating instance [ 1034.753988] env[61957]: DEBUG nova.compute.manager [None req-5a8aaf0c-6884-47c6-aa4a-9ccf495b4481 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: 6509707e-afc6-4c4a-88f9-a471865d510c] Start destroying the instance on the hypervisor. {{(pid=61957) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1034.754252] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-5a8aaf0c-6884-47c6-aa4a-9ccf495b4481 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: 6509707e-afc6-4c4a-88f9-a471865d510c] Destroying instance {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1034.755136] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76b99218-1aeb-4900-9700-d0ae2ade7da8 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.763602] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a8aaf0c-6884-47c6-aa4a-9ccf495b4481 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: 6509707e-afc6-4c4a-88f9-a471865d510c] Powering off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1034.763909] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9328c5ca-1927-42ec-8564-74c358232dd2 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.771455] env[61957]: DEBUG oslo_vmware.api [None req-5a8aaf0c-6884-47c6-aa4a-9ccf495b4481 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Waiting for the task: (returnval){ [ 1034.771455] env[61957]: value = "task-1278014" [ 1034.771455] env[61957]: _type = "Task" [ 1034.771455] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1034.779459] env[61957]: DEBUG oslo_vmware.api [None req-5a8aaf0c-6884-47c6-aa4a-9ccf495b4481 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': task-1278014, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.842500] env[61957]: DEBUG nova.scheduler.client.report [None req-507275cf-57ab-4489-9d8f-aa0676728f56 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1034.897651] env[61957]: DEBUG oslo_service.periodic_task [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61957) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1034.897970] env[61957]: DEBUG oslo_service.periodic_task [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61957) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1034.939865] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d07fdeef-3a24-45b1-a008-52046b4459b1 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1035.051900] env[61957]: DEBUG oslo_vmware.api [None req-3ec44fbf-1f22-4ad1-8b26-d0253bbe3176 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Task: {'id': task-1278013, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.318984} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.052281] env[61957]: INFO nova.virt.vmwareapi.ds_util [None req-3ec44fbf-1f22-4ad1-8b26-d0253bbe3176 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Moved virtual disk from [datastore2] OSTACK_IMG_e5927446-8842-4986-883f-0c074bfdd8b3/OSTACK_IMG_e5927446-8842-4986-883f-0c074bfdd8b3.vmdk to [datastore2] devstack-image-cache_base/f04c1641-6ff1-4b2c-9748-9cf176307db0/f04c1641-6ff1-4b2c-9748-9cf176307db0.vmdk. [ 1035.052424] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-3ec44fbf-1f22-4ad1-8b26-d0253bbe3176 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] [instance: 9b0053fe-420e-4b92-86a8-ed44f6a6ec49] Cleaning up location [datastore2] OSTACK_IMG_e5927446-8842-4986-883f-0c074bfdd8b3 {{(pid=61957) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1035.052608] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-3ec44fbf-1f22-4ad1-8b26-d0253bbe3176 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Deleting the datastore file [datastore2] OSTACK_IMG_e5927446-8842-4986-883f-0c074bfdd8b3 {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1035.052856] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-82c0ab4e-f36f-4ad9-87cf-db2ca4b4f32f {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.058923] env[61957]: DEBUG oslo_vmware.api [None req-3ec44fbf-1f22-4ad1-8b26-d0253bbe3176 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Waiting for the task: (returnval){ [ 1035.058923] env[61957]: value = "task-1278015" [ 1035.058923] env[61957]: _type = "Task" [ 1035.058923] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.065910] env[61957]: DEBUG oslo_vmware.api [None req-3ec44fbf-1f22-4ad1-8b26-d0253bbe3176 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Task: {'id': task-1278015, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.235805] env[61957]: DEBUG oslo_concurrency.lockutils [None req-4a6d948f-b9a8-4fc6-bd2c-f6dd4bec3479 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Acquiring lock "a7419b30-1539-4977-a5d9-93a49d84914a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1035.236106] env[61957]: DEBUG oslo_concurrency.lockutils [None req-4a6d948f-b9a8-4fc6-bd2c-f6dd4bec3479 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Lock "a7419b30-1539-4977-a5d9-93a49d84914a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1035.283238] env[61957]: DEBUG oslo_vmware.api [None req-5a8aaf0c-6884-47c6-aa4a-9ccf495b4481 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': task-1278014, 'name': PowerOffVM_Task, 'duration_secs': 0.240039} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.283520] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a8aaf0c-6884-47c6-aa4a-9ccf495b4481 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: 6509707e-afc6-4c4a-88f9-a471865d510c] Powered off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1035.283691] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-5a8aaf0c-6884-47c6-aa4a-9ccf495b4481 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: 6509707e-afc6-4c4a-88f9-a471865d510c] Unregistering the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1035.283934] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-dfa135b4-a4e2-4973-a3df-73c974c615fd {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.343184] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-5a8aaf0c-6884-47c6-aa4a-9ccf495b4481 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: 6509707e-afc6-4c4a-88f9-a471865d510c] Unregistered the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1035.343506] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-5a8aaf0c-6884-47c6-aa4a-9ccf495b4481 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: 6509707e-afc6-4c4a-88f9-a471865d510c] Deleting contents of the VM from datastore datastore1 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1035.343705] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-5a8aaf0c-6884-47c6-aa4a-9ccf495b4481 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Deleting the datastore file [datastore1] 6509707e-afc6-4c4a-88f9-a471865d510c {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1035.343971] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a949ee9b-6587-450c-8f6f-238e0002f3cb {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.347698] env[61957]: DEBUG oslo_concurrency.lockutils [None req-507275cf-57ab-4489-9d8f-aa0676728f56 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.762s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1035.352024] env[61957]: DEBUG oslo_concurrency.lockutils [None req-fa4af5a4-f456-4969-8ee2-40fce4f1cd43 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.284s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1035.352024] env[61957]: DEBUG nova.objects.instance [None req-fa4af5a4-f456-4969-8ee2-40fce4f1cd43 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Lazy-loading 'resources' on Instance uuid 3f144012-c0cc-43ce-8c73-4422fe8c1fe6 {{(pid=61957) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1035.353439] env[61957]: DEBUG oslo_vmware.api [None req-5a8aaf0c-6884-47c6-aa4a-9ccf495b4481 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Waiting for the task: (returnval){ [ 1035.353439] env[61957]: value = "task-1278017" [ 1035.353439] env[61957]: _type = "Task" [ 1035.353439] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.361237] env[61957]: DEBUG oslo_vmware.api [None req-5a8aaf0c-6884-47c6-aa4a-9ccf495b4481 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': task-1278017, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.367344] env[61957]: INFO nova.scheduler.client.report [None req-507275cf-57ab-4489-9d8f-aa0676728f56 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Deleted allocations for instance 08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2 [ 1035.403679] env[61957]: DEBUG oslo_service.periodic_task [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61957) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1035.404656] env[61957]: DEBUG nova.compute.manager [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Starting heal instance info cache {{(pid=61957) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 1035.404656] env[61957]: DEBUG nova.compute.manager [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Rebuilding the list of instances to heal {{(pid=61957) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9930}} [ 1035.568614] env[61957]: DEBUG oslo_vmware.api [None req-3ec44fbf-1f22-4ad1-8b26-d0253bbe3176 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Task: {'id': task-1278015, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.031148} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.568894] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-3ec44fbf-1f22-4ad1-8b26-d0253bbe3176 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Deleted the datastore file {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1035.569094] env[61957]: DEBUG oslo_concurrency.lockutils [None req-3ec44fbf-1f22-4ad1-8b26-d0253bbe3176 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Releasing lock "[datastore2] devstack-image-cache_base/f04c1641-6ff1-4b2c-9748-9cf176307db0/f04c1641-6ff1-4b2c-9748-9cf176307db0.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1035.569340] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-3ec44fbf-1f22-4ad1-8b26-d0253bbe3176 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/f04c1641-6ff1-4b2c-9748-9cf176307db0/f04c1641-6ff1-4b2c-9748-9cf176307db0.vmdk to [datastore2] 9b0053fe-420e-4b92-86a8-ed44f6a6ec49/9b0053fe-420e-4b92-86a8-ed44f6a6ec49.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1035.569634] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-87380d52-2811-47a7-b7fe-5b0493e1fa0a {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.575476] env[61957]: DEBUG oslo_vmware.api [None req-3ec44fbf-1f22-4ad1-8b26-d0253bbe3176 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Waiting for the task: (returnval){ [ 1035.575476] env[61957]: value = "task-1278018" [ 1035.575476] env[61957]: _type = "Task" [ 1035.575476] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.582663] env[61957]: DEBUG oslo_vmware.api [None req-3ec44fbf-1f22-4ad1-8b26-d0253bbe3176 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Task: {'id': task-1278018, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.738649] env[61957]: DEBUG nova.compute.manager [None req-4a6d948f-b9a8-4fc6-bd2c-f6dd4bec3479 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: a7419b30-1539-4977-a5d9-93a49d84914a] Starting instance... {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1035.869108] env[61957]: DEBUG oslo_vmware.api [None req-5a8aaf0c-6884-47c6-aa4a-9ccf495b4481 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': task-1278017, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.142572} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.869636] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-5a8aaf0c-6884-47c6-aa4a-9ccf495b4481 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Deleted the datastore file {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1035.869914] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-5a8aaf0c-6884-47c6-aa4a-9ccf495b4481 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: 6509707e-afc6-4c4a-88f9-a471865d510c] Deleted contents of the VM from datastore datastore1 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1035.870167] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-5a8aaf0c-6884-47c6-aa4a-9ccf495b4481 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: 6509707e-afc6-4c4a-88f9-a471865d510c] Instance destroyed {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1035.870414] env[61957]: INFO nova.compute.manager [None req-5a8aaf0c-6884-47c6-aa4a-9ccf495b4481 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: 6509707e-afc6-4c4a-88f9-a471865d510c] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1035.870734] env[61957]: DEBUG oslo.service.loopingcall [None req-5a8aaf0c-6884-47c6-aa4a-9ccf495b4481 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1035.870996] env[61957]: DEBUG nova.compute.manager [-] [instance: 6509707e-afc6-4c4a-88f9-a471865d510c] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1035.871127] env[61957]: DEBUG nova.network.neutron [-] [instance: 6509707e-afc6-4c4a-88f9-a471865d510c] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1035.876070] env[61957]: DEBUG oslo_concurrency.lockutils [None req-507275cf-57ab-4489-9d8f-aa0676728f56 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Lock "08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.308s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1035.876801] env[61957]: DEBUG oslo_concurrency.lockutils [req-d0775d92-7d76-40e7-9688-796e165eac56 req-333d4f0f-cd6b-419c-acdd-56111ba4d89f service nova] Acquired lock "08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1035.877547] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2608f711-dd01-410e-bd73-4044bd7ee722 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.888301] env[61957]: WARNING suds.client [-] Web service reported a SOAP processing fault using an unexpected HTTP status code 200. Reporting as an internal server error. [ 1035.888491] env[61957]: DEBUG oslo_vmware.api [-] Fault list: [ManagedObjectNotFound] {{(pid=61957) _invoke_api /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:337}} [ 1035.888929] env[61957]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ca5ff4f4-5465-4799-8d92-f218fe9241e4 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.897931] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcd6aadc-f078-439f-bab3-b9c94220ae22 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.913494] env[61957]: DEBUG nova.compute.manager [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] [instance: 6509707e-afc6-4c4a-88f9-a471865d510c] Skipping network cache update for instance because it is being deleted. {{(pid=61957) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9943}} [ 1035.935493] env[61957]: ERROR root [req-d0775d92-7d76-40e7-9688-796e165eac56 req-333d4f0f-cd6b-419c-acdd-56111ba4d89f service nova] Original exception being dropped: ['Traceback (most recent call last):\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py", line 377, in request_handler\n response = request(managed_object, **kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 586, in __call__\n return client.invoke(args, kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 728, in invoke\n result = self.send(soapenv, timeout=timeout)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 777, in send\n return self.process_reply(reply.message, None, None)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 840, in process_reply\n raise WebFault(fault, replyroot)\n', "suds.WebFault: Server raised fault: 'The object 'vim.VirtualMachine:vm-274571' has already been deleted or has not been completely created'\n", '\nDuring handling of the above exception, another exception occurred:\n\n', 'Traceback (most recent call last):\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 301, in _invoke_api\n return api_method(*args, **kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/vim_util.py", line 480, in get_object_property\n props = get_object_properties(vim, moref, [property_name],\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/vim_util.py", line 360, in get_object_properties\n retrieve_result = vim.RetrievePropertiesEx(\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py", line 413, in request_handler\n raise exceptions.VimFaultException(fault_list, fault_string,\n', "oslo_vmware.exceptions.VimFaultException: The object 'vim.VirtualMachine:vm-274571' has already been deleted or has not been completely created\nCause: Server raised fault: 'The object 'vim.VirtualMachine:vm-274571' has already been deleted or has not been completely created'\nFaults: [ManagedObjectNotFound]\nDetails: {'obj': 'vm-274571'}\n", '\nDuring handling of the above exception, another exception occurred:\n\n', 'Traceback (most recent call last):\n', ' File "/opt/stack/nova/nova/virt/vmwareapi/session.py", line 123, in _call_method\n return self.invoke_api(module, method, self.vim, *args,\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 358, in invoke_api\n return _invoke_api(module, method, *args, **kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 122, in func\n return evt.wait()\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait\n result = hub.switch()\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch\n return self.greenlet.switch()\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/common/loopingcall.py", line 122, in _inner\n idle = self.f(*self.args, **self.kw)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 96, in _func\n result = f(*args, **kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 341, in _invoke_api\n raise clazz(str(excep),\n', "oslo_vmware.exceptions.ManagedObjectNotFoundException: The object 'vim.VirtualMachine:vm-274571' has already been deleted or has not been completely created\nCause: Server raised fault: 'The object 'vim.VirtualMachine:vm-274571' has already been deleted or has not been completely created'\nFaults: [ManagedObjectNotFound]\nDetails: {'obj': 'vm-274571'}\n"]: nova.exception.InstanceNotFound: Instance 08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2 could not be found. [ 1035.935725] env[61957]: DEBUG oslo_concurrency.lockutils [req-d0775d92-7d76-40e7-9688-796e165eac56 req-333d4f0f-cd6b-419c-acdd-56111ba4d89f service nova] Releasing lock "08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1035.935960] env[61957]: DEBUG nova.compute.manager [req-d0775d92-7d76-40e7-9688-796e165eac56 req-333d4f0f-cd6b-419c-acdd-56111ba4d89f service nova] [instance: 08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2] Detach interface failed, port_id=00c8220d-5633-4de7-9795-8a6a8543c4ec, reason: Instance 08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2 could not be found. {{(pid=61957) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1035.936392] env[61957]: DEBUG oslo_concurrency.lockutils [req-f87cfb1f-b2a2-4db6-880c-a3fa7f8c0df7 req-da22b3c4-4128-486f-9423-a8115e382e2f service nova] Acquired lock "08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1035.937247] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8b82890-0c8a-4314-b020-b83412350143 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.946895] env[61957]: WARNING suds.client [-] Web service reported a SOAP processing fault using an unexpected HTTP status code 200. Reporting as an internal server error. [ 1035.947085] env[61957]: DEBUG oslo_vmware.api [-] Fault list: [ManagedObjectNotFound] {{(pid=61957) _invoke_api /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:337}} [ 1035.947499] env[61957]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4f8d76c9-22fa-42ed-82b9-7b2ff360eb23 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.960668] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df390e0a-fa71-4970-a6c6-d819dd9388f0 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.971818] env[61957]: DEBUG oslo_concurrency.lockutils [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Acquiring lock "refresh_cache-1a0327c2-5671-4970-9db7-c7cc912d8678" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1035.971964] env[61957]: DEBUG oslo_concurrency.lockutils [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Acquired lock "refresh_cache-1a0327c2-5671-4970-9db7-c7cc912d8678" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1035.972124] env[61957]: DEBUG nova.network.neutron [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] [instance: 1a0327c2-5671-4970-9db7-c7cc912d8678] Forcefully refreshing network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 1035.972319] env[61957]: DEBUG nova.objects.instance [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Lazy-loading 'info_cache' on Instance uuid 1a0327c2-5671-4970-9db7-c7cc912d8678 {{(pid=61957) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1035.991509] env[61957]: ERROR root [req-f87cfb1f-b2a2-4db6-880c-a3fa7f8c0df7 req-da22b3c4-4128-486f-9423-a8115e382e2f service nova] Original exception being dropped: ['Traceback (most recent call last):\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py", line 377, in request_handler\n response = request(managed_object, **kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 586, in __call__\n return client.invoke(args, kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 728, in invoke\n result = self.send(soapenv, timeout=timeout)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 777, in send\n return self.process_reply(reply.message, None, None)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 840, in process_reply\n raise WebFault(fault, replyroot)\n', "suds.WebFault: Server raised fault: 'The object 'vim.VirtualMachine:vm-274571' has already been deleted or has not been completely created'\n", '\nDuring handling of the above exception, another exception occurred:\n\n', 'Traceback (most recent call last):\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 301, in _invoke_api\n return api_method(*args, **kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/vim_util.py", line 480, in get_object_property\n props = get_object_properties(vim, moref, [property_name],\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/vim_util.py", line 360, in get_object_properties\n retrieve_result = vim.RetrievePropertiesEx(\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py", line 413, in request_handler\n raise exceptions.VimFaultException(fault_list, fault_string,\n', "oslo_vmware.exceptions.VimFaultException: The object 'vim.VirtualMachine:vm-274571' has already been deleted or has not been completely created\nCause: Server raised fault: 'The object 'vim.VirtualMachine:vm-274571' has already been deleted or has not been completely created'\nFaults: [ManagedObjectNotFound]\nDetails: {'obj': 'vm-274571'}\n", '\nDuring handling of the above exception, another exception occurred:\n\n', 'Traceback (most recent call last):\n', ' File "/opt/stack/nova/nova/virt/vmwareapi/session.py", line 123, in _call_method\n return self.invoke_api(module, method, self.vim, *args,\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 358, in invoke_api\n return _invoke_api(module, method, *args, **kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 122, in func\n return evt.wait()\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait\n result = hub.switch()\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch\n return self.greenlet.switch()\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/common/loopingcall.py", line 122, in _inner\n idle = self.f(*self.args, **self.kw)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 96, in _func\n result = f(*args, **kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 341, in _invoke_api\n raise clazz(str(excep),\n', "oslo_vmware.exceptions.ManagedObjectNotFoundException: The object 'vim.VirtualMachine:vm-274571' has already been deleted or has not been completely created\nCause: Server raised fault: 'The object 'vim.VirtualMachine:vm-274571' has already been deleted or has not been completely created'\nFaults: [ManagedObjectNotFound]\nDetails: {'obj': 'vm-274571'}\n"]: nova.exception.InstanceNotFound: Instance 08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2 could not be found. [ 1035.991742] env[61957]: DEBUG oslo_concurrency.lockutils [req-f87cfb1f-b2a2-4db6-880c-a3fa7f8c0df7 req-da22b3c4-4128-486f-9423-a8115e382e2f service nova] Releasing lock "08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1035.991965] env[61957]: DEBUG nova.compute.manager [req-f87cfb1f-b2a2-4db6-880c-a3fa7f8c0df7 req-da22b3c4-4128-486f-9423-a8115e382e2f service nova] [instance: 08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2] Detach interface failed, port_id=6a70445e-1266-453e-a6ff-50e8cbdcfd08, reason: Instance 08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2 could not be found. {{(pid=61957) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1035.992171] env[61957]: DEBUG nova.compute.manager [req-f87cfb1f-b2a2-4db6-880c-a3fa7f8c0df7 req-da22b3c4-4128-486f-9423-a8115e382e2f service nova] [instance: 9b0053fe-420e-4b92-86a8-ed44f6a6ec49] Received event network-vif-plugged-219230f0-0b14-4be9-81bb-67a765d8ced9 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1035.992401] env[61957]: DEBUG oslo_concurrency.lockutils [req-f87cfb1f-b2a2-4db6-880c-a3fa7f8c0df7 req-da22b3c4-4128-486f-9423-a8115e382e2f service nova] Acquiring lock "9b0053fe-420e-4b92-86a8-ed44f6a6ec49-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1035.992615] env[61957]: DEBUG oslo_concurrency.lockutils [req-f87cfb1f-b2a2-4db6-880c-a3fa7f8c0df7 req-da22b3c4-4128-486f-9423-a8115e382e2f service nova] Lock "9b0053fe-420e-4b92-86a8-ed44f6a6ec49-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1035.992779] env[61957]: DEBUG oslo_concurrency.lockutils [req-f87cfb1f-b2a2-4db6-880c-a3fa7f8c0df7 req-da22b3c4-4128-486f-9423-a8115e382e2f service nova] Lock "9b0053fe-420e-4b92-86a8-ed44f6a6ec49-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1035.992949] env[61957]: DEBUG nova.compute.manager [req-f87cfb1f-b2a2-4db6-880c-a3fa7f8c0df7 req-da22b3c4-4128-486f-9423-a8115e382e2f service nova] [instance: 9b0053fe-420e-4b92-86a8-ed44f6a6ec49] No waiting events found dispatching network-vif-plugged-219230f0-0b14-4be9-81bb-67a765d8ced9 {{(pid=61957) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1035.993140] env[61957]: WARNING nova.compute.manager [req-f87cfb1f-b2a2-4db6-880c-a3fa7f8c0df7 req-da22b3c4-4128-486f-9423-a8115e382e2f service nova] [instance: 9b0053fe-420e-4b92-86a8-ed44f6a6ec49] Received unexpected event network-vif-plugged-219230f0-0b14-4be9-81bb-67a765d8ced9 for instance with vm_state shelved_offloaded and task_state spawning. [ 1035.993364] env[61957]: DEBUG nova.compute.manager [req-f87cfb1f-b2a2-4db6-880c-a3fa7f8c0df7 req-da22b3c4-4128-486f-9423-a8115e382e2f service nova] [instance: 9b0053fe-420e-4b92-86a8-ed44f6a6ec49] Received event network-changed-219230f0-0b14-4be9-81bb-67a765d8ced9 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1035.993543] env[61957]: DEBUG nova.compute.manager [req-f87cfb1f-b2a2-4db6-880c-a3fa7f8c0df7 req-da22b3c4-4128-486f-9423-a8115e382e2f service nova] [instance: 9b0053fe-420e-4b92-86a8-ed44f6a6ec49] Refreshing instance network info cache due to event network-changed-219230f0-0b14-4be9-81bb-67a765d8ced9. {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1035.993733] env[61957]: DEBUG oslo_concurrency.lockutils [req-f87cfb1f-b2a2-4db6-880c-a3fa7f8c0df7 req-da22b3c4-4128-486f-9423-a8115e382e2f service nova] Acquiring lock "refresh_cache-9b0053fe-420e-4b92-86a8-ed44f6a6ec49" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1035.993874] env[61957]: DEBUG oslo_concurrency.lockutils [req-f87cfb1f-b2a2-4db6-880c-a3fa7f8c0df7 req-da22b3c4-4128-486f-9423-a8115e382e2f service nova] Acquired lock "refresh_cache-9b0053fe-420e-4b92-86a8-ed44f6a6ec49" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1035.994048] env[61957]: DEBUG nova.network.neutron [req-f87cfb1f-b2a2-4db6-880c-a3fa7f8c0df7 req-da22b3c4-4128-486f-9423-a8115e382e2f service nova] [instance: 9b0053fe-420e-4b92-86a8-ed44f6a6ec49] Refreshing network info cache for port 219230f0-0b14-4be9-81bb-67a765d8ced9 {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1036.059235] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f2240eb-bf0d-41a1-a80d-1bc8d74469da {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.067304] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4cb53e2-5862-46aa-b47b-b7a8647ac5f9 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.100714] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de457649-70cc-4867-875f-0cef4d52a866 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.108375] env[61957]: DEBUG oslo_vmware.api [None req-3ec44fbf-1f22-4ad1-8b26-d0253bbe3176 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Task: {'id': task-1278018, 'name': CopyVirtualDisk_Task} progress is 21%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.111718] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a068c9e1-41d5-422f-8802-7e88d354269b {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.127080] env[61957]: DEBUG nova.compute.provider_tree [None req-fa4af5a4-f456-4969-8ee2-40fce4f1cd43 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1036.204103] env[61957]: DEBUG nova.compute.manager [req-3bc8d35d-cf9f-451a-bbab-3e57c1825e05 req-63095e77-a153-4134-a967-654d67b222c7 service nova] [instance: 6509707e-afc6-4c4a-88f9-a471865d510c] Received event network-vif-deleted-7df2736b-10d0-4f9d-83aa-e2bbdf1ef45b {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1036.204436] env[61957]: INFO nova.compute.manager [req-3bc8d35d-cf9f-451a-bbab-3e57c1825e05 req-63095e77-a153-4134-a967-654d67b222c7 service nova] [instance: 6509707e-afc6-4c4a-88f9-a471865d510c] Neutron deleted interface 7df2736b-10d0-4f9d-83aa-e2bbdf1ef45b; detaching it from the instance and deleting it from the info cache [ 1036.204506] env[61957]: DEBUG nova.network.neutron [req-3bc8d35d-cf9f-451a-bbab-3e57c1825e05 req-63095e77-a153-4134-a967-654d67b222c7 service nova] [instance: 6509707e-afc6-4c4a-88f9-a471865d510c] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1036.260894] env[61957]: DEBUG oslo_concurrency.lockutils [None req-4a6d948f-b9a8-4fc6-bd2c-f6dd4bec3479 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1036.607433] env[61957]: DEBUG oslo_vmware.api [None req-3ec44fbf-1f22-4ad1-8b26-d0253bbe3176 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Task: {'id': task-1278018, 'name': CopyVirtualDisk_Task} progress is 43%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.630493] env[61957]: DEBUG nova.scheduler.client.report [None req-fa4af5a4-f456-4969-8ee2-40fce4f1cd43 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1036.682275] env[61957]: DEBUG nova.network.neutron [-] [instance: 6509707e-afc6-4c4a-88f9-a471865d510c] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1036.707384] env[61957]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-15f0e95d-968d-46f3-8b48-35af0229163a {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.717507] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c983f3b1-bd11-4845-87d5-e0e784f2190d {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.745312] env[61957]: DEBUG nova.compute.manager [req-3bc8d35d-cf9f-451a-bbab-3e57c1825e05 req-63095e77-a153-4134-a967-654d67b222c7 service nova] [instance: 6509707e-afc6-4c4a-88f9-a471865d510c] Detach interface failed, port_id=7df2736b-10d0-4f9d-83aa-e2bbdf1ef45b, reason: Instance 6509707e-afc6-4c4a-88f9-a471865d510c could not be found. {{(pid=61957) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1036.867808] env[61957]: DEBUG nova.network.neutron [req-f87cfb1f-b2a2-4db6-880c-a3fa7f8c0df7 req-da22b3c4-4128-486f-9423-a8115e382e2f service nova] [instance: 9b0053fe-420e-4b92-86a8-ed44f6a6ec49] Updated VIF entry in instance network info cache for port 219230f0-0b14-4be9-81bb-67a765d8ced9. {{(pid=61957) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1036.867966] env[61957]: DEBUG nova.network.neutron [req-f87cfb1f-b2a2-4db6-880c-a3fa7f8c0df7 req-da22b3c4-4128-486f-9423-a8115e382e2f service nova] [instance: 9b0053fe-420e-4b92-86a8-ed44f6a6ec49] Updating instance_info_cache with network_info: [{"id": "219230f0-0b14-4be9-81bb-67a765d8ced9", "address": "fa:16:3e:f6:9e:ce", "network": {"id": "cf378e9c-eaae-4168-bd60-9818fcfaf405", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1243326371-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f7c20ab3c822423fa837a03f066e4239", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8cb478a6-872c-4a90-a8db-526b374e82ce", "external-id": "nsx-vlan-transportzone-835", "segmentation_id": 835, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap219230f0-0b", "ovs_interfaceid": "219230f0-0b14-4be9-81bb-67a765d8ced9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1037.108025] env[61957]: DEBUG oslo_vmware.api [None req-3ec44fbf-1f22-4ad1-8b26-d0253bbe3176 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Task: {'id': task-1278018, 'name': CopyVirtualDisk_Task} progress is 66%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.135406] env[61957]: DEBUG oslo_concurrency.lockutils [None req-fa4af5a4-f456-4969-8ee2-40fce4f1cd43 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.784s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1037.137855] env[61957]: DEBUG oslo_concurrency.lockutils [None req-de4650d3-4738-443f-8147-36e90c68fb19 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.001s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1037.139717] env[61957]: INFO nova.compute.claims [None req-de4650d3-4738-443f-8147-36e90c68fb19 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] [instance: 039a6b01-a745-478c-9ff7-c8c37afe2ce4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1037.167640] env[61957]: INFO nova.scheduler.client.report [None req-fa4af5a4-f456-4969-8ee2-40fce4f1cd43 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Deleted allocations for instance 3f144012-c0cc-43ce-8c73-4422fe8c1fe6 [ 1037.185202] env[61957]: INFO nova.compute.manager [-] [instance: 6509707e-afc6-4c4a-88f9-a471865d510c] Took 1.31 seconds to deallocate network for instance. [ 1037.371273] env[61957]: DEBUG oslo_concurrency.lockutils [req-f87cfb1f-b2a2-4db6-880c-a3fa7f8c0df7 req-da22b3c4-4128-486f-9423-a8115e382e2f service nova] Releasing lock "refresh_cache-9b0053fe-420e-4b92-86a8-ed44f6a6ec49" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1037.404143] env[61957]: DEBUG oslo_concurrency.lockutils [None req-6c409e2b-c7c8-4a9a-8e7d-1343e1ae2437 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Acquiring lock "736eabcf-70f8-4e1e-9830-270d9971fc9a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1037.404443] env[61957]: DEBUG oslo_concurrency.lockutils [None req-6c409e2b-c7c8-4a9a-8e7d-1343e1ae2437 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Lock "736eabcf-70f8-4e1e-9830-270d9971fc9a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1037.404684] env[61957]: DEBUG oslo_concurrency.lockutils [None req-6c409e2b-c7c8-4a9a-8e7d-1343e1ae2437 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Acquiring lock "736eabcf-70f8-4e1e-9830-270d9971fc9a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1037.404875] env[61957]: DEBUG oslo_concurrency.lockutils [None req-6c409e2b-c7c8-4a9a-8e7d-1343e1ae2437 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Lock "736eabcf-70f8-4e1e-9830-270d9971fc9a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1037.405060] env[61957]: DEBUG oslo_concurrency.lockutils [None req-6c409e2b-c7c8-4a9a-8e7d-1343e1ae2437 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Lock "736eabcf-70f8-4e1e-9830-270d9971fc9a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1037.407383] env[61957]: INFO nova.compute.manager [None req-6c409e2b-c7c8-4a9a-8e7d-1343e1ae2437 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] [instance: 736eabcf-70f8-4e1e-9830-270d9971fc9a] Terminating instance [ 1037.409743] env[61957]: DEBUG nova.compute.manager [None req-6c409e2b-c7c8-4a9a-8e7d-1343e1ae2437 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] [instance: 736eabcf-70f8-4e1e-9830-270d9971fc9a] Start destroying the instance on the hypervisor. {{(pid=61957) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1037.409981] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-6c409e2b-c7c8-4a9a-8e7d-1343e1ae2437 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] [instance: 736eabcf-70f8-4e1e-9830-270d9971fc9a] Destroying instance {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1037.410913] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcdaa734-884f-4539-9ae0-473a08de20de {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.420281] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c409e2b-c7c8-4a9a-8e7d-1343e1ae2437 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] [instance: 736eabcf-70f8-4e1e-9830-270d9971fc9a] Powering off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1037.420583] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8291cd2b-b3e2-4ccf-b4b2-0b28274a539f {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.427634] env[61957]: DEBUG oslo_vmware.api [None req-6c409e2b-c7c8-4a9a-8e7d-1343e1ae2437 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Waiting for the task: (returnval){ [ 1037.427634] env[61957]: value = "task-1278019" [ 1037.427634] env[61957]: _type = "Task" [ 1037.427634] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.435970] env[61957]: DEBUG oslo_vmware.api [None req-6c409e2b-c7c8-4a9a-8e7d-1343e1ae2437 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Task: {'id': task-1278019, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.609481] env[61957]: DEBUG oslo_vmware.api [None req-3ec44fbf-1f22-4ad1-8b26-d0253bbe3176 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Task: {'id': task-1278018, 'name': CopyVirtualDisk_Task} progress is 88%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.675921] env[61957]: DEBUG oslo_concurrency.lockutils [None req-fa4af5a4-f456-4969-8ee2-40fce4f1cd43 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Lock "3f144012-c0cc-43ce-8c73-4422fe8c1fe6" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.727s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1037.692847] env[61957]: DEBUG oslo_concurrency.lockutils [None req-5a8aaf0c-6884-47c6-aa4a-9ccf495b4481 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1037.720403] env[61957]: DEBUG nova.network.neutron [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] [instance: 1a0327c2-5671-4970-9db7-c7cc912d8678] Updating instance_info_cache with network_info: [{"id": "e84fe58e-f595-4b03-9e8a-88238d003168", "address": "fa:16:3e:7e:5a:a2", "network": {"id": "4f0e57be-8906-48b5-865f-59ff12250a8b", "bridge": "br-int", "label": "tempest-ServersTestJSON-1916182687-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "65fd71bcd1bf41238a9cc3a5d6dd4924", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "939c05b6-8f31-4f3a-95ac-6297e0bd243e", "external-id": "nsx-vlan-transportzone-825", "segmentation_id": 825, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape84fe58e-f5", "ovs_interfaceid": "e84fe58e-f595-4b03-9e8a-88238d003168", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1037.937688] env[61957]: DEBUG oslo_vmware.api [None req-6c409e2b-c7c8-4a9a-8e7d-1343e1ae2437 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Task: {'id': task-1278019, 'name': PowerOffVM_Task} progress is 100%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.107815] env[61957]: DEBUG oslo_vmware.api [None req-3ec44fbf-1f22-4ad1-8b26-d0253bbe3176 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Task: {'id': task-1278018, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.289027} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.108102] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-3ec44fbf-1f22-4ad1-8b26-d0253bbe3176 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/f04c1641-6ff1-4b2c-9748-9cf176307db0/f04c1641-6ff1-4b2c-9748-9cf176307db0.vmdk to [datastore2] 9b0053fe-420e-4b92-86a8-ed44f6a6ec49/9b0053fe-420e-4b92-86a8-ed44f6a6ec49.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1038.108870] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41311402-a1e0-4157-9f70-e732c04a35e4 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.130791] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-3ec44fbf-1f22-4ad1-8b26-d0253bbe3176 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] [instance: 9b0053fe-420e-4b92-86a8-ed44f6a6ec49] Reconfiguring VM instance instance-00000051 to attach disk [datastore2] 9b0053fe-420e-4b92-86a8-ed44f6a6ec49/9b0053fe-420e-4b92-86a8-ed44f6a6ec49.vmdk or device None with type streamOptimized {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1038.131230] env[61957]: DEBUG oslo_concurrency.lockutils [None req-b6b5029a-df23-45ff-9311-ae897f587f74 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Acquiring lock "84cc9489-2a11-4510-8ad6-13dc41814856" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1038.131469] env[61957]: DEBUG oslo_concurrency.lockutils [None req-b6b5029a-df23-45ff-9311-ae897f587f74 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Lock "84cc9489-2a11-4510-8ad6-13dc41814856" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1038.131670] env[61957]: DEBUG oslo_concurrency.lockutils [None req-b6b5029a-df23-45ff-9311-ae897f587f74 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Acquiring lock "84cc9489-2a11-4510-8ad6-13dc41814856-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1038.131851] env[61957]: DEBUG oslo_concurrency.lockutils [None req-b6b5029a-df23-45ff-9311-ae897f587f74 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Lock "84cc9489-2a11-4510-8ad6-13dc41814856-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1038.132031] env[61957]: DEBUG oslo_concurrency.lockutils [None req-b6b5029a-df23-45ff-9311-ae897f587f74 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Lock "84cc9489-2a11-4510-8ad6-13dc41814856-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1038.133633] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f63d5320-201a-4dc4-b2c6-5b9e20a5cf5d {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.148988] env[61957]: INFO nova.compute.manager [None req-b6b5029a-df23-45ff-9311-ae897f587f74 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 84cc9489-2a11-4510-8ad6-13dc41814856] Terminating instance [ 1038.151885] env[61957]: DEBUG nova.compute.manager [None req-b6b5029a-df23-45ff-9311-ae897f587f74 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 84cc9489-2a11-4510-8ad6-13dc41814856] Start destroying the instance on the hypervisor. {{(pid=61957) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1038.151885] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-b6b5029a-df23-45ff-9311-ae897f587f74 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 84cc9489-2a11-4510-8ad6-13dc41814856] Destroying instance {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1038.155192] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-007f26bd-7c83-44e3-b27d-ed525e36fac9 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.159721] env[61957]: DEBUG oslo_vmware.api [None req-3ec44fbf-1f22-4ad1-8b26-d0253bbe3176 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Waiting for the task: (returnval){ [ 1038.159721] env[61957]: value = "task-1278020" [ 1038.159721] env[61957]: _type = "Task" [ 1038.159721] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1038.166381] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-b6b5029a-df23-45ff-9311-ae897f587f74 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 84cc9489-2a11-4510-8ad6-13dc41814856] Powering off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1038.166902] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1d4fd1ad-1b39-41c2-9374-5d4a45b8822a {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.172146] env[61957]: DEBUG oslo_vmware.api [None req-3ec44fbf-1f22-4ad1-8b26-d0253bbe3176 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Task: {'id': task-1278020, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.174320] env[61957]: DEBUG oslo_vmware.api [None req-b6b5029a-df23-45ff-9311-ae897f587f74 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Waiting for the task: (returnval){ [ 1038.174320] env[61957]: value = "task-1278021" [ 1038.174320] env[61957]: _type = "Task" [ 1038.174320] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1038.185339] env[61957]: DEBUG oslo_vmware.api [None req-b6b5029a-df23-45ff-9311-ae897f587f74 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': task-1278021, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.223080] env[61957]: DEBUG oslo_concurrency.lockutils [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Releasing lock "refresh_cache-1a0327c2-5671-4970-9db7-c7cc912d8678" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1038.223335] env[61957]: DEBUG nova.compute.manager [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] [instance: 1a0327c2-5671-4970-9db7-c7cc912d8678] Updated the network info_cache for instance {{(pid=61957) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9997}} [ 1038.223573] env[61957]: DEBUG oslo_service.periodic_task [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61957) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1038.223750] env[61957]: DEBUG oslo_service.periodic_task [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61957) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1038.223953] env[61957]: DEBUG oslo_service.periodic_task [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61957) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1038.224139] env[61957]: DEBUG oslo_service.periodic_task [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61957) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1038.224327] env[61957]: DEBUG oslo_service.periodic_task [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61957) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1038.225036] env[61957]: DEBUG oslo_service.periodic_task [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61957) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1038.225036] env[61957]: DEBUG nova.compute.manager [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61957) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 1038.225036] env[61957]: DEBUG oslo_service.periodic_task [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61957) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1038.283771] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d10a62ad-e6bd-480b-9341-39c337795952 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.292496] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c1bdc2a-d10e-44ae-87cb-0dfa2880d18d {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.323613] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-912711b9-2964-4945-baaa-7df189017573 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.331366] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c08e51b1-8559-4c0a-959d-b1fd8b169c33 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.345051] env[61957]: DEBUG nova.compute.provider_tree [None req-de4650d3-4738-443f-8147-36e90c68fb19 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1038.439612] env[61957]: DEBUG oslo_vmware.api [None req-6c409e2b-c7c8-4a9a-8e7d-1343e1ae2437 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Task: {'id': task-1278019, 'name': PowerOffVM_Task, 'duration_secs': 0.573481} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.439886] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c409e2b-c7c8-4a9a-8e7d-1343e1ae2437 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] [instance: 736eabcf-70f8-4e1e-9830-270d9971fc9a] Powered off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1038.440078] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-6c409e2b-c7c8-4a9a-8e7d-1343e1ae2437 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] [instance: 736eabcf-70f8-4e1e-9830-270d9971fc9a] Unregistering the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1038.440326] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c84e30e8-8679-498f-af8b-359ec0b4e221 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.534404] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-6c409e2b-c7c8-4a9a-8e7d-1343e1ae2437 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] [instance: 736eabcf-70f8-4e1e-9830-270d9971fc9a] Unregistered the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1038.534404] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-6c409e2b-c7c8-4a9a-8e7d-1343e1ae2437 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] [instance: 736eabcf-70f8-4e1e-9830-270d9971fc9a] Deleting contents of the VM from datastore datastore2 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1038.534814] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-6c409e2b-c7c8-4a9a-8e7d-1343e1ae2437 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Deleting the datastore file [datastore2] 736eabcf-70f8-4e1e-9830-270d9971fc9a {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1038.534879] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-13618339-f8f5-42ab-82b8-9d9c6117df37 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.541591] env[61957]: DEBUG oslo_vmware.api [None req-6c409e2b-c7c8-4a9a-8e7d-1343e1ae2437 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Waiting for the task: (returnval){ [ 1038.541591] env[61957]: value = "task-1278023" [ 1038.541591] env[61957]: _type = "Task" [ 1038.541591] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1038.549259] env[61957]: DEBUG oslo_vmware.api [None req-6c409e2b-c7c8-4a9a-8e7d-1343e1ae2437 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Task: {'id': task-1278023, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.669824] env[61957]: DEBUG oslo_vmware.api [None req-3ec44fbf-1f22-4ad1-8b26-d0253bbe3176 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Task: {'id': task-1278020, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.683266] env[61957]: DEBUG oslo_vmware.api [None req-b6b5029a-df23-45ff-9311-ae897f587f74 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': task-1278021, 'name': PowerOffVM_Task, 'duration_secs': 0.224336} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.683614] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-b6b5029a-df23-45ff-9311-ae897f587f74 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 84cc9489-2a11-4510-8ad6-13dc41814856] Powered off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1038.683826] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-b6b5029a-df23-45ff-9311-ae897f587f74 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 84cc9489-2a11-4510-8ad6-13dc41814856] Unregistering the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1038.684169] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b5391ffa-cc77-4926-94eb-8e7ad9c26b95 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.727908] env[61957]: DEBUG oslo_concurrency.lockutils [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1038.747363] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-b6b5029a-df23-45ff-9311-ae897f587f74 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 84cc9489-2a11-4510-8ad6-13dc41814856] Unregistered the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1038.747643] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-b6b5029a-df23-45ff-9311-ae897f587f74 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 84cc9489-2a11-4510-8ad6-13dc41814856] Deleting contents of the VM from datastore datastore1 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1038.747936] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-b6b5029a-df23-45ff-9311-ae897f587f74 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Deleting the datastore file [datastore1] 84cc9489-2a11-4510-8ad6-13dc41814856 {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1038.748143] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c4ed0c0a-d226-42ff-9757-c1a392ee47d9 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.754605] env[61957]: DEBUG oslo_vmware.api [None req-b6b5029a-df23-45ff-9311-ae897f587f74 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Waiting for the task: (returnval){ [ 1038.754605] env[61957]: value = "task-1278025" [ 1038.754605] env[61957]: _type = "Task" [ 1038.754605] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1038.764889] env[61957]: DEBUG oslo_vmware.api [None req-b6b5029a-df23-45ff-9311-ae897f587f74 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': task-1278025, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.790864] env[61957]: DEBUG oslo_concurrency.lockutils [None req-8a0a01e3-6392-4a5c-b585-5e26a0a02629 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Acquiring lock "4625d1dc-a621-4a33-9537-9e8cce96c914" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1038.791253] env[61957]: DEBUG oslo_concurrency.lockutils [None req-8a0a01e3-6392-4a5c-b585-5e26a0a02629 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Lock "4625d1dc-a621-4a33-9537-9e8cce96c914" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1038.848554] env[61957]: DEBUG nova.scheduler.client.report [None req-de4650d3-4738-443f-8147-36e90c68fb19 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1039.052184] env[61957]: DEBUG oslo_vmware.api [None req-6c409e2b-c7c8-4a9a-8e7d-1343e1ae2437 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Task: {'id': task-1278023, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.509007} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1039.052530] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-6c409e2b-c7c8-4a9a-8e7d-1343e1ae2437 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Deleted the datastore file {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1039.052728] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-6c409e2b-c7c8-4a9a-8e7d-1343e1ae2437 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] [instance: 736eabcf-70f8-4e1e-9830-270d9971fc9a] Deleted contents of the VM from datastore datastore2 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1039.052915] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-6c409e2b-c7c8-4a9a-8e7d-1343e1ae2437 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] [instance: 736eabcf-70f8-4e1e-9830-270d9971fc9a] Instance destroyed {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1039.053101] env[61957]: INFO nova.compute.manager [None req-6c409e2b-c7c8-4a9a-8e7d-1343e1ae2437 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] [instance: 736eabcf-70f8-4e1e-9830-270d9971fc9a] Took 1.64 seconds to destroy the instance on the hypervisor. [ 1039.053353] env[61957]: DEBUG oslo.service.loopingcall [None req-6c409e2b-c7c8-4a9a-8e7d-1343e1ae2437 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1039.053554] env[61957]: DEBUG nova.compute.manager [-] [instance: 736eabcf-70f8-4e1e-9830-270d9971fc9a] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1039.053649] env[61957]: DEBUG nova.network.neutron [-] [instance: 736eabcf-70f8-4e1e-9830-270d9971fc9a] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1039.169460] env[61957]: DEBUG oslo_vmware.api [None req-3ec44fbf-1f22-4ad1-8b26-d0253bbe3176 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Task: {'id': task-1278020, 'name': ReconfigVM_Task, 'duration_secs': 0.869678} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1039.169755] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-3ec44fbf-1f22-4ad1-8b26-d0253bbe3176 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] [instance: 9b0053fe-420e-4b92-86a8-ed44f6a6ec49] Reconfigured VM instance instance-00000051 to attach disk [datastore2] 9b0053fe-420e-4b92-86a8-ed44f6a6ec49/9b0053fe-420e-4b92-86a8-ed44f6a6ec49.vmdk or device None with type streamOptimized {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1039.170430] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6f9cd548-6617-43f5-83c2-fe55b8068f0a {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.176451] env[61957]: DEBUG oslo_vmware.api [None req-3ec44fbf-1f22-4ad1-8b26-d0253bbe3176 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Waiting for the task: (returnval){ [ 1039.176451] env[61957]: value = "task-1278026" [ 1039.176451] env[61957]: _type = "Task" [ 1039.176451] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1039.183930] env[61957]: DEBUG oslo_vmware.api [None req-3ec44fbf-1f22-4ad1-8b26-d0253bbe3176 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Task: {'id': task-1278026, 'name': Rename_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.264467] env[61957]: DEBUG oslo_vmware.api [None req-b6b5029a-df23-45ff-9311-ae897f587f74 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': task-1278025, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.378276} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1039.264816] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-b6b5029a-df23-45ff-9311-ae897f587f74 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Deleted the datastore file {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1039.265170] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-b6b5029a-df23-45ff-9311-ae897f587f74 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 84cc9489-2a11-4510-8ad6-13dc41814856] Deleted contents of the VM from datastore datastore1 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1039.265429] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-b6b5029a-df23-45ff-9311-ae897f587f74 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 84cc9489-2a11-4510-8ad6-13dc41814856] Instance destroyed {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1039.265673] env[61957]: INFO nova.compute.manager [None req-b6b5029a-df23-45ff-9311-ae897f587f74 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 84cc9489-2a11-4510-8ad6-13dc41814856] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1039.266008] env[61957]: DEBUG oslo.service.loopingcall [None req-b6b5029a-df23-45ff-9311-ae897f587f74 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1039.266298] env[61957]: DEBUG nova.compute.manager [-] [instance: 84cc9489-2a11-4510-8ad6-13dc41814856] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1039.266473] env[61957]: DEBUG nova.network.neutron [-] [instance: 84cc9489-2a11-4510-8ad6-13dc41814856] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1039.294068] env[61957]: DEBUG nova.compute.manager [None req-8a0a01e3-6392-4a5c-b585-5e26a0a02629 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 4625d1dc-a621-4a33-9537-9e8cce96c914] Starting instance... {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1039.353445] env[61957]: DEBUG oslo_concurrency.lockutils [None req-de4650d3-4738-443f-8147-36e90c68fb19 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.215s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1039.354045] env[61957]: DEBUG nova.compute.manager [None req-de4650d3-4738-443f-8147-36e90c68fb19 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] [instance: 039a6b01-a745-478c-9ff7-c8c37afe2ce4] Start building networks asynchronously for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1039.356913] env[61957]: DEBUG oslo_concurrency.lockutils [None req-0d170cdb-983c-43d4-9998-8676a1404b75 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.318s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1039.358422] env[61957]: INFO nova.compute.claims [None req-0d170cdb-983c-43d4-9998-8676a1404b75 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 42860944-c700-43be-b291-970803fc7955] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1039.687221] env[61957]: DEBUG oslo_vmware.api [None req-3ec44fbf-1f22-4ad1-8b26-d0253bbe3176 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Task: {'id': task-1278026, 'name': Rename_Task, 'duration_secs': 0.23681} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1039.687797] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-3ec44fbf-1f22-4ad1-8b26-d0253bbe3176 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] [instance: 9b0053fe-420e-4b92-86a8-ed44f6a6ec49] Powering on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1039.687797] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7db47d16-8485-4970-8226-eed20f34ac5d {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.697921] env[61957]: DEBUG oslo_vmware.api [None req-3ec44fbf-1f22-4ad1-8b26-d0253bbe3176 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Waiting for the task: (returnval){ [ 1039.697921] env[61957]: value = "task-1278027" [ 1039.697921] env[61957]: _type = "Task" [ 1039.697921] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1039.705821] env[61957]: DEBUG oslo_vmware.api [None req-3ec44fbf-1f22-4ad1-8b26-d0253bbe3176 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Task: {'id': task-1278027, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.826897] env[61957]: DEBUG oslo_concurrency.lockutils [None req-8a0a01e3-6392-4a5c-b585-5e26a0a02629 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1039.852634] env[61957]: DEBUG nova.compute.manager [req-b26a2b40-b69c-435a-80b3-38ba902b70df req-a851ad0a-13c9-4247-9e05-3572b70ad3a6 service nova] [instance: 84cc9489-2a11-4510-8ad6-13dc41814856] Received event network-vif-deleted-b602c88c-d9b2-44bd-ba43-0ba7d976d8ac {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1039.852838] env[61957]: INFO nova.compute.manager [req-b26a2b40-b69c-435a-80b3-38ba902b70df req-a851ad0a-13c9-4247-9e05-3572b70ad3a6 service nova] [instance: 84cc9489-2a11-4510-8ad6-13dc41814856] Neutron deleted interface b602c88c-d9b2-44bd-ba43-0ba7d976d8ac; detaching it from the instance and deleting it from the info cache [ 1039.853031] env[61957]: DEBUG nova.network.neutron [req-b26a2b40-b69c-435a-80b3-38ba902b70df req-a851ad0a-13c9-4247-9e05-3572b70ad3a6 service nova] [instance: 84cc9489-2a11-4510-8ad6-13dc41814856] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1039.862545] env[61957]: DEBUG nova.compute.utils [None req-de4650d3-4738-443f-8147-36e90c68fb19 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Using /dev/sd instead of None {{(pid=61957) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1039.866083] env[61957]: DEBUG nova.compute.manager [None req-de4650d3-4738-443f-8147-36e90c68fb19 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] [instance: 039a6b01-a745-478c-9ff7-c8c37afe2ce4] Allocating IP information in the background. {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1039.867146] env[61957]: DEBUG nova.network.neutron [None req-de4650d3-4738-443f-8147-36e90c68fb19 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] [instance: 039a6b01-a745-478c-9ff7-c8c37afe2ce4] allocate_for_instance() {{(pid=61957) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1039.904963] env[61957]: DEBUG nova.policy [None req-de4650d3-4738-443f-8147-36e90c68fb19 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '965923b6f23744a484629c752b0d11d1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ea955bd4d84548ff8e6a0c21542d6ee7', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61957) authorize /opt/stack/nova/nova/policy.py:203}} [ 1040.185900] env[61957]: DEBUG nova.network.neutron [None req-de4650d3-4738-443f-8147-36e90c68fb19 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] [instance: 039a6b01-a745-478c-9ff7-c8c37afe2ce4] Successfully created port: e49b8de4-47b9-400b-a0b0-9e0a0e68dc84 {{(pid=61957) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1040.208107] env[61957]: DEBUG oslo_vmware.api [None req-3ec44fbf-1f22-4ad1-8b26-d0253bbe3176 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Task: {'id': task-1278027, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.328745] env[61957]: DEBUG nova.network.neutron [-] [instance: 84cc9489-2a11-4510-8ad6-13dc41814856] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1040.340965] env[61957]: DEBUG nova.network.neutron [-] [instance: 736eabcf-70f8-4e1e-9830-270d9971fc9a] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1040.356296] env[61957]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c3a13da4-86b9-407f-8c30-a9b16df153d1 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.367273] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8c2c4c0-9268-4211-a2d9-1340c30d29a5 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.384178] env[61957]: DEBUG nova.compute.manager [None req-de4650d3-4738-443f-8147-36e90c68fb19 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] [instance: 039a6b01-a745-478c-9ff7-c8c37afe2ce4] Start building block device mappings for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1040.407778] env[61957]: DEBUG nova.compute.manager [req-b26a2b40-b69c-435a-80b3-38ba902b70df req-a851ad0a-13c9-4247-9e05-3572b70ad3a6 service nova] [instance: 84cc9489-2a11-4510-8ad6-13dc41814856] Detach interface failed, port_id=b602c88c-d9b2-44bd-ba43-0ba7d976d8ac, reason: Instance 84cc9489-2a11-4510-8ad6-13dc41814856 could not be found. {{(pid=61957) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1040.534524] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71f691bb-125d-4eac-92a3-89c7307d1d20 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.542029] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c69c5ab-95e3-4fa6-bd4b-b3be9156f21a {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.571189] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61df749d-06df-442c-9c9c-efc3aad6baf2 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.578457] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-704ffa52-50e5-4714-a425-f14a0572b86a {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.592960] env[61957]: DEBUG nova.compute.provider_tree [None req-0d170cdb-983c-43d4-9998-8676a1404b75 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1040.712557] env[61957]: DEBUG oslo_vmware.api [None req-3ec44fbf-1f22-4ad1-8b26-d0253bbe3176 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Task: {'id': task-1278027, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.837826] env[61957]: INFO nova.compute.manager [-] [instance: 84cc9489-2a11-4510-8ad6-13dc41814856] Took 1.57 seconds to deallocate network for instance. [ 1040.844360] env[61957]: INFO nova.compute.manager [-] [instance: 736eabcf-70f8-4e1e-9830-270d9971fc9a] Took 1.79 seconds to deallocate network for instance. [ 1041.096560] env[61957]: DEBUG nova.scheduler.client.report [None req-0d170cdb-983c-43d4-9998-8676a1404b75 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1041.209291] env[61957]: DEBUG oslo_vmware.api [None req-3ec44fbf-1f22-4ad1-8b26-d0253bbe3176 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Task: {'id': task-1278027, 'name': PowerOnVM_Task, 'duration_secs': 1.102379} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1041.209573] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-3ec44fbf-1f22-4ad1-8b26-d0253bbe3176 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] [instance: 9b0053fe-420e-4b92-86a8-ed44f6a6ec49] Powered on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1041.338013] env[61957]: DEBUG nova.compute.manager [None req-3ec44fbf-1f22-4ad1-8b26-d0253bbe3176 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] [instance: 9b0053fe-420e-4b92-86a8-ed44f6a6ec49] Checking state {{(pid=61957) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1041.339264] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c40a096e-0e7c-49e9-8c48-b9faa8e11069 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.343950] env[61957]: DEBUG oslo_concurrency.lockutils [None req-b6b5029a-df23-45ff-9311-ae897f587f74 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1041.350769] env[61957]: DEBUG oslo_concurrency.lockutils [None req-6c409e2b-c7c8-4a9a-8e7d-1343e1ae2437 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1041.398433] env[61957]: DEBUG nova.compute.manager [None req-de4650d3-4738-443f-8147-36e90c68fb19 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] [instance: 039a6b01-a745-478c-9ff7-c8c37afe2ce4] Start spawning the instance on the hypervisor. {{(pid=61957) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1041.419118] env[61957]: DEBUG nova.virt.hardware [None req-de4650d3-4738-443f-8147-36e90c68fb19 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T17:22:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T17:21:50Z,direct_url=,disk_format='vmdk',id=11c76a2c-f705-470a-ba9d-4657858bab38,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='81d3096d0f094373913fc3dc8f5c3c6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T17:21:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1041.419373] env[61957]: DEBUG nova.virt.hardware [None req-de4650d3-4738-443f-8147-36e90c68fb19 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Flavor limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1041.419581] env[61957]: DEBUG nova.virt.hardware [None req-de4650d3-4738-443f-8147-36e90c68fb19 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Image limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1041.419783] env[61957]: DEBUG nova.virt.hardware [None req-de4650d3-4738-443f-8147-36e90c68fb19 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Flavor pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1041.419935] env[61957]: DEBUG nova.virt.hardware [None req-de4650d3-4738-443f-8147-36e90c68fb19 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Image pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1041.420104] env[61957]: DEBUG nova.virt.hardware [None req-de4650d3-4738-443f-8147-36e90c68fb19 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1041.420393] env[61957]: DEBUG nova.virt.hardware [None req-de4650d3-4738-443f-8147-36e90c68fb19 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1041.420585] env[61957]: DEBUG nova.virt.hardware [None req-de4650d3-4738-443f-8147-36e90c68fb19 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1041.420760] env[61957]: DEBUG nova.virt.hardware [None req-de4650d3-4738-443f-8147-36e90c68fb19 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Got 1 possible topologies {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1041.420926] env[61957]: DEBUG nova.virt.hardware [None req-de4650d3-4738-443f-8147-36e90c68fb19 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1041.421114] env[61957]: DEBUG nova.virt.hardware [None req-de4650d3-4738-443f-8147-36e90c68fb19 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1041.422304] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59b950eb-8fb4-4b4e-b569-1792f26ea683 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.431174] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-574c2319-8d9b-4a16-85ee-cbd969753924 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.601800] env[61957]: DEBUG oslo_concurrency.lockutils [None req-0d170cdb-983c-43d4-9998-8676a1404b75 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.245s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1041.602419] env[61957]: DEBUG nova.compute.manager [None req-0d170cdb-983c-43d4-9998-8676a1404b75 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 42860944-c700-43be-b291-970803fc7955] Start building networks asynchronously for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1041.605195] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d07fdeef-3a24-45b1-a008-52046b4459b1 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.666s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1041.605415] env[61957]: DEBUG nova.objects.instance [None req-d07fdeef-3a24-45b1-a008-52046b4459b1 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Lazy-loading 'resources' on Instance uuid 964c8c9d-0721-4ad3-a146-fce1fd1d6513 {{(pid=61957) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1041.697193] env[61957]: DEBUG nova.network.neutron [None req-de4650d3-4738-443f-8147-36e90c68fb19 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] [instance: 039a6b01-a745-478c-9ff7-c8c37afe2ce4] Successfully updated port: e49b8de4-47b9-400b-a0b0-9e0a0e68dc84 {{(pid=61957) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1041.858963] env[61957]: DEBUG oslo_concurrency.lockutils [None req-3ec44fbf-1f22-4ad1-8b26-d0253bbe3176 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Lock "9b0053fe-420e-4b92-86a8-ed44f6a6ec49" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 20.589s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1041.882199] env[61957]: DEBUG nova.compute.manager [req-b51d83b6-0d9c-4a2f-b766-4f2efe19ad34 req-b4a7c860-ea3b-4e0d-a7dc-c1c25da9baac service nova] [instance: 736eabcf-70f8-4e1e-9830-270d9971fc9a] Received event network-vif-deleted-95763a65-6383-409f-a9d6-1f37aaf44d34 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1041.882494] env[61957]: DEBUG nova.compute.manager [req-b51d83b6-0d9c-4a2f-b766-4f2efe19ad34 req-b4a7c860-ea3b-4e0d-a7dc-c1c25da9baac service nova] [instance: 039a6b01-a745-478c-9ff7-c8c37afe2ce4] Received event network-vif-plugged-e49b8de4-47b9-400b-a0b0-9e0a0e68dc84 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1041.882725] env[61957]: DEBUG oslo_concurrency.lockutils [req-b51d83b6-0d9c-4a2f-b766-4f2efe19ad34 req-b4a7c860-ea3b-4e0d-a7dc-c1c25da9baac service nova] Acquiring lock "039a6b01-a745-478c-9ff7-c8c37afe2ce4-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1041.882892] env[61957]: DEBUG oslo_concurrency.lockutils [req-b51d83b6-0d9c-4a2f-b766-4f2efe19ad34 req-b4a7c860-ea3b-4e0d-a7dc-c1c25da9baac service nova] Lock "039a6b01-a745-478c-9ff7-c8c37afe2ce4-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1041.883068] env[61957]: DEBUG oslo_concurrency.lockutils [req-b51d83b6-0d9c-4a2f-b766-4f2efe19ad34 req-b4a7c860-ea3b-4e0d-a7dc-c1c25da9baac service nova] Lock "039a6b01-a745-478c-9ff7-c8c37afe2ce4-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1041.883240] env[61957]: DEBUG nova.compute.manager [req-b51d83b6-0d9c-4a2f-b766-4f2efe19ad34 req-b4a7c860-ea3b-4e0d-a7dc-c1c25da9baac service nova] [instance: 039a6b01-a745-478c-9ff7-c8c37afe2ce4] No waiting events found dispatching network-vif-plugged-e49b8de4-47b9-400b-a0b0-9e0a0e68dc84 {{(pid=61957) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1041.883405] env[61957]: WARNING nova.compute.manager [req-b51d83b6-0d9c-4a2f-b766-4f2efe19ad34 req-b4a7c860-ea3b-4e0d-a7dc-c1c25da9baac service nova] [instance: 039a6b01-a745-478c-9ff7-c8c37afe2ce4] Received unexpected event network-vif-plugged-e49b8de4-47b9-400b-a0b0-9e0a0e68dc84 for instance with vm_state building and task_state spawning. [ 1041.883567] env[61957]: DEBUG nova.compute.manager [req-b51d83b6-0d9c-4a2f-b766-4f2efe19ad34 req-b4a7c860-ea3b-4e0d-a7dc-c1c25da9baac service nova] [instance: 039a6b01-a745-478c-9ff7-c8c37afe2ce4] Received event network-changed-e49b8de4-47b9-400b-a0b0-9e0a0e68dc84 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1041.883719] env[61957]: DEBUG nova.compute.manager [req-b51d83b6-0d9c-4a2f-b766-4f2efe19ad34 req-b4a7c860-ea3b-4e0d-a7dc-c1c25da9baac service nova] [instance: 039a6b01-a745-478c-9ff7-c8c37afe2ce4] Refreshing instance network info cache due to event network-changed-e49b8de4-47b9-400b-a0b0-9e0a0e68dc84. {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1041.883907] env[61957]: DEBUG oslo_concurrency.lockutils [req-b51d83b6-0d9c-4a2f-b766-4f2efe19ad34 req-b4a7c860-ea3b-4e0d-a7dc-c1c25da9baac service nova] Acquiring lock "refresh_cache-039a6b01-a745-478c-9ff7-c8c37afe2ce4" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1041.884048] env[61957]: DEBUG oslo_concurrency.lockutils [req-b51d83b6-0d9c-4a2f-b766-4f2efe19ad34 req-b4a7c860-ea3b-4e0d-a7dc-c1c25da9baac service nova] Acquired lock "refresh_cache-039a6b01-a745-478c-9ff7-c8c37afe2ce4" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1041.884213] env[61957]: DEBUG nova.network.neutron [req-b51d83b6-0d9c-4a2f-b766-4f2efe19ad34 req-b4a7c860-ea3b-4e0d-a7dc-c1c25da9baac service nova] [instance: 039a6b01-a745-478c-9ff7-c8c37afe2ce4] Refreshing network info cache for port e49b8de4-47b9-400b-a0b0-9e0a0e68dc84 {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1042.110762] env[61957]: DEBUG nova.compute.utils [None req-0d170cdb-983c-43d4-9998-8676a1404b75 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Using /dev/sd instead of None {{(pid=61957) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1042.111219] env[61957]: DEBUG nova.compute.manager [None req-0d170cdb-983c-43d4-9998-8676a1404b75 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 42860944-c700-43be-b291-970803fc7955] Allocating IP information in the background. {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1042.111427] env[61957]: DEBUG nova.network.neutron [None req-0d170cdb-983c-43d4-9998-8676a1404b75 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 42860944-c700-43be-b291-970803fc7955] allocate_for_instance() {{(pid=61957) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1042.154780] env[61957]: DEBUG nova.policy [None req-0d170cdb-983c-43d4-9998-8676a1404b75 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '85d4f8bb3aa64e7285aac42d14087ff1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e16f6dce3f0e44fb96de516f17d4c6f5', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61957) authorize /opt/stack/nova/nova/policy.py:203}} [ 1042.199624] env[61957]: DEBUG oslo_concurrency.lockutils [None req-de4650d3-4738-443f-8147-36e90c68fb19 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Acquiring lock "refresh_cache-039a6b01-a745-478c-9ff7-c8c37afe2ce4" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1042.255432] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04857546-e203-4223-b2c9-0b4cd690b930 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.264144] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03221c94-a45a-4839-95bf-c2fa6ce56e1e {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.294564] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93efebf9-2b5f-4a59-aa14-54ea71ab3b33 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.302892] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf6c8cae-1a1c-42e5-b637-5a3cafdf9072 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.318811] env[61957]: DEBUG nova.compute.provider_tree [None req-d07fdeef-3a24-45b1-a008-52046b4459b1 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1042.433410] env[61957]: DEBUG nova.network.neutron [req-b51d83b6-0d9c-4a2f-b766-4f2efe19ad34 req-b4a7c860-ea3b-4e0d-a7dc-c1c25da9baac service nova] [instance: 039a6b01-a745-478c-9ff7-c8c37afe2ce4] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1042.435843] env[61957]: DEBUG nova.network.neutron [None req-0d170cdb-983c-43d4-9998-8676a1404b75 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 42860944-c700-43be-b291-970803fc7955] Successfully created port: 0b2a4b91-7550-44b3-b4de-0fccf960ea77 {{(pid=61957) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1042.584194] env[61957]: DEBUG nova.network.neutron [req-b51d83b6-0d9c-4a2f-b766-4f2efe19ad34 req-b4a7c860-ea3b-4e0d-a7dc-c1c25da9baac service nova] [instance: 039a6b01-a745-478c-9ff7-c8c37afe2ce4] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1042.618793] env[61957]: DEBUG nova.compute.manager [None req-0d170cdb-983c-43d4-9998-8676a1404b75 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 42860944-c700-43be-b291-970803fc7955] Start building block device mappings for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1042.827612] env[61957]: DEBUG nova.scheduler.client.report [None req-d07fdeef-3a24-45b1-a008-52046b4459b1 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1043.087975] env[61957]: DEBUG oslo_concurrency.lockutils [req-b51d83b6-0d9c-4a2f-b766-4f2efe19ad34 req-b4a7c860-ea3b-4e0d-a7dc-c1c25da9baac service nova] Releasing lock "refresh_cache-039a6b01-a745-478c-9ff7-c8c37afe2ce4" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1043.088400] env[61957]: DEBUG oslo_concurrency.lockutils [None req-de4650d3-4738-443f-8147-36e90c68fb19 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Acquired lock "refresh_cache-039a6b01-a745-478c-9ff7-c8c37afe2ce4" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1043.088569] env[61957]: DEBUG nova.network.neutron [None req-de4650d3-4738-443f-8147-36e90c68fb19 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] [instance: 039a6b01-a745-478c-9ff7-c8c37afe2ce4] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1043.332579] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d07fdeef-3a24-45b1-a008-52046b4459b1 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.727s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1043.334976] env[61957]: DEBUG oslo_concurrency.lockutils [None req-4a6d948f-b9a8-4fc6-bd2c-f6dd4bec3479 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.074s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1043.336830] env[61957]: INFO nova.compute.claims [None req-4a6d948f-b9a8-4fc6-bd2c-f6dd4bec3479 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: a7419b30-1539-4977-a5d9-93a49d84914a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1043.355920] env[61957]: INFO nova.scheduler.client.report [None req-d07fdeef-3a24-45b1-a008-52046b4459b1 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Deleted allocations for instance 964c8c9d-0721-4ad3-a146-fce1fd1d6513 [ 1043.390284] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b7523c5-73d7-4ba4-8342-130e4e204cc6 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.398323] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-5257bf84-ad3e-4eff-bdeb-a7f7fbf613f6 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] [instance: 9b0053fe-420e-4b92-86a8-ed44f6a6ec49] Suspending the VM {{(pid=61957) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1163}} [ 1043.398632] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-77692655-4fcf-40b4-83ab-ff923bf25b20 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.405033] env[61957]: DEBUG oslo_vmware.api [None req-5257bf84-ad3e-4eff-bdeb-a7f7fbf613f6 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Waiting for the task: (returnval){ [ 1043.405033] env[61957]: value = "task-1278028" [ 1043.405033] env[61957]: _type = "Task" [ 1043.405033] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1043.414714] env[61957]: DEBUG oslo_vmware.api [None req-5257bf84-ad3e-4eff-bdeb-a7f7fbf613f6 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Task: {'id': task-1278028, 'name': SuspendVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.630009] env[61957]: DEBUG nova.network.neutron [None req-de4650d3-4738-443f-8147-36e90c68fb19 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] [instance: 039a6b01-a745-478c-9ff7-c8c37afe2ce4] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1043.634339] env[61957]: DEBUG nova.compute.manager [None req-0d170cdb-983c-43d4-9998-8676a1404b75 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 42860944-c700-43be-b291-970803fc7955] Start spawning the instance on the hypervisor. {{(pid=61957) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1043.663687] env[61957]: DEBUG nova.virt.hardware [None req-0d170cdb-983c-43d4-9998-8676a1404b75 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T17:22:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T17:21:50Z,direct_url=,disk_format='vmdk',id=11c76a2c-f705-470a-ba9d-4657858bab38,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='81d3096d0f094373913fc3dc8f5c3c6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T17:21:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1043.664047] env[61957]: DEBUG nova.virt.hardware [None req-0d170cdb-983c-43d4-9998-8676a1404b75 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Flavor limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1043.664263] env[61957]: DEBUG nova.virt.hardware [None req-0d170cdb-983c-43d4-9998-8676a1404b75 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Image limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1043.664493] env[61957]: DEBUG nova.virt.hardware [None req-0d170cdb-983c-43d4-9998-8676a1404b75 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Flavor pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1043.664764] env[61957]: DEBUG nova.virt.hardware [None req-0d170cdb-983c-43d4-9998-8676a1404b75 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Image pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1043.664967] env[61957]: DEBUG nova.virt.hardware [None req-0d170cdb-983c-43d4-9998-8676a1404b75 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1043.665305] env[61957]: DEBUG nova.virt.hardware [None req-0d170cdb-983c-43d4-9998-8676a1404b75 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1043.665481] env[61957]: DEBUG nova.virt.hardware [None req-0d170cdb-983c-43d4-9998-8676a1404b75 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1043.665693] env[61957]: DEBUG nova.virt.hardware [None req-0d170cdb-983c-43d4-9998-8676a1404b75 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Got 1 possible topologies {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1043.665896] env[61957]: DEBUG nova.virt.hardware [None req-0d170cdb-983c-43d4-9998-8676a1404b75 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1043.666143] env[61957]: DEBUG nova.virt.hardware [None req-0d170cdb-983c-43d4-9998-8676a1404b75 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1043.667114] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-104c2b45-53a8-42a5-af35-3341aa25b696 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.675654] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c926a4cd-6add-49e9-ac0a-b0c2429dcb09 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.848272] env[61957]: DEBUG nova.network.neutron [None req-de4650d3-4738-443f-8147-36e90c68fb19 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] [instance: 039a6b01-a745-478c-9ff7-c8c37afe2ce4] Updating instance_info_cache with network_info: [{"id": "e49b8de4-47b9-400b-a0b0-9e0a0e68dc84", "address": "fa:16:3e:c2:c7:e5", "network": {"id": "8e60ab13-6dd3-4cda-9cdf-974ad8bebe1f", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1404238496-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ea955bd4d84548ff8e6a0c21542d6ee7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9f208df-1fb5-4403-9796-7fd19e4bfb85", "external-id": "cl2-zone-400", "segmentation_id": 400, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape49b8de4-47", "ovs_interfaceid": "e49b8de4-47b9-400b-a0b0-9e0a0e68dc84", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1043.863302] env[61957]: DEBUG oslo_concurrency.lockutils [None req-d07fdeef-3a24-45b1-a008-52046b4459b1 tempest-AttachVolumeNegativeTest-1827655201 tempest-AttachVolumeNegativeTest-1827655201-project-member] Lock "964c8c9d-0721-4ad3-a146-fce1fd1d6513" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 12.048s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1043.915608] env[61957]: DEBUG oslo_vmware.api [None req-5257bf84-ad3e-4eff-bdeb-a7f7fbf613f6 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Task: {'id': task-1278028, 'name': SuspendVM_Task} progress is 70%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.941608] env[61957]: DEBUG nova.compute.manager [req-e0333265-e8b4-44b2-ba90-9cd207cc51e8 req-756aa891-52f0-427d-bf8f-063c009f9ee0 service nova] [instance: 42860944-c700-43be-b291-970803fc7955] Received event network-vif-plugged-0b2a4b91-7550-44b3-b4de-0fccf960ea77 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1043.941917] env[61957]: DEBUG oslo_concurrency.lockutils [req-e0333265-e8b4-44b2-ba90-9cd207cc51e8 req-756aa891-52f0-427d-bf8f-063c009f9ee0 service nova] Acquiring lock "42860944-c700-43be-b291-970803fc7955-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1043.942159] env[61957]: DEBUG oslo_concurrency.lockutils [req-e0333265-e8b4-44b2-ba90-9cd207cc51e8 req-756aa891-52f0-427d-bf8f-063c009f9ee0 service nova] Lock "42860944-c700-43be-b291-970803fc7955-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1043.942365] env[61957]: DEBUG oslo_concurrency.lockutils [req-e0333265-e8b4-44b2-ba90-9cd207cc51e8 req-756aa891-52f0-427d-bf8f-063c009f9ee0 service nova] Lock "42860944-c700-43be-b291-970803fc7955-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1043.942552] env[61957]: DEBUG nova.compute.manager [req-e0333265-e8b4-44b2-ba90-9cd207cc51e8 req-756aa891-52f0-427d-bf8f-063c009f9ee0 service nova] [instance: 42860944-c700-43be-b291-970803fc7955] No waiting events found dispatching network-vif-plugged-0b2a4b91-7550-44b3-b4de-0fccf960ea77 {{(pid=61957) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1043.942728] env[61957]: WARNING nova.compute.manager [req-e0333265-e8b4-44b2-ba90-9cd207cc51e8 req-756aa891-52f0-427d-bf8f-063c009f9ee0 service nova] [instance: 42860944-c700-43be-b291-970803fc7955] Received unexpected event network-vif-plugged-0b2a4b91-7550-44b3-b4de-0fccf960ea77 for instance with vm_state building and task_state spawning. [ 1044.032499] env[61957]: DEBUG nova.network.neutron [None req-0d170cdb-983c-43d4-9998-8676a1404b75 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 42860944-c700-43be-b291-970803fc7955] Successfully updated port: 0b2a4b91-7550-44b3-b4de-0fccf960ea77 {{(pid=61957) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1044.352343] env[61957]: DEBUG oslo_concurrency.lockutils [None req-de4650d3-4738-443f-8147-36e90c68fb19 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Releasing lock "refresh_cache-039a6b01-a745-478c-9ff7-c8c37afe2ce4" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1044.352750] env[61957]: DEBUG nova.compute.manager [None req-de4650d3-4738-443f-8147-36e90c68fb19 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] [instance: 039a6b01-a745-478c-9ff7-c8c37afe2ce4] Instance network_info: |[{"id": "e49b8de4-47b9-400b-a0b0-9e0a0e68dc84", "address": "fa:16:3e:c2:c7:e5", "network": {"id": "8e60ab13-6dd3-4cda-9cdf-974ad8bebe1f", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1404238496-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ea955bd4d84548ff8e6a0c21542d6ee7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9f208df-1fb5-4403-9796-7fd19e4bfb85", "external-id": "cl2-zone-400", "segmentation_id": 400, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape49b8de4-47", "ovs_interfaceid": "e49b8de4-47b9-400b-a0b0-9e0a0e68dc84", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1044.353487] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-de4650d3-4738-443f-8147-36e90c68fb19 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] [instance: 039a6b01-a745-478c-9ff7-c8c37afe2ce4] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c2:c7:e5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c9f208df-1fb5-4403-9796-7fd19e4bfb85', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e49b8de4-47b9-400b-a0b0-9e0a0e68dc84', 'vif_model': 'vmxnet3'}] {{(pid=61957) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1044.361238] env[61957]: DEBUG oslo.service.loopingcall [None req-de4650d3-4738-443f-8147-36e90c68fb19 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1044.361478] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 039a6b01-a745-478c-9ff7-c8c37afe2ce4] Creating VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1044.361743] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7ce02d42-1dd4-4e52-ae84-1aefcd4b4fee {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.386326] env[61957]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1044.386326] env[61957]: value = "task-1278030" [ 1044.386326] env[61957]: _type = "Task" [ 1044.386326] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1044.395577] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1278030, 'name': CreateVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.415892] env[61957]: DEBUG oslo_vmware.api [None req-5257bf84-ad3e-4eff-bdeb-a7f7fbf613f6 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Task: {'id': task-1278028, 'name': SuspendVM_Task, 'duration_secs': 0.590367} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1044.415983] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-5257bf84-ad3e-4eff-bdeb-a7f7fbf613f6 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] [instance: 9b0053fe-420e-4b92-86a8-ed44f6a6ec49] Suspended the VM {{(pid=61957) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1167}} [ 1044.416932] env[61957]: DEBUG nova.compute.manager [None req-5257bf84-ad3e-4eff-bdeb-a7f7fbf613f6 tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] [instance: 9b0053fe-420e-4b92-86a8-ed44f6a6ec49] Checking state {{(pid=61957) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1044.416932] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2d2c14c-d21a-48fd-92b4-d6f3aab9573b {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.486871] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43a26acb-0ec4-4f24-91f2-69b17b242aa0 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.494441] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd4ade21-3c3e-482a-9b86-579061670150 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.524245] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3c4241e-45f8-4fc2-8da4-9fd6ea948348 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.531422] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11c13dc0-d2b1-4703-a2c4-c79059706e76 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.535319] env[61957]: DEBUG oslo_concurrency.lockutils [None req-0d170cdb-983c-43d4-9998-8676a1404b75 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Acquiring lock "refresh_cache-42860944-c700-43be-b291-970803fc7955" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1044.535484] env[61957]: DEBUG oslo_concurrency.lockutils [None req-0d170cdb-983c-43d4-9998-8676a1404b75 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Acquired lock "refresh_cache-42860944-c700-43be-b291-970803fc7955" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1044.535716] env[61957]: DEBUG nova.network.neutron [None req-0d170cdb-983c-43d4-9998-8676a1404b75 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 42860944-c700-43be-b291-970803fc7955] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1044.546300] env[61957]: DEBUG nova.compute.provider_tree [None req-4a6d948f-b9a8-4fc6-bd2c-f6dd4bec3479 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1044.896784] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1278030, 'name': CreateVM_Task, 'duration_secs': 0.356025} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1044.897035] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 039a6b01-a745-478c-9ff7-c8c37afe2ce4] Created VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1044.897764] env[61957]: DEBUG oslo_concurrency.lockutils [None req-de4650d3-4738-443f-8147-36e90c68fb19 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1044.897953] env[61957]: DEBUG oslo_concurrency.lockutils [None req-de4650d3-4738-443f-8147-36e90c68fb19 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Acquired lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1044.898297] env[61957]: DEBUG oslo_concurrency.lockutils [None req-de4650d3-4738-443f-8147-36e90c68fb19 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1044.898551] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-86609749-693f-47d0-9c44-7ded9f28a8df {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.903229] env[61957]: DEBUG oslo_vmware.api [None req-de4650d3-4738-443f-8147-36e90c68fb19 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Waiting for the task: (returnval){ [ 1044.903229] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]5224d71f-9295-7d1b-40fe-94002ed82bff" [ 1044.903229] env[61957]: _type = "Task" [ 1044.903229] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1044.911980] env[61957]: DEBUG oslo_vmware.api [None req-de4650d3-4738-443f-8147-36e90c68fb19 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]5224d71f-9295-7d1b-40fe-94002ed82bff, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.049134] env[61957]: DEBUG nova.scheduler.client.report [None req-4a6d948f-b9a8-4fc6-bd2c-f6dd4bec3479 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1045.067836] env[61957]: DEBUG nova.network.neutron [None req-0d170cdb-983c-43d4-9998-8676a1404b75 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 42860944-c700-43be-b291-970803fc7955] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1045.204874] env[61957]: DEBUG nova.network.neutron [None req-0d170cdb-983c-43d4-9998-8676a1404b75 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 42860944-c700-43be-b291-970803fc7955] Updating instance_info_cache with network_info: [{"id": "0b2a4b91-7550-44b3-b4de-0fccf960ea77", "address": "fa:16:3e:25:f2:60", "network": {"id": "efe08823-51fe-4620-a849-5345e262cc71", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1035436819-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e16f6dce3f0e44fb96de516f17d4c6f5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "667a2e97-c1be-421d-9941-6b84c2629b43", "external-id": "nsx-vlan-transportzone-484", "segmentation_id": 484, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0b2a4b91-75", "ovs_interfaceid": "0b2a4b91-7550-44b3-b4de-0fccf960ea77", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1045.413561] env[61957]: DEBUG oslo_vmware.api [None req-de4650d3-4738-443f-8147-36e90c68fb19 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]5224d71f-9295-7d1b-40fe-94002ed82bff, 'name': SearchDatastore_Task, 'duration_secs': 0.009255} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1045.413889] env[61957]: DEBUG oslo_concurrency.lockutils [None req-de4650d3-4738-443f-8147-36e90c68fb19 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Releasing lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1045.414146] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-de4650d3-4738-443f-8147-36e90c68fb19 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] [instance: 039a6b01-a745-478c-9ff7-c8c37afe2ce4] Processing image 11c76a2c-f705-470a-ba9d-4657858bab38 {{(pid=61957) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1045.414380] env[61957]: DEBUG oslo_concurrency.lockutils [None req-de4650d3-4738-443f-8147-36e90c68fb19 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1045.414532] env[61957]: DEBUG oslo_concurrency.lockutils [None req-de4650d3-4738-443f-8147-36e90c68fb19 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Acquired lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1045.414713] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-de4650d3-4738-443f-8147-36e90c68fb19 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1045.414966] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4bc23bcc-39ee-4619-9830-01ed2e59c724 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.422314] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-de4650d3-4738-443f-8147-36e90c68fb19 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1045.422531] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-de4650d3-4738-443f-8147-36e90c68fb19 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61957) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1045.423156] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e64be169-e6bf-446b-ab66-1cd61312caea {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.428011] env[61957]: DEBUG oslo_vmware.api [None req-de4650d3-4738-443f-8147-36e90c68fb19 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Waiting for the task: (returnval){ [ 1045.428011] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52e61eb6-dfaf-2011-72d1-b87a81da019c" [ 1045.428011] env[61957]: _type = "Task" [ 1045.428011] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1045.435275] env[61957]: DEBUG oslo_vmware.api [None req-de4650d3-4738-443f-8147-36e90c68fb19 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52e61eb6-dfaf-2011-72d1-b87a81da019c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.554888] env[61957]: DEBUG oslo_concurrency.lockutils [None req-4a6d948f-b9a8-4fc6-bd2c-f6dd4bec3479 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.220s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1045.555397] env[61957]: DEBUG nova.compute.manager [None req-4a6d948f-b9a8-4fc6-bd2c-f6dd4bec3479 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: a7419b30-1539-4977-a5d9-93a49d84914a] Start building networks asynchronously for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1045.557917] env[61957]: DEBUG oslo_concurrency.lockutils [None req-5a8aaf0c-6884-47c6-aa4a-9ccf495b4481 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 7.865s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1045.558134] env[61957]: DEBUG oslo_concurrency.lockutils [None req-5a8aaf0c-6884-47c6-aa4a-9ccf495b4481 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1045.560196] env[61957]: DEBUG oslo_concurrency.lockutils [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 6.832s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1045.560361] env[61957]: DEBUG oslo_concurrency.lockutils [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1045.560508] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61957) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1045.560787] env[61957]: DEBUG oslo_concurrency.lockutils [None req-8a0a01e3-6392-4a5c-b585-5e26a0a02629 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.734s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1045.562630] env[61957]: INFO nova.compute.claims [None req-8a0a01e3-6392-4a5c-b585-5e26a0a02629 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 4625d1dc-a621-4a33-9537-9e8cce96c914] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1045.565978] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f614c8a-76cb-4bc2-acf0-b7053141648d {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.573992] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d273c1d-3191-4e3a-b11b-4b29668083d2 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.587999] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e382bd3c-50f4-4a4a-963c-4afdfc8914bc {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.591584] env[61957]: INFO nova.scheduler.client.report [None req-5a8aaf0c-6884-47c6-aa4a-9ccf495b4481 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Deleted allocations for instance 6509707e-afc6-4c4a-88f9-a471865d510c [ 1045.597364] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4307ee9-c1e1-4669-bbe6-6987b5402234 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.628530] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=179634MB free_disk=142GB free_vcpus=48 pci_devices=None {{(pid=61957) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1045.628798] env[61957]: DEBUG oslo_concurrency.lockutils [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1045.707451] env[61957]: DEBUG oslo_concurrency.lockutils [None req-0d170cdb-983c-43d4-9998-8676a1404b75 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Releasing lock "refresh_cache-42860944-c700-43be-b291-970803fc7955" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1045.707780] env[61957]: DEBUG nova.compute.manager [None req-0d170cdb-983c-43d4-9998-8676a1404b75 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 42860944-c700-43be-b291-970803fc7955] Instance network_info: |[{"id": "0b2a4b91-7550-44b3-b4de-0fccf960ea77", "address": "fa:16:3e:25:f2:60", "network": {"id": "efe08823-51fe-4620-a849-5345e262cc71", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1035436819-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e16f6dce3f0e44fb96de516f17d4c6f5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "667a2e97-c1be-421d-9941-6b84c2629b43", "external-id": "nsx-vlan-transportzone-484", "segmentation_id": 484, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0b2a4b91-75", "ovs_interfaceid": "0b2a4b91-7550-44b3-b4de-0fccf960ea77", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1045.708236] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-0d170cdb-983c-43d4-9998-8676a1404b75 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 42860944-c700-43be-b291-970803fc7955] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:25:f2:60', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '667a2e97-c1be-421d-9941-6b84c2629b43', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0b2a4b91-7550-44b3-b4de-0fccf960ea77', 'vif_model': 'vmxnet3'}] {{(pid=61957) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1045.715667] env[61957]: DEBUG oslo.service.loopingcall [None req-0d170cdb-983c-43d4-9998-8676a1404b75 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1045.716265] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 42860944-c700-43be-b291-970803fc7955] Creating VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1045.716265] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-be80db72-33e7-44df-a8a3-b86131f6f6f6 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.736286] env[61957]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1045.736286] env[61957]: value = "task-1278031" [ 1045.736286] env[61957]: _type = "Task" [ 1045.736286] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1045.743433] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1278031, 'name': CreateVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.795664] env[61957]: INFO nova.compute.manager [None req-7e98aea5-9e6c-437d-b311-a1ce6e8e055a tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] [instance: 9b0053fe-420e-4b92-86a8-ed44f6a6ec49] Resuming [ 1045.796359] env[61957]: DEBUG nova.objects.instance [None req-7e98aea5-9e6c-437d-b311-a1ce6e8e055a tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Lazy-loading 'flavor' on Instance uuid 9b0053fe-420e-4b92-86a8-ed44f6a6ec49 {{(pid=61957) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1045.940657] env[61957]: DEBUG oslo_vmware.api [None req-de4650d3-4738-443f-8147-36e90c68fb19 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52e61eb6-dfaf-2011-72d1-b87a81da019c, 'name': SearchDatastore_Task, 'duration_secs': 0.008366} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1045.941639] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-60788ed7-e246-4fab-b81d-4617f39f2359 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.950766] env[61957]: DEBUG oslo_vmware.api [None req-de4650d3-4738-443f-8147-36e90c68fb19 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Waiting for the task: (returnval){ [ 1045.950766] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]525b946e-6034-95d3-4f33-50e8a3e4397f" [ 1045.950766] env[61957]: _type = "Task" [ 1045.950766] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1045.961132] env[61957]: DEBUG oslo_vmware.api [None req-de4650d3-4738-443f-8147-36e90c68fb19 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]525b946e-6034-95d3-4f33-50e8a3e4397f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.968616] env[61957]: DEBUG nova.compute.manager [req-9c4469d6-9247-4d5e-9f94-27670509f778 req-a594ae8c-d0cf-4970-8ede-d2ee3d02ba55 service nova] [instance: 42860944-c700-43be-b291-970803fc7955] Received event network-changed-0b2a4b91-7550-44b3-b4de-0fccf960ea77 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1045.968708] env[61957]: DEBUG nova.compute.manager [req-9c4469d6-9247-4d5e-9f94-27670509f778 req-a594ae8c-d0cf-4970-8ede-d2ee3d02ba55 service nova] [instance: 42860944-c700-43be-b291-970803fc7955] Refreshing instance network info cache due to event network-changed-0b2a4b91-7550-44b3-b4de-0fccf960ea77. {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1045.968929] env[61957]: DEBUG oslo_concurrency.lockutils [req-9c4469d6-9247-4d5e-9f94-27670509f778 req-a594ae8c-d0cf-4970-8ede-d2ee3d02ba55 service nova] Acquiring lock "refresh_cache-42860944-c700-43be-b291-970803fc7955" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1045.969093] env[61957]: DEBUG oslo_concurrency.lockutils [req-9c4469d6-9247-4d5e-9f94-27670509f778 req-a594ae8c-d0cf-4970-8ede-d2ee3d02ba55 service nova] Acquired lock "refresh_cache-42860944-c700-43be-b291-970803fc7955" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1045.969256] env[61957]: DEBUG nova.network.neutron [req-9c4469d6-9247-4d5e-9f94-27670509f778 req-a594ae8c-d0cf-4970-8ede-d2ee3d02ba55 service nova] [instance: 42860944-c700-43be-b291-970803fc7955] Refreshing network info cache for port 0b2a4b91-7550-44b3-b4de-0fccf960ea77 {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1046.067151] env[61957]: DEBUG nova.compute.utils [None req-4a6d948f-b9a8-4fc6-bd2c-f6dd4bec3479 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Using /dev/sd instead of None {{(pid=61957) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1046.068509] env[61957]: DEBUG nova.compute.manager [None req-4a6d948f-b9a8-4fc6-bd2c-f6dd4bec3479 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: a7419b30-1539-4977-a5d9-93a49d84914a] Allocating IP information in the background. {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1046.068659] env[61957]: DEBUG nova.network.neutron [None req-4a6d948f-b9a8-4fc6-bd2c-f6dd4bec3479 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: a7419b30-1539-4977-a5d9-93a49d84914a] allocate_for_instance() {{(pid=61957) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1046.104065] env[61957]: DEBUG oslo_concurrency.lockutils [None req-5a8aaf0c-6884-47c6-aa4a-9ccf495b4481 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Lock "6509707e-afc6-4c4a-88f9-a471865d510c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.355s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1046.108560] env[61957]: DEBUG nova.policy [None req-4a6d948f-b9a8-4fc6-bd2c-f6dd4bec3479 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3136bf0d1db3411fb1ff1ebe9c6045f6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1dbdc699741a48af93c52068d88a5357', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61957) authorize /opt/stack/nova/nova/policy.py:203}} [ 1046.246155] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1278031, 'name': CreateVM_Task, 'duration_secs': 0.280999} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1046.246322] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 42860944-c700-43be-b291-970803fc7955] Created VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1046.247100] env[61957]: DEBUG oslo_concurrency.lockutils [None req-0d170cdb-983c-43d4-9998-8676a1404b75 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1046.247276] env[61957]: DEBUG oslo_concurrency.lockutils [None req-0d170cdb-983c-43d4-9998-8676a1404b75 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Acquired lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1046.247598] env[61957]: DEBUG oslo_concurrency.lockutils [None req-0d170cdb-983c-43d4-9998-8676a1404b75 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1046.247840] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-64304ba8-5357-4c49-8739-73ea2dab4d4a {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.252504] env[61957]: DEBUG oslo_vmware.api [None req-0d170cdb-983c-43d4-9998-8676a1404b75 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Waiting for the task: (returnval){ [ 1046.252504] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]521cade0-3b38-a7f3-39db-bdab5f8e4fdd" [ 1046.252504] env[61957]: _type = "Task" [ 1046.252504] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.259484] env[61957]: DEBUG oslo_vmware.api [None req-0d170cdb-983c-43d4-9998-8676a1404b75 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]521cade0-3b38-a7f3-39db-bdab5f8e4fdd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.343787] env[61957]: DEBUG nova.network.neutron [None req-4a6d948f-b9a8-4fc6-bd2c-f6dd4bec3479 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: a7419b30-1539-4977-a5d9-93a49d84914a] Successfully created port: 6bfe39fb-7cc2-4933-bdf0-f8b2a121aab2 {{(pid=61957) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1046.462322] env[61957]: DEBUG oslo_vmware.api [None req-de4650d3-4738-443f-8147-36e90c68fb19 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]525b946e-6034-95d3-4f33-50e8a3e4397f, 'name': SearchDatastore_Task, 'duration_secs': 0.010819} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1046.462322] env[61957]: DEBUG oslo_concurrency.lockutils [None req-de4650d3-4738-443f-8147-36e90c68fb19 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Releasing lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1046.462672] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-de4650d3-4738-443f-8147-36e90c68fb19 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore1] 039a6b01-a745-478c-9ff7-c8c37afe2ce4/039a6b01-a745-478c-9ff7-c8c37afe2ce4.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1046.462837] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c8fbe1ed-125c-4cf7-b47f-84c2d9438cca {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.473414] env[61957]: DEBUG oslo_vmware.api [None req-de4650d3-4738-443f-8147-36e90c68fb19 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Waiting for the task: (returnval){ [ 1046.473414] env[61957]: value = "task-1278032" [ 1046.473414] env[61957]: _type = "Task" [ 1046.473414] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.482091] env[61957]: DEBUG oslo_vmware.api [None req-de4650d3-4738-443f-8147-36e90c68fb19 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Task: {'id': task-1278032, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.572259] env[61957]: DEBUG nova.compute.manager [None req-4a6d948f-b9a8-4fc6-bd2c-f6dd4bec3479 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: a7419b30-1539-4977-a5d9-93a49d84914a] Start building block device mappings for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1046.737563] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e5cf626-099b-4f0e-a12c-4a616df43a23 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.761493] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03b1e7e9-a8c5-4cd8-83d8-d9121fc36d39 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.773666] env[61957]: DEBUG oslo_vmware.api [None req-0d170cdb-983c-43d4-9998-8676a1404b75 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]521cade0-3b38-a7f3-39db-bdab5f8e4fdd, 'name': SearchDatastore_Task, 'duration_secs': 0.008451} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1046.805127] env[61957]: DEBUG oslo_concurrency.lockutils [None req-0d170cdb-983c-43d4-9998-8676a1404b75 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Releasing lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1046.805521] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-0d170cdb-983c-43d4-9998-8676a1404b75 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 42860944-c700-43be-b291-970803fc7955] Processing image 11c76a2c-f705-470a-ba9d-4657858bab38 {{(pid=61957) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1046.805849] env[61957]: DEBUG oslo_concurrency.lockutils [None req-0d170cdb-983c-43d4-9998-8676a1404b75 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1046.806020] env[61957]: DEBUG oslo_concurrency.lockutils [None req-0d170cdb-983c-43d4-9998-8676a1404b75 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Acquired lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1046.806272] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-0d170cdb-983c-43d4-9998-8676a1404b75 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1046.807569] env[61957]: DEBUG oslo_concurrency.lockutils [None req-7e98aea5-9e6c-437d-b311-a1ce6e8e055a tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Acquiring lock "refresh_cache-9b0053fe-420e-4b92-86a8-ed44f6a6ec49" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1046.807755] env[61957]: DEBUG oslo_concurrency.lockutils [None req-7e98aea5-9e6c-437d-b311-a1ce6e8e055a tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Acquired lock "refresh_cache-9b0053fe-420e-4b92-86a8-ed44f6a6ec49" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1046.807994] env[61957]: DEBUG nova.network.neutron [None req-7e98aea5-9e6c-437d-b311-a1ce6e8e055a tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] [instance: 9b0053fe-420e-4b92-86a8-ed44f6a6ec49] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1046.810898] env[61957]: DEBUG nova.network.neutron [req-9c4469d6-9247-4d5e-9f94-27670509f778 req-a594ae8c-d0cf-4970-8ede-d2ee3d02ba55 service nova] [instance: 42860944-c700-43be-b291-970803fc7955] Updated VIF entry in instance network info cache for port 0b2a4b91-7550-44b3-b4de-0fccf960ea77. {{(pid=61957) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1046.811302] env[61957]: DEBUG nova.network.neutron [req-9c4469d6-9247-4d5e-9f94-27670509f778 req-a594ae8c-d0cf-4970-8ede-d2ee3d02ba55 service nova] [instance: 42860944-c700-43be-b291-970803fc7955] Updating instance_info_cache with network_info: [{"id": "0b2a4b91-7550-44b3-b4de-0fccf960ea77", "address": "fa:16:3e:25:f2:60", "network": {"id": "efe08823-51fe-4620-a849-5345e262cc71", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1035436819-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e16f6dce3f0e44fb96de516f17d4c6f5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "667a2e97-c1be-421d-9941-6b84c2629b43", "external-id": "nsx-vlan-transportzone-484", "segmentation_id": 484, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0b2a4b91-75", "ovs_interfaceid": "0b2a4b91-7550-44b3-b4de-0fccf960ea77", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1046.812639] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a1e8d40b-7091-4883-a5e3-8292ddf767ba {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.815861] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbad82fe-3484-4e67-8d56-34aa4689551c {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.827793] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-235cb63c-bdc0-4fe3-9ccb-2e47ff49637c {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.834167] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-0d170cdb-983c-43d4-9998-8676a1404b75 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1046.834453] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-0d170cdb-983c-43d4-9998-8676a1404b75 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61957) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1046.835953] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-05415608-13df-42fc-8136-74b6b1de4d8c {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.849755] env[61957]: DEBUG nova.compute.provider_tree [None req-8a0a01e3-6392-4a5c-b585-5e26a0a02629 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1046.857544] env[61957]: DEBUG oslo_vmware.api [None req-0d170cdb-983c-43d4-9998-8676a1404b75 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Waiting for the task: (returnval){ [ 1046.857544] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]5239d9af-65ba-5994-1fbe-d7fa8fdc4caa" [ 1046.857544] env[61957]: _type = "Task" [ 1046.857544] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.871683] env[61957]: DEBUG oslo_vmware.api [None req-0d170cdb-983c-43d4-9998-8676a1404b75 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]5239d9af-65ba-5994-1fbe-d7fa8fdc4caa, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.983905] env[61957]: DEBUG oslo_vmware.api [None req-de4650d3-4738-443f-8147-36e90c68fb19 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Task: {'id': task-1278032, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.430471} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1046.984197] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-de4650d3-4738-443f-8147-36e90c68fb19 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore1] 039a6b01-a745-478c-9ff7-c8c37afe2ce4/039a6b01-a745-478c-9ff7-c8c37afe2ce4.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1046.984416] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-de4650d3-4738-443f-8147-36e90c68fb19 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] [instance: 039a6b01-a745-478c-9ff7-c8c37afe2ce4] Extending root virtual disk to 1048576 {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1046.984675] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-fc816a36-5257-431e-81af-e5c2af381f32 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.991124] env[61957]: DEBUG oslo_vmware.api [None req-de4650d3-4738-443f-8147-36e90c68fb19 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Waiting for the task: (returnval){ [ 1046.991124] env[61957]: value = "task-1278033" [ 1046.991124] env[61957]: _type = "Task" [ 1046.991124] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.998780] env[61957]: DEBUG oslo_vmware.api [None req-de4650d3-4738-443f-8147-36e90c68fb19 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Task: {'id': task-1278033, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.321023] env[61957]: DEBUG oslo_concurrency.lockutils [req-9c4469d6-9247-4d5e-9f94-27670509f778 req-a594ae8c-d0cf-4970-8ede-d2ee3d02ba55 service nova] Releasing lock "refresh_cache-42860944-c700-43be-b291-970803fc7955" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1047.353642] env[61957]: DEBUG nova.scheduler.client.report [None req-8a0a01e3-6392-4a5c-b585-5e26a0a02629 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1047.367561] env[61957]: DEBUG oslo_vmware.api [None req-0d170cdb-983c-43d4-9998-8676a1404b75 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]5239d9af-65ba-5994-1fbe-d7fa8fdc4caa, 'name': SearchDatastore_Task, 'duration_secs': 0.032864} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.368357] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-27bd58fa-5d8d-44ab-b660-f6522d69598b {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.373603] env[61957]: DEBUG oslo_vmware.api [None req-0d170cdb-983c-43d4-9998-8676a1404b75 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Waiting for the task: (returnval){ [ 1047.373603] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52c7fe02-7a2d-44f4-fc6f-49116d1ed806" [ 1047.373603] env[61957]: _type = "Task" [ 1047.373603] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.383835] env[61957]: DEBUG oslo_vmware.api [None req-0d170cdb-983c-43d4-9998-8676a1404b75 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52c7fe02-7a2d-44f4-fc6f-49116d1ed806, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.501032] env[61957]: DEBUG oslo_vmware.api [None req-de4650d3-4738-443f-8147-36e90c68fb19 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Task: {'id': task-1278033, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.059792} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.501327] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-de4650d3-4738-443f-8147-36e90c68fb19 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] [instance: 039a6b01-a745-478c-9ff7-c8c37afe2ce4] Extended root virtual disk {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1047.502113] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f8cafb9-ff77-4768-a4a4-911e9fa110f6 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.524362] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-de4650d3-4738-443f-8147-36e90c68fb19 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] [instance: 039a6b01-a745-478c-9ff7-c8c37afe2ce4] Reconfiguring VM instance instance-00000062 to attach disk [datastore1] 039a6b01-a745-478c-9ff7-c8c37afe2ce4/039a6b01-a745-478c-9ff7-c8c37afe2ce4.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1047.525462] env[61957]: DEBUG nova.network.neutron [None req-7e98aea5-9e6c-437d-b311-a1ce6e8e055a tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] [instance: 9b0053fe-420e-4b92-86a8-ed44f6a6ec49] Updating instance_info_cache with network_info: [{"id": "219230f0-0b14-4be9-81bb-67a765d8ced9", "address": "fa:16:3e:f6:9e:ce", "network": {"id": "cf378e9c-eaae-4168-bd60-9818fcfaf405", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1243326371-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f7c20ab3c822423fa837a03f066e4239", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8cb478a6-872c-4a90-a8db-526b374e82ce", "external-id": "nsx-vlan-transportzone-835", "segmentation_id": 835, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap219230f0-0b", "ovs_interfaceid": "219230f0-0b14-4be9-81bb-67a765d8ced9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1047.526706] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4a21b536-690d-4f73-8909-c2b3fa0d031d {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.547587] env[61957]: DEBUG oslo_vmware.api [None req-de4650d3-4738-443f-8147-36e90c68fb19 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Waiting for the task: (returnval){ [ 1047.547587] env[61957]: value = "task-1278035" [ 1047.547587] env[61957]: _type = "Task" [ 1047.547587] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.555972] env[61957]: DEBUG oslo_vmware.api [None req-de4650d3-4738-443f-8147-36e90c68fb19 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Task: {'id': task-1278035, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.585222] env[61957]: DEBUG nova.compute.manager [None req-4a6d948f-b9a8-4fc6-bd2c-f6dd4bec3479 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: a7419b30-1539-4977-a5d9-93a49d84914a] Start spawning the instance on the hypervisor. {{(pid=61957) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1047.610657] env[61957]: DEBUG nova.virt.hardware [None req-4a6d948f-b9a8-4fc6-bd2c-f6dd4bec3479 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T17:22:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T17:21:50Z,direct_url=,disk_format='vmdk',id=11c76a2c-f705-470a-ba9d-4657858bab38,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='81d3096d0f094373913fc3dc8f5c3c6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T17:21:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1047.610911] env[61957]: DEBUG nova.virt.hardware [None req-4a6d948f-b9a8-4fc6-bd2c-f6dd4bec3479 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Flavor limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1047.611082] env[61957]: DEBUG nova.virt.hardware [None req-4a6d948f-b9a8-4fc6-bd2c-f6dd4bec3479 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Image limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1047.611273] env[61957]: DEBUG nova.virt.hardware [None req-4a6d948f-b9a8-4fc6-bd2c-f6dd4bec3479 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Flavor pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1047.611420] env[61957]: DEBUG nova.virt.hardware [None req-4a6d948f-b9a8-4fc6-bd2c-f6dd4bec3479 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Image pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1047.611568] env[61957]: DEBUG nova.virt.hardware [None req-4a6d948f-b9a8-4fc6-bd2c-f6dd4bec3479 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1047.611773] env[61957]: DEBUG nova.virt.hardware [None req-4a6d948f-b9a8-4fc6-bd2c-f6dd4bec3479 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1047.611934] env[61957]: DEBUG nova.virt.hardware [None req-4a6d948f-b9a8-4fc6-bd2c-f6dd4bec3479 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1047.612118] env[61957]: DEBUG nova.virt.hardware [None req-4a6d948f-b9a8-4fc6-bd2c-f6dd4bec3479 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Got 1 possible topologies {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1047.612285] env[61957]: DEBUG nova.virt.hardware [None req-4a6d948f-b9a8-4fc6-bd2c-f6dd4bec3479 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1047.612518] env[61957]: DEBUG nova.virt.hardware [None req-4a6d948f-b9a8-4fc6-bd2c-f6dd4bec3479 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1047.613756] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1e2ba65-d663-4758-8a90-c4ee0ccdb034 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.621455] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-467a7869-17f3-412b-93e7-e90670cd1943 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.858238] env[61957]: DEBUG oslo_concurrency.lockutils [None req-8a0a01e3-6392-4a5c-b585-5e26a0a02629 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.297s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1047.858779] env[61957]: DEBUG nova.compute.manager [None req-8a0a01e3-6392-4a5c-b585-5e26a0a02629 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 4625d1dc-a621-4a33-9537-9e8cce96c914] Start building networks asynchronously for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1047.861651] env[61957]: DEBUG oslo_concurrency.lockutils [None req-b6b5029a-df23-45ff-9311-ae897f587f74 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.518s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1047.861874] env[61957]: DEBUG nova.objects.instance [None req-b6b5029a-df23-45ff-9311-ae897f587f74 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Lazy-loading 'resources' on Instance uuid 84cc9489-2a11-4510-8ad6-13dc41814856 {{(pid=61957) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1047.883422] env[61957]: DEBUG oslo_vmware.api [None req-0d170cdb-983c-43d4-9998-8676a1404b75 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52c7fe02-7a2d-44f4-fc6f-49116d1ed806, 'name': SearchDatastore_Task, 'duration_secs': 0.009567} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.883811] env[61957]: DEBUG oslo_concurrency.lockutils [None req-0d170cdb-983c-43d4-9998-8676a1404b75 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Releasing lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1047.884211] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-0d170cdb-983c-43d4-9998-8676a1404b75 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore1] 42860944-c700-43be-b291-970803fc7955/42860944-c700-43be-b291-970803fc7955.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1047.884579] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-eddb2f5e-6e57-490e-88bb-a03f6b96475a {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.890171] env[61957]: DEBUG oslo_vmware.api [None req-0d170cdb-983c-43d4-9998-8676a1404b75 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Waiting for the task: (returnval){ [ 1047.890171] env[61957]: value = "task-1278036" [ 1047.890171] env[61957]: _type = "Task" [ 1047.890171] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.897823] env[61957]: DEBUG oslo_vmware.api [None req-0d170cdb-983c-43d4-9998-8676a1404b75 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1278036, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.978347] env[61957]: DEBUG nova.network.neutron [None req-4a6d948f-b9a8-4fc6-bd2c-f6dd4bec3479 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: a7419b30-1539-4977-a5d9-93a49d84914a] Successfully updated port: 6bfe39fb-7cc2-4933-bdf0-f8b2a121aab2 {{(pid=61957) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1047.994263] env[61957]: DEBUG nova.compute.manager [req-42206133-af19-4f4e-8412-6942369792dc req-59911953-6101-493c-8b1a-8ef47fd4d078 service nova] [instance: a7419b30-1539-4977-a5d9-93a49d84914a] Received event network-vif-plugged-6bfe39fb-7cc2-4933-bdf0-f8b2a121aab2 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1047.994577] env[61957]: DEBUG oslo_concurrency.lockutils [req-42206133-af19-4f4e-8412-6942369792dc req-59911953-6101-493c-8b1a-8ef47fd4d078 service nova] Acquiring lock "a7419b30-1539-4977-a5d9-93a49d84914a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1047.994880] env[61957]: DEBUG oslo_concurrency.lockutils [req-42206133-af19-4f4e-8412-6942369792dc req-59911953-6101-493c-8b1a-8ef47fd4d078 service nova] Lock "a7419b30-1539-4977-a5d9-93a49d84914a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1047.995199] env[61957]: DEBUG oslo_concurrency.lockutils [req-42206133-af19-4f4e-8412-6942369792dc req-59911953-6101-493c-8b1a-8ef47fd4d078 service nova] Lock "a7419b30-1539-4977-a5d9-93a49d84914a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1047.995413] env[61957]: DEBUG nova.compute.manager [req-42206133-af19-4f4e-8412-6942369792dc req-59911953-6101-493c-8b1a-8ef47fd4d078 service nova] [instance: a7419b30-1539-4977-a5d9-93a49d84914a] No waiting events found dispatching network-vif-plugged-6bfe39fb-7cc2-4933-bdf0-f8b2a121aab2 {{(pid=61957) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1047.995670] env[61957]: WARNING nova.compute.manager [req-42206133-af19-4f4e-8412-6942369792dc req-59911953-6101-493c-8b1a-8ef47fd4d078 service nova] [instance: a7419b30-1539-4977-a5d9-93a49d84914a] Received unexpected event network-vif-plugged-6bfe39fb-7cc2-4933-bdf0-f8b2a121aab2 for instance with vm_state building and task_state spawning. [ 1048.042354] env[61957]: DEBUG oslo_concurrency.lockutils [None req-7e98aea5-9e6c-437d-b311-a1ce6e8e055a tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Releasing lock "refresh_cache-9b0053fe-420e-4b92-86a8-ed44f6a6ec49" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1048.043499] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7469edde-d55c-466c-a1e0-e4ff01732d45 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.054327] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-7e98aea5-9e6c-437d-b311-a1ce6e8e055a tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] [instance: 9b0053fe-420e-4b92-86a8-ed44f6a6ec49] Resuming the VM {{(pid=61957) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1184}} [ 1048.055536] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4cbbbd5a-b03e-4d73-83f5-72b9cf4d0d94 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.061221] env[61957]: DEBUG oslo_vmware.api [None req-de4650d3-4738-443f-8147-36e90c68fb19 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Task: {'id': task-1278035, 'name': ReconfigVM_Task, 'duration_secs': 0.320233} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.062368] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-de4650d3-4738-443f-8147-36e90c68fb19 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] [instance: 039a6b01-a745-478c-9ff7-c8c37afe2ce4] Reconfigured VM instance instance-00000062 to attach disk [datastore1] 039a6b01-a745-478c-9ff7-c8c37afe2ce4/039a6b01-a745-478c-9ff7-c8c37afe2ce4.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1048.063164] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-67289d3b-6afc-4679-bc58-21f90d228ace {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.068467] env[61957]: DEBUG oslo_vmware.api [None req-7e98aea5-9e6c-437d-b311-a1ce6e8e055a tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Waiting for the task: (returnval){ [ 1048.068467] env[61957]: value = "task-1278037" [ 1048.068467] env[61957]: _type = "Task" [ 1048.068467] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.070069] env[61957]: DEBUG oslo_vmware.api [None req-de4650d3-4738-443f-8147-36e90c68fb19 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Waiting for the task: (returnval){ [ 1048.070069] env[61957]: value = "task-1278038" [ 1048.070069] env[61957]: _type = "Task" [ 1048.070069] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.083538] env[61957]: DEBUG oslo_vmware.api [None req-7e98aea5-9e6c-437d-b311-a1ce6e8e055a tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Task: {'id': task-1278037, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.087835] env[61957]: DEBUG oslo_vmware.api [None req-de4650d3-4738-443f-8147-36e90c68fb19 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Task: {'id': task-1278038, 'name': Rename_Task} progress is 6%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.365270] env[61957]: DEBUG nova.compute.utils [None req-8a0a01e3-6392-4a5c-b585-5e26a0a02629 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Using /dev/sd instead of None {{(pid=61957) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1048.366721] env[61957]: DEBUG nova.compute.manager [None req-8a0a01e3-6392-4a5c-b585-5e26a0a02629 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 4625d1dc-a621-4a33-9537-9e8cce96c914] Allocating IP information in the background. {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1048.366891] env[61957]: DEBUG nova.network.neutron [None req-8a0a01e3-6392-4a5c-b585-5e26a0a02629 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 4625d1dc-a621-4a33-9537-9e8cce96c914] allocate_for_instance() {{(pid=61957) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1048.401015] env[61957]: DEBUG oslo_vmware.api [None req-0d170cdb-983c-43d4-9998-8676a1404b75 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1278036, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.464635} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.403759] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-0d170cdb-983c-43d4-9998-8676a1404b75 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore1] 42860944-c700-43be-b291-970803fc7955/42860944-c700-43be-b291-970803fc7955.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1048.403980] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-0d170cdb-983c-43d4-9998-8676a1404b75 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 42860944-c700-43be-b291-970803fc7955] Extending root virtual disk to 1048576 {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1048.404413] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-06626454-178a-4f53-897e-dca723a054aa {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.410381] env[61957]: DEBUG oslo_vmware.api [None req-0d170cdb-983c-43d4-9998-8676a1404b75 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Waiting for the task: (returnval){ [ 1048.410381] env[61957]: value = "task-1278039" [ 1048.410381] env[61957]: _type = "Task" [ 1048.410381] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.416672] env[61957]: DEBUG nova.policy [None req-8a0a01e3-6392-4a5c-b585-5e26a0a02629 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6fd1c19dc3a44212ada44445e0919106', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8713aa35bcb24b86ad0b58ca9fc991ba', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61957) authorize /opt/stack/nova/nova/policy.py:203}} [ 1048.423190] env[61957]: DEBUG oslo_vmware.api [None req-0d170cdb-983c-43d4-9998-8676a1404b75 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1278039, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.481490] env[61957]: DEBUG oslo_concurrency.lockutils [None req-4a6d948f-b9a8-4fc6-bd2c-f6dd4bec3479 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Acquiring lock "refresh_cache-a7419b30-1539-4977-a5d9-93a49d84914a" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1048.481700] env[61957]: DEBUG oslo_concurrency.lockutils [None req-4a6d948f-b9a8-4fc6-bd2c-f6dd4bec3479 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Acquired lock "refresh_cache-a7419b30-1539-4977-a5d9-93a49d84914a" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1048.481779] env[61957]: DEBUG nova.network.neutron [None req-4a6d948f-b9a8-4fc6-bd2c-f6dd4bec3479 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: a7419b30-1539-4977-a5d9-93a49d84914a] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1048.487920] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6e9fda8-16ee-4a32-8496-8aa8d71d9705 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.496487] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f18b9631-0bb3-477f-b2f9-7d4263e1e3bf {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.545207] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-406a51c9-41ac-4302-9d38-1e31b2a1b438 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.553540] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ddb1fcd-7527-421b-9dbc-a45060cda560 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.567756] env[61957]: DEBUG nova.compute.provider_tree [None req-b6b5029a-df23-45ff-9311-ae897f587f74 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Updating inventory in ProviderTree for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1048.580929] env[61957]: DEBUG oslo_vmware.api [None req-7e98aea5-9e6c-437d-b311-a1ce6e8e055a tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Task: {'id': task-1278037, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.588021] env[61957]: DEBUG oslo_vmware.api [None req-de4650d3-4738-443f-8147-36e90c68fb19 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Task: {'id': task-1278038, 'name': Rename_Task, 'duration_secs': 0.159355} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.588021] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-de4650d3-4738-443f-8147-36e90c68fb19 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] [instance: 039a6b01-a745-478c-9ff7-c8c37afe2ce4] Powering on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1048.588021] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d1d8facb-2709-4d0f-92ca-1777dceb0fc9 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.593597] env[61957]: DEBUG oslo_vmware.api [None req-de4650d3-4738-443f-8147-36e90c68fb19 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Waiting for the task: (returnval){ [ 1048.593597] env[61957]: value = "task-1278040" [ 1048.593597] env[61957]: _type = "Task" [ 1048.593597] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.602612] env[61957]: DEBUG oslo_vmware.api [None req-de4650d3-4738-443f-8147-36e90c68fb19 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Task: {'id': task-1278040, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.728855] env[61957]: DEBUG nova.network.neutron [None req-8a0a01e3-6392-4a5c-b585-5e26a0a02629 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 4625d1dc-a621-4a33-9537-9e8cce96c914] Successfully created port: 8365f86e-8b14-47de-a6f4-297107be492e {{(pid=61957) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1048.874224] env[61957]: DEBUG nova.compute.manager [None req-8a0a01e3-6392-4a5c-b585-5e26a0a02629 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 4625d1dc-a621-4a33-9537-9e8cce96c914] Start building block device mappings for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1048.928859] env[61957]: DEBUG oslo_vmware.api [None req-0d170cdb-983c-43d4-9998-8676a1404b75 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1278039, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065773} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.930015] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-0d170cdb-983c-43d4-9998-8676a1404b75 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 42860944-c700-43be-b291-970803fc7955] Extended root virtual disk {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1048.930906] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-827c93c4-77ee-44fe-a600-2f6f8d1fec29 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.960220] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-0d170cdb-983c-43d4-9998-8676a1404b75 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 42860944-c700-43be-b291-970803fc7955] Reconfiguring VM instance instance-00000063 to attach disk [datastore1] 42860944-c700-43be-b291-970803fc7955/42860944-c700-43be-b291-970803fc7955.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1048.961078] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d72fa73b-fb99-4e67-8d1f-04aaa8202d5f {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.985864] env[61957]: DEBUG oslo_vmware.api [None req-0d170cdb-983c-43d4-9998-8676a1404b75 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Waiting for the task: (returnval){ [ 1048.985864] env[61957]: value = "task-1278041" [ 1048.985864] env[61957]: _type = "Task" [ 1048.985864] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.997188] env[61957]: DEBUG oslo_vmware.api [None req-0d170cdb-983c-43d4-9998-8676a1404b75 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1278041, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.033406] env[61957]: DEBUG nova.network.neutron [None req-4a6d948f-b9a8-4fc6-bd2c-f6dd4bec3479 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: a7419b30-1539-4977-a5d9-93a49d84914a] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1049.084936] env[61957]: DEBUG oslo_vmware.api [None req-7e98aea5-9e6c-437d-b311-a1ce6e8e055a tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Task: {'id': task-1278037, 'name': PowerOnVM_Task, 'duration_secs': 0.564806} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1049.085270] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-7e98aea5-9e6c-437d-b311-a1ce6e8e055a tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] [instance: 9b0053fe-420e-4b92-86a8-ed44f6a6ec49] Resumed the VM {{(pid=61957) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1189}} [ 1049.085478] env[61957]: DEBUG nova.compute.manager [None req-7e98aea5-9e6c-437d-b311-a1ce6e8e055a tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] [instance: 9b0053fe-420e-4b92-86a8-ed44f6a6ec49] Checking state {{(pid=61957) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1049.086308] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5641576-0223-44d8-a6e5-90e1c5fe85af {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.097109] env[61957]: ERROR nova.scheduler.client.report [None req-b6b5029a-df23-45ff-9311-ae897f587f74 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [req-8587997c-a7f8-4a2c-ae48-f198fa753346] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 0ceb6c9e-61c6-496d-8579-9d32627e96da. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-8587997c-a7f8-4a2c-ae48-f198fa753346"}]} [ 1049.108967] env[61957]: DEBUG oslo_vmware.api [None req-de4650d3-4738-443f-8147-36e90c68fb19 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Task: {'id': task-1278040, 'name': PowerOnVM_Task, 'duration_secs': 0.50847} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1049.109288] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-de4650d3-4738-443f-8147-36e90c68fb19 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] [instance: 039a6b01-a745-478c-9ff7-c8c37afe2ce4] Powered on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1049.109521] env[61957]: INFO nova.compute.manager [None req-de4650d3-4738-443f-8147-36e90c68fb19 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] [instance: 039a6b01-a745-478c-9ff7-c8c37afe2ce4] Took 7.71 seconds to spawn the instance on the hypervisor. [ 1049.109708] env[61957]: DEBUG nova.compute.manager [None req-de4650d3-4738-443f-8147-36e90c68fb19 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] [instance: 039a6b01-a745-478c-9ff7-c8c37afe2ce4] Checking state {{(pid=61957) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1049.110509] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79edaf79-3bff-45d4-86f0-0a8c1ff5f731 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.122178] env[61957]: DEBUG nova.scheduler.client.report [None req-b6b5029a-df23-45ff-9311-ae897f587f74 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Refreshing inventories for resource provider 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 1049.136580] env[61957]: DEBUG nova.scheduler.client.report [None req-b6b5029a-df23-45ff-9311-ae897f587f74 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Updating ProviderTree inventory for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 1049.136866] env[61957]: DEBUG nova.compute.provider_tree [None req-b6b5029a-df23-45ff-9311-ae897f587f74 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Updating inventory in ProviderTree for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1049.151231] env[61957]: DEBUG nova.scheduler.client.report [None req-b6b5029a-df23-45ff-9311-ae897f587f74 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Refreshing aggregate associations for resource provider 0ceb6c9e-61c6-496d-8579-9d32627e96da, aggregates: None {{(pid=61957) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 1049.170034] env[61957]: DEBUG nova.scheduler.client.report [None req-b6b5029a-df23-45ff-9311-ae897f587f74 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Refreshing trait associations for resource provider 0ceb6c9e-61c6-496d-8579-9d32627e96da, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO,HW_ARCH_X86_64,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=61957) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 1049.202585] env[61957]: DEBUG nova.network.neutron [None req-4a6d948f-b9a8-4fc6-bd2c-f6dd4bec3479 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: a7419b30-1539-4977-a5d9-93a49d84914a] Updating instance_info_cache with network_info: [{"id": "6bfe39fb-7cc2-4933-bdf0-f8b2a121aab2", "address": "fa:16:3e:79:7a:69", "network": {"id": "6fbccc17-6674-4cd5-8f1b-aa73570baac7", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-394544811-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1dbdc699741a48af93c52068d88a5357", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "20e3f794-c7a3-4696-9488-ecf34c570ef9", "external-id": "nsx-vlan-transportzone-509", "segmentation_id": 509, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6bfe39fb-7c", "ovs_interfaceid": "6bfe39fb-7cc2-4933-bdf0-f8b2a121aab2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1049.279852] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc53b007-36a1-4f89-8f30-be0070c6025e {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.288855] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f4a853a-d7e3-4ffc-aa1a-c183840377dc {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.318347] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3be6f262-8c44-474e-9da3-5f1b6c6553a9 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.325996] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-878dd15d-394d-4a9b-8ec1-aa55ab3d330c {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.339531] env[61957]: DEBUG nova.compute.provider_tree [None req-b6b5029a-df23-45ff-9311-ae897f587f74 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Updating inventory in ProviderTree for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1049.496158] env[61957]: DEBUG oslo_vmware.api [None req-0d170cdb-983c-43d4-9998-8676a1404b75 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1278041, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.632306] env[61957]: INFO nova.compute.manager [None req-de4650d3-4738-443f-8147-36e90c68fb19 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] [instance: 039a6b01-a745-478c-9ff7-c8c37afe2ce4] Took 16.51 seconds to build instance. [ 1049.707091] env[61957]: DEBUG oslo_concurrency.lockutils [None req-4a6d948f-b9a8-4fc6-bd2c-f6dd4bec3479 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Releasing lock "refresh_cache-a7419b30-1539-4977-a5d9-93a49d84914a" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1049.707394] env[61957]: DEBUG nova.compute.manager [None req-4a6d948f-b9a8-4fc6-bd2c-f6dd4bec3479 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: a7419b30-1539-4977-a5d9-93a49d84914a] Instance network_info: |[{"id": "6bfe39fb-7cc2-4933-bdf0-f8b2a121aab2", "address": "fa:16:3e:79:7a:69", "network": {"id": "6fbccc17-6674-4cd5-8f1b-aa73570baac7", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-394544811-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1dbdc699741a48af93c52068d88a5357", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "20e3f794-c7a3-4696-9488-ecf34c570ef9", "external-id": "nsx-vlan-transportzone-509", "segmentation_id": 509, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6bfe39fb-7c", "ovs_interfaceid": "6bfe39fb-7cc2-4933-bdf0-f8b2a121aab2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1049.707831] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-4a6d948f-b9a8-4fc6-bd2c-f6dd4bec3479 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: a7419b30-1539-4977-a5d9-93a49d84914a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:79:7a:69', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '20e3f794-c7a3-4696-9488-ecf34c570ef9', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6bfe39fb-7cc2-4933-bdf0-f8b2a121aab2', 'vif_model': 'vmxnet3'}] {{(pid=61957) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1049.715576] env[61957]: DEBUG oslo.service.loopingcall [None req-4a6d948f-b9a8-4fc6-bd2c-f6dd4bec3479 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1049.716120] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a7419b30-1539-4977-a5d9-93a49d84914a] Creating VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1049.716403] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2c514745-61b0-44ab-80ca-5068569944b3 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.736333] env[61957]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1049.736333] env[61957]: value = "task-1278042" [ 1049.736333] env[61957]: _type = "Task" [ 1049.736333] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.744348] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1278042, 'name': CreateVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.876105] env[61957]: DEBUG nova.scheduler.client.report [None req-b6b5029a-df23-45ff-9311-ae897f587f74 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Updated inventory for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da with generation 121 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 1049.876401] env[61957]: DEBUG nova.compute.provider_tree [None req-b6b5029a-df23-45ff-9311-ae897f587f74 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Updating resource provider 0ceb6c9e-61c6-496d-8579-9d32627e96da generation from 121 to 122 during operation: update_inventory {{(pid=61957) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1049.876737] env[61957]: DEBUG nova.compute.provider_tree [None req-b6b5029a-df23-45ff-9311-ae897f587f74 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Updating inventory in ProviderTree for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1049.885457] env[61957]: DEBUG nova.compute.manager [None req-8a0a01e3-6392-4a5c-b585-5e26a0a02629 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 4625d1dc-a621-4a33-9537-9e8cce96c914] Start spawning the instance on the hypervisor. {{(pid=61957) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1049.915321] env[61957]: DEBUG nova.virt.hardware [None req-8a0a01e3-6392-4a5c-b585-5e26a0a02629 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T17:22:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T17:21:50Z,direct_url=,disk_format='vmdk',id=11c76a2c-f705-470a-ba9d-4657858bab38,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='81d3096d0f094373913fc3dc8f5c3c6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T17:21:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1049.915601] env[61957]: DEBUG nova.virt.hardware [None req-8a0a01e3-6392-4a5c-b585-5e26a0a02629 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Flavor limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1049.915778] env[61957]: DEBUG nova.virt.hardware [None req-8a0a01e3-6392-4a5c-b585-5e26a0a02629 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Image limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1049.915970] env[61957]: DEBUG nova.virt.hardware [None req-8a0a01e3-6392-4a5c-b585-5e26a0a02629 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Flavor pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1049.916166] env[61957]: DEBUG nova.virt.hardware [None req-8a0a01e3-6392-4a5c-b585-5e26a0a02629 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Image pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1049.916326] env[61957]: DEBUG nova.virt.hardware [None req-8a0a01e3-6392-4a5c-b585-5e26a0a02629 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1049.916539] env[61957]: DEBUG nova.virt.hardware [None req-8a0a01e3-6392-4a5c-b585-5e26a0a02629 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1049.916699] env[61957]: DEBUG nova.virt.hardware [None req-8a0a01e3-6392-4a5c-b585-5e26a0a02629 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1049.916866] env[61957]: DEBUG nova.virt.hardware [None req-8a0a01e3-6392-4a5c-b585-5e26a0a02629 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Got 1 possible topologies {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1049.917040] env[61957]: DEBUG nova.virt.hardware [None req-8a0a01e3-6392-4a5c-b585-5e26a0a02629 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1049.917223] env[61957]: DEBUG nova.virt.hardware [None req-8a0a01e3-6392-4a5c-b585-5e26a0a02629 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1049.918312] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7664c10-881b-485e-8c6a-d275a0f364e6 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.928103] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea34baac-75f5-4006-ac6c-ffad1ff32f53 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.996087] env[61957]: DEBUG oslo_vmware.api [None req-0d170cdb-983c-43d4-9998-8676a1404b75 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1278041, 'name': ReconfigVM_Task, 'duration_secs': 0.954981} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1049.996338] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-0d170cdb-983c-43d4-9998-8676a1404b75 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 42860944-c700-43be-b291-970803fc7955] Reconfigured VM instance instance-00000063 to attach disk [datastore1] 42860944-c700-43be-b291-970803fc7955/42860944-c700-43be-b291-970803fc7955.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1049.996976] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-90b790ed-46b4-48de-9c34-6f1f0a38250c {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.003796] env[61957]: DEBUG oslo_vmware.api [None req-0d170cdb-983c-43d4-9998-8676a1404b75 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Waiting for the task: (returnval){ [ 1050.003796] env[61957]: value = "task-1278043" [ 1050.003796] env[61957]: _type = "Task" [ 1050.003796] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.012435] env[61957]: DEBUG oslo_vmware.api [None req-0d170cdb-983c-43d4-9998-8676a1404b75 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1278043, 'name': Rename_Task} progress is 5%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.056918] env[61957]: DEBUG nova.compute.manager [req-939673c5-a655-41d4-8e7c-336c26266d21 req-39e4b5eb-fab8-49a9-a493-c4bfc21e7ac7 service nova] [instance: 039a6b01-a745-478c-9ff7-c8c37afe2ce4] Received event network-changed-e49b8de4-47b9-400b-a0b0-9e0a0e68dc84 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1050.057267] env[61957]: DEBUG nova.compute.manager [req-939673c5-a655-41d4-8e7c-336c26266d21 req-39e4b5eb-fab8-49a9-a493-c4bfc21e7ac7 service nova] [instance: 039a6b01-a745-478c-9ff7-c8c37afe2ce4] Refreshing instance network info cache due to event network-changed-e49b8de4-47b9-400b-a0b0-9e0a0e68dc84. {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1050.057642] env[61957]: DEBUG oslo_concurrency.lockutils [req-939673c5-a655-41d4-8e7c-336c26266d21 req-39e4b5eb-fab8-49a9-a493-c4bfc21e7ac7 service nova] Acquiring lock "refresh_cache-039a6b01-a745-478c-9ff7-c8c37afe2ce4" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1050.057888] env[61957]: DEBUG oslo_concurrency.lockutils [req-939673c5-a655-41d4-8e7c-336c26266d21 req-39e4b5eb-fab8-49a9-a493-c4bfc21e7ac7 service nova] Acquired lock "refresh_cache-039a6b01-a745-478c-9ff7-c8c37afe2ce4" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1050.058198] env[61957]: DEBUG nova.network.neutron [req-939673c5-a655-41d4-8e7c-336c26266d21 req-39e4b5eb-fab8-49a9-a493-c4bfc21e7ac7 service nova] [instance: 039a6b01-a745-478c-9ff7-c8c37afe2ce4] Refreshing network info cache for port e49b8de4-47b9-400b-a0b0-9e0a0e68dc84 {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1050.088452] env[61957]: DEBUG nova.compute.manager [req-de22d6c1-91e7-430e-834a-369eba62e4d7 req-85b1bf84-2bdd-4b3e-82e6-73273505a362 service nova] [instance: a7419b30-1539-4977-a5d9-93a49d84914a] Received event network-changed-6bfe39fb-7cc2-4933-bdf0-f8b2a121aab2 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1050.088667] env[61957]: DEBUG nova.compute.manager [req-de22d6c1-91e7-430e-834a-369eba62e4d7 req-85b1bf84-2bdd-4b3e-82e6-73273505a362 service nova] [instance: a7419b30-1539-4977-a5d9-93a49d84914a] Refreshing instance network info cache due to event network-changed-6bfe39fb-7cc2-4933-bdf0-f8b2a121aab2. {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1050.088926] env[61957]: DEBUG oslo_concurrency.lockutils [req-de22d6c1-91e7-430e-834a-369eba62e4d7 req-85b1bf84-2bdd-4b3e-82e6-73273505a362 service nova] Acquiring lock "refresh_cache-a7419b30-1539-4977-a5d9-93a49d84914a" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1050.089041] env[61957]: DEBUG oslo_concurrency.lockutils [req-de22d6c1-91e7-430e-834a-369eba62e4d7 req-85b1bf84-2bdd-4b3e-82e6-73273505a362 service nova] Acquired lock "refresh_cache-a7419b30-1539-4977-a5d9-93a49d84914a" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1050.089223] env[61957]: DEBUG nova.network.neutron [req-de22d6c1-91e7-430e-834a-369eba62e4d7 req-85b1bf84-2bdd-4b3e-82e6-73273505a362 service nova] [instance: a7419b30-1539-4977-a5d9-93a49d84914a] Refreshing network info cache for port 6bfe39fb-7cc2-4933-bdf0-f8b2a121aab2 {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1050.134126] env[61957]: DEBUG oslo_concurrency.lockutils [None req-de4650d3-4738-443f-8147-36e90c68fb19 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Lock "039a6b01-a745-478c-9ff7-c8c37afe2ce4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.022s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1050.251997] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1278042, 'name': CreateVM_Task} progress is 99%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.323992] env[61957]: DEBUG nova.network.neutron [None req-8a0a01e3-6392-4a5c-b585-5e26a0a02629 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 4625d1dc-a621-4a33-9537-9e8cce96c914] Successfully updated port: 8365f86e-8b14-47de-a6f4-297107be492e {{(pid=61957) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1050.386146] env[61957]: DEBUG oslo_concurrency.lockutils [None req-b6b5029a-df23-45ff-9311-ae897f587f74 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.524s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1050.389234] env[61957]: DEBUG oslo_concurrency.lockutils [None req-6c409e2b-c7c8-4a9a-8e7d-1343e1ae2437 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 9.038s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1050.389792] env[61957]: DEBUG nova.objects.instance [None req-6c409e2b-c7c8-4a9a-8e7d-1343e1ae2437 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Lazy-loading 'resources' on Instance uuid 736eabcf-70f8-4e1e-9830-270d9971fc9a {{(pid=61957) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1050.412232] env[61957]: INFO nova.scheduler.client.report [None req-b6b5029a-df23-45ff-9311-ae897f587f74 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Deleted allocations for instance 84cc9489-2a11-4510-8ad6-13dc41814856 [ 1050.514301] env[61957]: DEBUG oslo_vmware.api [None req-0d170cdb-983c-43d4-9998-8676a1404b75 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1278043, 'name': Rename_Task, 'duration_secs': 0.260037} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.514579] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-0d170cdb-983c-43d4-9998-8676a1404b75 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 42860944-c700-43be-b291-970803fc7955] Powering on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1050.514837] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-49c45e0e-95e8-49c7-b1b8-2c0da385cf58 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.521650] env[61957]: DEBUG oslo_vmware.api [None req-0d170cdb-983c-43d4-9998-8676a1404b75 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Waiting for the task: (returnval){ [ 1050.521650] env[61957]: value = "task-1278044" [ 1050.521650] env[61957]: _type = "Task" [ 1050.521650] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.530189] env[61957]: DEBUG oslo_vmware.api [None req-0d170cdb-983c-43d4-9998-8676a1404b75 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1278044, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.748575] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1278042, 'name': CreateVM_Task} progress is 99%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.827574] env[61957]: DEBUG oslo_concurrency.lockutils [None req-8a0a01e3-6392-4a5c-b585-5e26a0a02629 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Acquiring lock "refresh_cache-4625d1dc-a621-4a33-9537-9e8cce96c914" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1050.827574] env[61957]: DEBUG oslo_concurrency.lockutils [None req-8a0a01e3-6392-4a5c-b585-5e26a0a02629 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Acquired lock "refresh_cache-4625d1dc-a621-4a33-9537-9e8cce96c914" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1050.827574] env[61957]: DEBUG nova.network.neutron [None req-8a0a01e3-6392-4a5c-b585-5e26a0a02629 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 4625d1dc-a621-4a33-9537-9e8cce96c914] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1050.916062] env[61957]: DEBUG nova.network.neutron [req-939673c5-a655-41d4-8e7c-336c26266d21 req-39e4b5eb-fab8-49a9-a493-c4bfc21e7ac7 service nova] [instance: 039a6b01-a745-478c-9ff7-c8c37afe2ce4] Updated VIF entry in instance network info cache for port e49b8de4-47b9-400b-a0b0-9e0a0e68dc84. {{(pid=61957) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1050.916466] env[61957]: DEBUG nova.network.neutron [req-939673c5-a655-41d4-8e7c-336c26266d21 req-39e4b5eb-fab8-49a9-a493-c4bfc21e7ac7 service nova] [instance: 039a6b01-a745-478c-9ff7-c8c37afe2ce4] Updating instance_info_cache with network_info: [{"id": "e49b8de4-47b9-400b-a0b0-9e0a0e68dc84", "address": "fa:16:3e:c2:c7:e5", "network": {"id": "8e60ab13-6dd3-4cda-9cdf-974ad8bebe1f", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1404238496-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.245", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ea955bd4d84548ff8e6a0c21542d6ee7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9f208df-1fb5-4403-9796-7fd19e4bfb85", "external-id": "cl2-zone-400", "segmentation_id": 400, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape49b8de4-47", "ovs_interfaceid": "e49b8de4-47b9-400b-a0b0-9e0a0e68dc84", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1050.919701] env[61957]: DEBUG oslo_concurrency.lockutils [None req-b6b5029a-df23-45ff-9311-ae897f587f74 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Lock "84cc9489-2a11-4510-8ad6-13dc41814856" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 12.788s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1050.936815] env[61957]: DEBUG nova.network.neutron [req-de22d6c1-91e7-430e-834a-369eba62e4d7 req-85b1bf84-2bdd-4b3e-82e6-73273505a362 service nova] [instance: a7419b30-1539-4977-a5d9-93a49d84914a] Updated VIF entry in instance network info cache for port 6bfe39fb-7cc2-4933-bdf0-f8b2a121aab2. {{(pid=61957) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1050.937176] env[61957]: DEBUG nova.network.neutron [req-de22d6c1-91e7-430e-834a-369eba62e4d7 req-85b1bf84-2bdd-4b3e-82e6-73273505a362 service nova] [instance: a7419b30-1539-4977-a5d9-93a49d84914a] Updating instance_info_cache with network_info: [{"id": "6bfe39fb-7cc2-4933-bdf0-f8b2a121aab2", "address": "fa:16:3e:79:7a:69", "network": {"id": "6fbccc17-6674-4cd5-8f1b-aa73570baac7", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-394544811-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1dbdc699741a48af93c52068d88a5357", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "20e3f794-c7a3-4696-9488-ecf34c570ef9", "external-id": "nsx-vlan-transportzone-509", "segmentation_id": 509, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6bfe39fb-7c", "ovs_interfaceid": "6bfe39fb-7cc2-4933-bdf0-f8b2a121aab2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1051.001961] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57e6b0eb-32e2-42a7-ac6e-acc7b776195a {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.010211] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cff65e98-1480-46c0-8dc6-a60676c21588 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.048503] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92aa9887-688f-4789-976c-a7ae8fe7636b {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.058848] env[61957]: DEBUG oslo_vmware.api [None req-0d170cdb-983c-43d4-9998-8676a1404b75 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1278044, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.060108] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7998fe68-146e-4d59-b316-c012c48c3c26 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.074488] env[61957]: DEBUG nova.compute.provider_tree [None req-6c409e2b-c7c8-4a9a-8e7d-1343e1ae2437 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1051.249875] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1278042, 'name': CreateVM_Task, 'duration_secs': 1.48705} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1051.250161] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a7419b30-1539-4977-a5d9-93a49d84914a] Created VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1051.250991] env[61957]: DEBUG oslo_concurrency.lockutils [None req-4a6d948f-b9a8-4fc6-bd2c-f6dd4bec3479 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1051.251264] env[61957]: DEBUG oslo_concurrency.lockutils [None req-4a6d948f-b9a8-4fc6-bd2c-f6dd4bec3479 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Acquired lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1051.251657] env[61957]: DEBUG oslo_concurrency.lockutils [None req-4a6d948f-b9a8-4fc6-bd2c-f6dd4bec3479 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1051.251980] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-39b18592-bb56-4d69-9052-05b9abe486b5 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.256638] env[61957]: DEBUG oslo_vmware.api [None req-4a6d948f-b9a8-4fc6-bd2c-f6dd4bec3479 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Waiting for the task: (returnval){ [ 1051.256638] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]529e1df8-68e0-051e-fe7a-410f87bf6484" [ 1051.256638] env[61957]: _type = "Task" [ 1051.256638] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1051.264656] env[61957]: DEBUG oslo_vmware.api [None req-4a6d948f-b9a8-4fc6-bd2c-f6dd4bec3479 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]529e1df8-68e0-051e-fe7a-410f87bf6484, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.364686] env[61957]: DEBUG nova.network.neutron [None req-8a0a01e3-6392-4a5c-b585-5e26a0a02629 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 4625d1dc-a621-4a33-9537-9e8cce96c914] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1051.421088] env[61957]: DEBUG oslo_concurrency.lockutils [req-939673c5-a655-41d4-8e7c-336c26266d21 req-39e4b5eb-fab8-49a9-a493-c4bfc21e7ac7 service nova] Releasing lock "refresh_cache-039a6b01-a745-478c-9ff7-c8c37afe2ce4" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1051.440322] env[61957]: DEBUG oslo_concurrency.lockutils [req-de22d6c1-91e7-430e-834a-369eba62e4d7 req-85b1bf84-2bdd-4b3e-82e6-73273505a362 service nova] Releasing lock "refresh_cache-a7419b30-1539-4977-a5d9-93a49d84914a" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1051.519341] env[61957]: DEBUG nova.network.neutron [None req-8a0a01e3-6392-4a5c-b585-5e26a0a02629 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 4625d1dc-a621-4a33-9537-9e8cce96c914] Updating instance_info_cache with network_info: [{"id": "8365f86e-8b14-47de-a6f4-297107be492e", "address": "fa:16:3e:e8:59:9d", "network": {"id": "364cabcb-046d-4f91-b1ee-5dd3adfae6d7", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1432396967-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8713aa35bcb24b86ad0b58ca9fc991ba", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b7a73c01-1bb9-4612-a1a7-16d71b732e81", "external-id": "nsx-vlan-transportzone-711", "segmentation_id": 711, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8365f86e-8b", "ovs_interfaceid": "8365f86e-8b14-47de-a6f4-297107be492e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1051.554090] env[61957]: DEBUG oslo_vmware.api [None req-0d170cdb-983c-43d4-9998-8676a1404b75 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1278044, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.578521] env[61957]: DEBUG nova.scheduler.client.report [None req-6c409e2b-c7c8-4a9a-8e7d-1343e1ae2437 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1051.767496] env[61957]: DEBUG oslo_vmware.api [None req-4a6d948f-b9a8-4fc6-bd2c-f6dd4bec3479 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]529e1df8-68e0-051e-fe7a-410f87bf6484, 'name': SearchDatastore_Task, 'duration_secs': 0.0089} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1051.767957] env[61957]: DEBUG oslo_concurrency.lockutils [None req-4a6d948f-b9a8-4fc6-bd2c-f6dd4bec3479 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Releasing lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1051.768289] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-4a6d948f-b9a8-4fc6-bd2c-f6dd4bec3479 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: a7419b30-1539-4977-a5d9-93a49d84914a] Processing image 11c76a2c-f705-470a-ba9d-4657858bab38 {{(pid=61957) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1051.768602] env[61957]: DEBUG oslo_concurrency.lockutils [None req-4a6d948f-b9a8-4fc6-bd2c-f6dd4bec3479 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1051.768828] env[61957]: DEBUG oslo_concurrency.lockutils [None req-4a6d948f-b9a8-4fc6-bd2c-f6dd4bec3479 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Acquired lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1051.769110] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-4a6d948f-b9a8-4fc6-bd2c-f6dd4bec3479 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1051.773019] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8f758540-3a09-464f-aa73-9973d2adb970 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.779386] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-4a6d948f-b9a8-4fc6-bd2c-f6dd4bec3479 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1051.779701] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-4a6d948f-b9a8-4fc6-bd2c-f6dd4bec3479 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61957) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1051.780958] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6317ca5c-f0f7-47cc-abcc-490d9151d91b {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.789352] env[61957]: DEBUG oslo_vmware.api [None req-4a6d948f-b9a8-4fc6-bd2c-f6dd4bec3479 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Waiting for the task: (returnval){ [ 1051.789352] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52c47aa4-5e05-b90a-a802-d17802bb4b3c" [ 1051.789352] env[61957]: _type = "Task" [ 1051.789352] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1051.800869] env[61957]: DEBUG oslo_vmware.api [None req-4a6d948f-b9a8-4fc6-bd2c-f6dd4bec3479 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52c47aa4-5e05-b90a-a802-d17802bb4b3c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.022757] env[61957]: DEBUG oslo_concurrency.lockutils [None req-8a0a01e3-6392-4a5c-b585-5e26a0a02629 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Releasing lock "refresh_cache-4625d1dc-a621-4a33-9537-9e8cce96c914" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1052.023128] env[61957]: DEBUG nova.compute.manager [None req-8a0a01e3-6392-4a5c-b585-5e26a0a02629 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 4625d1dc-a621-4a33-9537-9e8cce96c914] Instance network_info: |[{"id": "8365f86e-8b14-47de-a6f4-297107be492e", "address": "fa:16:3e:e8:59:9d", "network": {"id": "364cabcb-046d-4f91-b1ee-5dd3adfae6d7", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1432396967-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8713aa35bcb24b86ad0b58ca9fc991ba", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b7a73c01-1bb9-4612-a1a7-16d71b732e81", "external-id": "nsx-vlan-transportzone-711", "segmentation_id": 711, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8365f86e-8b", "ovs_interfaceid": "8365f86e-8b14-47de-a6f4-297107be492e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1052.023674] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-8a0a01e3-6392-4a5c-b585-5e26a0a02629 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 4625d1dc-a621-4a33-9537-9e8cce96c914] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e8:59:9d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'b7a73c01-1bb9-4612-a1a7-16d71b732e81', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8365f86e-8b14-47de-a6f4-297107be492e', 'vif_model': 'vmxnet3'}] {{(pid=61957) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1052.031463] env[61957]: DEBUG oslo.service.loopingcall [None req-8a0a01e3-6392-4a5c-b585-5e26a0a02629 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1052.031750] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4625d1dc-a621-4a33-9537-9e8cce96c914] Creating VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1052.032057] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b78b7224-d010-4938-9307-3890dbc7b91b {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.055686] env[61957]: DEBUG oslo_vmware.api [None req-0d170cdb-983c-43d4-9998-8676a1404b75 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1278044, 'name': PowerOnVM_Task, 'duration_secs': 1.098218} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1052.057195] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-0d170cdb-983c-43d4-9998-8676a1404b75 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 42860944-c700-43be-b291-970803fc7955] Powered on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1052.057442] env[61957]: INFO nova.compute.manager [None req-0d170cdb-983c-43d4-9998-8676a1404b75 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 42860944-c700-43be-b291-970803fc7955] Took 8.42 seconds to spawn the instance on the hypervisor. [ 1052.057647] env[61957]: DEBUG nova.compute.manager [None req-0d170cdb-983c-43d4-9998-8676a1404b75 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 42860944-c700-43be-b291-970803fc7955] Checking state {{(pid=61957) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1052.057924] env[61957]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1052.057924] env[61957]: value = "task-1278045" [ 1052.057924] env[61957]: _type = "Task" [ 1052.057924] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1052.058631] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e30c1b1a-007e-4904-b0dc-824d40c6b2fa {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.069859] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1278045, 'name': CreateVM_Task} progress is 5%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.078807] env[61957]: DEBUG oslo_concurrency.lockutils [None req-bfdcb579-7ae1-4a05-9d08-d1a10d0fec8a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Acquiring lock "11d67afa-1d96-4a9b-8439-383d586d8e17" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1052.078807] env[61957]: DEBUG oslo_concurrency.lockutils [None req-bfdcb579-7ae1-4a05-9d08-d1a10d0fec8a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Lock "11d67afa-1d96-4a9b-8439-383d586d8e17" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1052.083853] env[61957]: DEBUG oslo_concurrency.lockutils [None req-6c409e2b-c7c8-4a9a-8e7d-1343e1ae2437 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.695s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1052.086112] env[61957]: DEBUG oslo_concurrency.lockutils [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 6.457s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1052.137818] env[61957]: DEBUG nova.compute.manager [req-3affb249-3b87-49c5-8252-32a12aa18852 req-4ffdee71-aca6-41fd-ad5a-2cf4f81fa59b service nova] [instance: 4625d1dc-a621-4a33-9537-9e8cce96c914] Received event network-vif-plugged-8365f86e-8b14-47de-a6f4-297107be492e {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1052.138067] env[61957]: DEBUG oslo_concurrency.lockutils [req-3affb249-3b87-49c5-8252-32a12aa18852 req-4ffdee71-aca6-41fd-ad5a-2cf4f81fa59b service nova] Acquiring lock "4625d1dc-a621-4a33-9537-9e8cce96c914-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1052.138398] env[61957]: DEBUG oslo_concurrency.lockutils [req-3affb249-3b87-49c5-8252-32a12aa18852 req-4ffdee71-aca6-41fd-ad5a-2cf4f81fa59b service nova] Lock "4625d1dc-a621-4a33-9537-9e8cce96c914-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1052.138536] env[61957]: DEBUG oslo_concurrency.lockutils [req-3affb249-3b87-49c5-8252-32a12aa18852 req-4ffdee71-aca6-41fd-ad5a-2cf4f81fa59b service nova] Lock "4625d1dc-a621-4a33-9537-9e8cce96c914-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1052.138626] env[61957]: DEBUG nova.compute.manager [req-3affb249-3b87-49c5-8252-32a12aa18852 req-4ffdee71-aca6-41fd-ad5a-2cf4f81fa59b service nova] [instance: 4625d1dc-a621-4a33-9537-9e8cce96c914] No waiting events found dispatching network-vif-plugged-8365f86e-8b14-47de-a6f4-297107be492e {{(pid=61957) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1052.138836] env[61957]: WARNING nova.compute.manager [req-3affb249-3b87-49c5-8252-32a12aa18852 req-4ffdee71-aca6-41fd-ad5a-2cf4f81fa59b service nova] [instance: 4625d1dc-a621-4a33-9537-9e8cce96c914] Received unexpected event network-vif-plugged-8365f86e-8b14-47de-a6f4-297107be492e for instance with vm_state building and task_state spawning. [ 1052.138945] env[61957]: DEBUG nova.compute.manager [req-3affb249-3b87-49c5-8252-32a12aa18852 req-4ffdee71-aca6-41fd-ad5a-2cf4f81fa59b service nova] [instance: 4625d1dc-a621-4a33-9537-9e8cce96c914] Received event network-changed-8365f86e-8b14-47de-a6f4-297107be492e {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1052.139165] env[61957]: DEBUG nova.compute.manager [req-3affb249-3b87-49c5-8252-32a12aa18852 req-4ffdee71-aca6-41fd-ad5a-2cf4f81fa59b service nova] [instance: 4625d1dc-a621-4a33-9537-9e8cce96c914] Refreshing instance network info cache due to event network-changed-8365f86e-8b14-47de-a6f4-297107be492e. {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1052.139297] env[61957]: DEBUG oslo_concurrency.lockutils [req-3affb249-3b87-49c5-8252-32a12aa18852 req-4ffdee71-aca6-41fd-ad5a-2cf4f81fa59b service nova] Acquiring lock "refresh_cache-4625d1dc-a621-4a33-9537-9e8cce96c914" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1052.139564] env[61957]: DEBUG oslo_concurrency.lockutils [req-3affb249-3b87-49c5-8252-32a12aa18852 req-4ffdee71-aca6-41fd-ad5a-2cf4f81fa59b service nova] Acquired lock "refresh_cache-4625d1dc-a621-4a33-9537-9e8cce96c914" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1052.139634] env[61957]: DEBUG nova.network.neutron [req-3affb249-3b87-49c5-8252-32a12aa18852 req-4ffdee71-aca6-41fd-ad5a-2cf4f81fa59b service nova] [instance: 4625d1dc-a621-4a33-9537-9e8cce96c914] Refreshing network info cache for port 8365f86e-8b14-47de-a6f4-297107be492e {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1052.145537] env[61957]: INFO nova.scheduler.client.report [None req-6c409e2b-c7c8-4a9a-8e7d-1343e1ae2437 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Deleted allocations for instance 736eabcf-70f8-4e1e-9830-270d9971fc9a [ 1052.301453] env[61957]: DEBUG oslo_vmware.api [None req-4a6d948f-b9a8-4fc6-bd2c-f6dd4bec3479 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52c47aa4-5e05-b90a-a802-d17802bb4b3c, 'name': SearchDatastore_Task, 'duration_secs': 0.008804} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1052.302319] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0b4135a7-7172-4667-af15-2cb8ff62f30f {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.307859] env[61957]: DEBUG oslo_vmware.api [None req-4a6d948f-b9a8-4fc6-bd2c-f6dd4bec3479 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Waiting for the task: (returnval){ [ 1052.307859] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]522aadb2-9710-64b9-0cf4-b309d35ad54f" [ 1052.307859] env[61957]: _type = "Task" [ 1052.307859] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1052.316460] env[61957]: DEBUG oslo_vmware.api [None req-4a6d948f-b9a8-4fc6-bd2c-f6dd4bec3479 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]522aadb2-9710-64b9-0cf4-b309d35ad54f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.571222] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1278045, 'name': CreateVM_Task, 'duration_secs': 0.407879} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1052.571499] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4625d1dc-a621-4a33-9537-9e8cce96c914] Created VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1052.572124] env[61957]: DEBUG oslo_concurrency.lockutils [None req-8a0a01e3-6392-4a5c-b585-5e26a0a02629 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1052.572298] env[61957]: DEBUG oslo_concurrency.lockutils [None req-8a0a01e3-6392-4a5c-b585-5e26a0a02629 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Acquired lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1052.572667] env[61957]: DEBUG oslo_concurrency.lockutils [None req-8a0a01e3-6392-4a5c-b585-5e26a0a02629 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1052.572926] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-981d5514-969c-4cb2-9c4d-b46d48c44c1f {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.578234] env[61957]: DEBUG oslo_vmware.api [None req-8a0a01e3-6392-4a5c-b585-5e26a0a02629 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Waiting for the task: (returnval){ [ 1052.578234] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52c71144-aec0-b557-2293-c154a17b39a8" [ 1052.578234] env[61957]: _type = "Task" [ 1052.578234] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1052.586719] env[61957]: DEBUG nova.compute.manager [None req-bfdcb579-7ae1-4a05-9d08-d1a10d0fec8a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 11d67afa-1d96-4a9b-8439-383d586d8e17] Starting instance... {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1052.594884] env[61957]: INFO nova.compute.manager [None req-0d170cdb-983c-43d4-9998-8676a1404b75 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 42860944-c700-43be-b291-970803fc7955] Took 18.58 seconds to build instance. [ 1052.599549] env[61957]: DEBUG oslo_vmware.api [None req-8a0a01e3-6392-4a5c-b585-5e26a0a02629 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52c71144-aec0-b557-2293-c154a17b39a8, 'name': SearchDatastore_Task, 'duration_secs': 0.012881} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1052.600132] env[61957]: DEBUG oslo_concurrency.lockutils [None req-8a0a01e3-6392-4a5c-b585-5e26a0a02629 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Releasing lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1052.600378] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-8a0a01e3-6392-4a5c-b585-5e26a0a02629 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 4625d1dc-a621-4a33-9537-9e8cce96c914] Processing image 11c76a2c-f705-470a-ba9d-4657858bab38 {{(pid=61957) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1052.600601] env[61957]: DEBUG oslo_concurrency.lockutils [None req-8a0a01e3-6392-4a5c-b585-5e26a0a02629 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1052.651044] env[61957]: DEBUG oslo_concurrency.lockutils [None req-6c409e2b-c7c8-4a9a-8e7d-1343e1ae2437 tempest-AttachVolumeShelveTestJSON-525518302 tempest-AttachVolumeShelveTestJSON-525518302-project-member] Lock "736eabcf-70f8-4e1e-9830-270d9971fc9a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 15.247s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1052.835898] env[61957]: DEBUG oslo_vmware.api [None req-4a6d948f-b9a8-4fc6-bd2c-f6dd4bec3479 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]522aadb2-9710-64b9-0cf4-b309d35ad54f, 'name': SearchDatastore_Task, 'duration_secs': 0.012052} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1052.839934] env[61957]: DEBUG oslo_concurrency.lockutils [None req-4a6d948f-b9a8-4fc6-bd2c-f6dd4bec3479 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Releasing lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1052.839934] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-4a6d948f-b9a8-4fc6-bd2c-f6dd4bec3479 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore1] a7419b30-1539-4977-a5d9-93a49d84914a/a7419b30-1539-4977-a5d9-93a49d84914a.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1052.839934] env[61957]: DEBUG oslo_concurrency.lockutils [None req-8a0a01e3-6392-4a5c-b585-5e26a0a02629 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Acquired lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1052.839934] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-8a0a01e3-6392-4a5c-b585-5e26a0a02629 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1052.839934] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ac1babe7-b56a-482a-96c4-7c34ada9eaa1 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.844317] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a3729b02-73b4-4a0f-addc-a0f271880142 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.853501] env[61957]: DEBUG oslo_vmware.api [None req-4a6d948f-b9a8-4fc6-bd2c-f6dd4bec3479 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Waiting for the task: (returnval){ [ 1052.853501] env[61957]: value = "task-1278046" [ 1052.853501] env[61957]: _type = "Task" [ 1052.853501] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1052.854725] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-8a0a01e3-6392-4a5c-b585-5e26a0a02629 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1052.854975] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-8a0a01e3-6392-4a5c-b585-5e26a0a02629 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61957) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1052.858278] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0a1b7958-51d4-4e46-9426-5c8a40091581 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.863799] env[61957]: DEBUG oslo_vmware.api [None req-8a0a01e3-6392-4a5c-b585-5e26a0a02629 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Waiting for the task: (returnval){ [ 1052.863799] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]521a7e97-18a7-d160-669f-6d0fb8c5896b" [ 1052.863799] env[61957]: _type = "Task" [ 1052.863799] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1052.866560] env[61957]: DEBUG oslo_vmware.api [None req-4a6d948f-b9a8-4fc6-bd2c-f6dd4bec3479 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': task-1278046, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.875062] env[61957]: DEBUG oslo_vmware.api [None req-8a0a01e3-6392-4a5c-b585-5e26a0a02629 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]521a7e97-18a7-d160-669f-6d0fb8c5896b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.875862] env[61957]: DEBUG nova.network.neutron [req-3affb249-3b87-49c5-8252-32a12aa18852 req-4ffdee71-aca6-41fd-ad5a-2cf4f81fa59b service nova] [instance: 4625d1dc-a621-4a33-9537-9e8cce96c914] Updated VIF entry in instance network info cache for port 8365f86e-8b14-47de-a6f4-297107be492e. {{(pid=61957) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1052.876201] env[61957]: DEBUG nova.network.neutron [req-3affb249-3b87-49c5-8252-32a12aa18852 req-4ffdee71-aca6-41fd-ad5a-2cf4f81fa59b service nova] [instance: 4625d1dc-a621-4a33-9537-9e8cce96c914] Updating instance_info_cache with network_info: [{"id": "8365f86e-8b14-47de-a6f4-297107be492e", "address": "fa:16:3e:e8:59:9d", "network": {"id": "364cabcb-046d-4f91-b1ee-5dd3adfae6d7", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1432396967-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8713aa35bcb24b86ad0b58ca9fc991ba", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b7a73c01-1bb9-4612-a1a7-16d71b732e81", "external-id": "nsx-vlan-transportzone-711", "segmentation_id": 711, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8365f86e-8b", "ovs_interfaceid": "8365f86e-8b14-47de-a6f4-297107be492e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1053.107248] env[61957]: DEBUG oslo_concurrency.lockutils [None req-0d170cdb-983c-43d4-9998-8676a1404b75 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Lock "42860944-c700-43be-b291-970803fc7955" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.094s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1053.136811] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Instance 1a0327c2-5671-4970-9db7-c7cc912d8678 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61957) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1053.136811] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Instance 9b0053fe-420e-4b92-86a8-ed44f6a6ec49 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61957) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1053.136811] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Instance 039a6b01-a745-478c-9ff7-c8c37afe2ce4 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61957) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1053.136811] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Instance 42860944-c700-43be-b291-970803fc7955 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61957) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1053.136811] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Instance a7419b30-1539-4977-a5d9-93a49d84914a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61957) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1053.136811] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Instance 4625d1dc-a621-4a33-9537-9e8cce96c914 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61957) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1053.136811] env[61957]: DEBUG oslo_concurrency.lockutils [None req-bfdcb579-7ae1-4a05-9d08-d1a10d0fec8a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1053.149955] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de6a1552-a947-4c08-beb8-7ba1f435082c {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.157628] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-688a51c1-1c5f-4ee1-b516-89ef93f5b35e tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 42860944-c700-43be-b291-970803fc7955] Suspending the VM {{(pid=61957) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1163}} [ 1053.157957] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-820070d9-1dd1-42ac-82f3-2ecbe83db0d6 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.164665] env[61957]: DEBUG oslo_vmware.api [None req-688a51c1-1c5f-4ee1-b516-89ef93f5b35e tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Waiting for the task: (returnval){ [ 1053.164665] env[61957]: value = "task-1278047" [ 1053.164665] env[61957]: _type = "Task" [ 1053.164665] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1053.174182] env[61957]: DEBUG oslo_vmware.api [None req-688a51c1-1c5f-4ee1-b516-89ef93f5b35e tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1278047, 'name': SuspendVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.367944] env[61957]: DEBUG oslo_vmware.api [None req-4a6d948f-b9a8-4fc6-bd2c-f6dd4bec3479 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': task-1278046, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.378458] env[61957]: DEBUG oslo_concurrency.lockutils [req-3affb249-3b87-49c5-8252-32a12aa18852 req-4ffdee71-aca6-41fd-ad5a-2cf4f81fa59b service nova] Releasing lock "refresh_cache-4625d1dc-a621-4a33-9537-9e8cce96c914" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1053.378996] env[61957]: DEBUG oslo_vmware.api [None req-8a0a01e3-6392-4a5c-b585-5e26a0a02629 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]521a7e97-18a7-d160-669f-6d0fb8c5896b, 'name': SearchDatastore_Task, 'duration_secs': 0.010186} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1053.379880] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ebee16f0-37f6-4ef0-90ae-654c4d0ed6c9 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.386178] env[61957]: DEBUG oslo_vmware.api [None req-8a0a01e3-6392-4a5c-b585-5e26a0a02629 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Waiting for the task: (returnval){ [ 1053.386178] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]524d5bfd-0baf-2fb6-696b-6ac1cc25dc86" [ 1053.386178] env[61957]: _type = "Task" [ 1053.386178] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1053.395547] env[61957]: DEBUG oslo_vmware.api [None req-8a0a01e3-6392-4a5c-b585-5e26a0a02629 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]524d5bfd-0baf-2fb6-696b-6ac1cc25dc86, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.638795] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Instance 11d67afa-1d96-4a9b-8439-383d586d8e17 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61957) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 1053.639034] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Total usable vcpus: 48, total allocated vcpus: 6 {{(pid=61957) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1053.639157] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1664MB phys_disk=200GB used_disk=6GB total_vcpus=48 used_vcpus=6 pci_stats=[] {{(pid=61957) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1053.675417] env[61957]: DEBUG oslo_vmware.api [None req-688a51c1-1c5f-4ee1-b516-89ef93f5b35e tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1278047, 'name': SuspendVM_Task} progress is 58%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.743479] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6979af5-b4af-4805-9716-88d70dd05f4d {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.750833] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70c5da28-7d8c-4c5a-96ba-0c3b719465f8 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.779874] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1f5bd84-f979-404a-b851-eb4e0a3dcc3b {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.786784] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efff42f3-2771-4148-ad16-99e974b79e2f {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.801321] env[61957]: DEBUG nova.compute.provider_tree [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1053.865139] env[61957]: DEBUG oslo_vmware.api [None req-4a6d948f-b9a8-4fc6-bd2c-f6dd4bec3479 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': task-1278046, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.798968} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1053.865447] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-4a6d948f-b9a8-4fc6-bd2c-f6dd4bec3479 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore1] a7419b30-1539-4977-a5d9-93a49d84914a/a7419b30-1539-4977-a5d9-93a49d84914a.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1053.865668] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-4a6d948f-b9a8-4fc6-bd2c-f6dd4bec3479 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: a7419b30-1539-4977-a5d9-93a49d84914a] Extending root virtual disk to 1048576 {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1053.865920] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-fae64b0a-2040-4564-8de0-0aa34b44e116 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.872718] env[61957]: DEBUG oslo_vmware.api [None req-4a6d948f-b9a8-4fc6-bd2c-f6dd4bec3479 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Waiting for the task: (returnval){ [ 1053.872718] env[61957]: value = "task-1278048" [ 1053.872718] env[61957]: _type = "Task" [ 1053.872718] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1053.880663] env[61957]: DEBUG oslo_vmware.api [None req-4a6d948f-b9a8-4fc6-bd2c-f6dd4bec3479 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': task-1278048, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.895906] env[61957]: DEBUG oslo_vmware.api [None req-8a0a01e3-6392-4a5c-b585-5e26a0a02629 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]524d5bfd-0baf-2fb6-696b-6ac1cc25dc86, 'name': SearchDatastore_Task, 'duration_secs': 0.047739} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1053.896210] env[61957]: DEBUG oslo_concurrency.lockutils [None req-8a0a01e3-6392-4a5c-b585-5e26a0a02629 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Releasing lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1053.896470] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-8a0a01e3-6392-4a5c-b585-5e26a0a02629 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore1] 4625d1dc-a621-4a33-9537-9e8cce96c914/4625d1dc-a621-4a33-9537-9e8cce96c914.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1053.896730] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3682759a-cee4-4a5b-b575-19cfa70d135f {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.903642] env[61957]: DEBUG oslo_vmware.api [None req-8a0a01e3-6392-4a5c-b585-5e26a0a02629 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Waiting for the task: (returnval){ [ 1053.903642] env[61957]: value = "task-1278049" [ 1053.903642] env[61957]: _type = "Task" [ 1053.903642] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1053.911446] env[61957]: DEBUG oslo_vmware.api [None req-8a0a01e3-6392-4a5c-b585-5e26a0a02629 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': task-1278049, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.176075] env[61957]: DEBUG oslo_vmware.api [None req-688a51c1-1c5f-4ee1-b516-89ef93f5b35e tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1278047, 'name': SuspendVM_Task, 'duration_secs': 0.843346} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1054.176316] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-688a51c1-1c5f-4ee1-b516-89ef93f5b35e tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 42860944-c700-43be-b291-970803fc7955] Suspended the VM {{(pid=61957) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1167}} [ 1054.176527] env[61957]: DEBUG nova.compute.manager [None req-688a51c1-1c5f-4ee1-b516-89ef93f5b35e tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 42860944-c700-43be-b291-970803fc7955] Checking state {{(pid=61957) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1054.177297] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff8a33b1-3b21-4a16-b9f7-50484f7b0608 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.304496] env[61957]: DEBUG nova.scheduler.client.report [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1054.383245] env[61957]: DEBUG oslo_vmware.api [None req-4a6d948f-b9a8-4fc6-bd2c-f6dd4bec3479 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': task-1278048, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072755} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1054.383932] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-4a6d948f-b9a8-4fc6-bd2c-f6dd4bec3479 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: a7419b30-1539-4977-a5d9-93a49d84914a] Extended root virtual disk {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1054.384954] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c5baf0f-4a24-4fe3-b472-d809ff1a921d {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.418346] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-4a6d948f-b9a8-4fc6-bd2c-f6dd4bec3479 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: a7419b30-1539-4977-a5d9-93a49d84914a] Reconfiguring VM instance instance-00000064 to attach disk [datastore1] a7419b30-1539-4977-a5d9-93a49d84914a/a7419b30-1539-4977-a5d9-93a49d84914a.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1054.421993] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-174a9184-8a85-418c-b5a3-aaa19ff9b246 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.446331] env[61957]: DEBUG oslo_vmware.api [None req-8a0a01e3-6392-4a5c-b585-5e26a0a02629 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': task-1278049, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.447747] env[61957]: DEBUG oslo_vmware.api [None req-4a6d948f-b9a8-4fc6-bd2c-f6dd4bec3479 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Waiting for the task: (returnval){ [ 1054.447747] env[61957]: value = "task-1278050" [ 1054.447747] env[61957]: _type = "Task" [ 1054.447747] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1054.455775] env[61957]: DEBUG oslo_vmware.api [None req-4a6d948f-b9a8-4fc6-bd2c-f6dd4bec3479 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': task-1278050, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.809245] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61957) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1054.809580] env[61957]: DEBUG oslo_concurrency.lockutils [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.723s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1054.809939] env[61957]: DEBUG oslo_concurrency.lockutils [None req-bfdcb579-7ae1-4a05-9d08-d1a10d0fec8a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.673s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1054.811573] env[61957]: INFO nova.compute.claims [None req-bfdcb579-7ae1-4a05-9d08-d1a10d0fec8a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 11d67afa-1d96-4a9b-8439-383d586d8e17] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1054.915315] env[61957]: DEBUG oslo_vmware.api [None req-8a0a01e3-6392-4a5c-b585-5e26a0a02629 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': task-1278049, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.879341} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1054.915614] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-8a0a01e3-6392-4a5c-b585-5e26a0a02629 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore1] 4625d1dc-a621-4a33-9537-9e8cce96c914/4625d1dc-a621-4a33-9537-9e8cce96c914.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1054.915814] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-8a0a01e3-6392-4a5c-b585-5e26a0a02629 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 4625d1dc-a621-4a33-9537-9e8cce96c914] Extending root virtual disk to 1048576 {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1054.916124] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0af71584-4961-4f9a-9d85-73c94e0b63d6 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.922146] env[61957]: DEBUG oslo_vmware.api [None req-8a0a01e3-6392-4a5c-b585-5e26a0a02629 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Waiting for the task: (returnval){ [ 1054.922146] env[61957]: value = "task-1278051" [ 1054.922146] env[61957]: _type = "Task" [ 1054.922146] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1054.930041] env[61957]: DEBUG oslo_vmware.api [None req-8a0a01e3-6392-4a5c-b585-5e26a0a02629 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': task-1278051, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.956785] env[61957]: DEBUG oslo_vmware.api [None req-4a6d948f-b9a8-4fc6-bd2c-f6dd4bec3479 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': task-1278050, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.433181] env[61957]: DEBUG oslo_vmware.api [None req-8a0a01e3-6392-4a5c-b585-5e26a0a02629 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': task-1278051, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.350136} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1055.433527] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-8a0a01e3-6392-4a5c-b585-5e26a0a02629 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 4625d1dc-a621-4a33-9537-9e8cce96c914] Extended root virtual disk {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1055.434531] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78885599-1992-45a1-b7fc-b5af13cb83ee {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.461723] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-8a0a01e3-6392-4a5c-b585-5e26a0a02629 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 4625d1dc-a621-4a33-9537-9e8cce96c914] Reconfiguring VM instance instance-00000065 to attach disk [datastore1] 4625d1dc-a621-4a33-9537-9e8cce96c914/4625d1dc-a621-4a33-9537-9e8cce96c914.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1055.464756] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2e0c11d2-2378-4523-aa1a-160a8adda063 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.485778] env[61957]: DEBUG oslo_vmware.api [None req-4a6d948f-b9a8-4fc6-bd2c-f6dd4bec3479 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': task-1278050, 'name': ReconfigVM_Task, 'duration_secs': 0.969905} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1055.487108] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-4a6d948f-b9a8-4fc6-bd2c-f6dd4bec3479 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: a7419b30-1539-4977-a5d9-93a49d84914a] Reconfigured VM instance instance-00000064 to attach disk [datastore1] a7419b30-1539-4977-a5d9-93a49d84914a/a7419b30-1539-4977-a5d9-93a49d84914a.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1055.487854] env[61957]: DEBUG oslo_vmware.api [None req-8a0a01e3-6392-4a5c-b585-5e26a0a02629 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Waiting for the task: (returnval){ [ 1055.487854] env[61957]: value = "task-1278052" [ 1055.487854] env[61957]: _type = "Task" [ 1055.487854] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1055.488079] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d14ea929-ba58-4f0d-97fe-bb6768d56afb {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.499196] env[61957]: DEBUG oslo_vmware.api [None req-8a0a01e3-6392-4a5c-b585-5e26a0a02629 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': task-1278052, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.500520] env[61957]: DEBUG oslo_vmware.api [None req-4a6d948f-b9a8-4fc6-bd2c-f6dd4bec3479 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Waiting for the task: (returnval){ [ 1055.500520] env[61957]: value = "task-1278053" [ 1055.500520] env[61957]: _type = "Task" [ 1055.500520] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1055.510704] env[61957]: DEBUG oslo_vmware.api [None req-4a6d948f-b9a8-4fc6-bd2c-f6dd4bec3479 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': task-1278053, 'name': Rename_Task} progress is 6%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.793102] env[61957]: DEBUG oslo_concurrency.lockutils [None req-71d3ee2e-25df-44b6-92cf-ec6938a2e3f1 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Acquiring lock "42860944-c700-43be-b291-970803fc7955" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1055.793428] env[61957]: DEBUG oslo_concurrency.lockutils [None req-71d3ee2e-25df-44b6-92cf-ec6938a2e3f1 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Lock "42860944-c700-43be-b291-970803fc7955" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1055.793658] env[61957]: DEBUG oslo_concurrency.lockutils [None req-71d3ee2e-25df-44b6-92cf-ec6938a2e3f1 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Acquiring lock "42860944-c700-43be-b291-970803fc7955-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1055.793914] env[61957]: DEBUG oslo_concurrency.lockutils [None req-71d3ee2e-25df-44b6-92cf-ec6938a2e3f1 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Lock "42860944-c700-43be-b291-970803fc7955-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1055.794109] env[61957]: DEBUG oslo_concurrency.lockutils [None req-71d3ee2e-25df-44b6-92cf-ec6938a2e3f1 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Lock "42860944-c700-43be-b291-970803fc7955-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1055.796443] env[61957]: INFO nova.compute.manager [None req-71d3ee2e-25df-44b6-92cf-ec6938a2e3f1 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 42860944-c700-43be-b291-970803fc7955] Terminating instance [ 1055.798782] env[61957]: DEBUG nova.compute.manager [None req-71d3ee2e-25df-44b6-92cf-ec6938a2e3f1 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 42860944-c700-43be-b291-970803fc7955] Start destroying the instance on the hypervisor. {{(pid=61957) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1055.799665] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-71d3ee2e-25df-44b6-92cf-ec6938a2e3f1 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 42860944-c700-43be-b291-970803fc7955] Destroying instance {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1055.799861] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89bbc9f6-3f97-4358-a009-32e3ad112066 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.807873] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-71d3ee2e-25df-44b6-92cf-ec6938a2e3f1 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 42860944-c700-43be-b291-970803fc7955] Unregistering the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1055.808156] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-98dc77d7-3b1c-456a-8586-11b680e92998 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.899512] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-71d3ee2e-25df-44b6-92cf-ec6938a2e3f1 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 42860944-c700-43be-b291-970803fc7955] Unregistered the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1055.899752] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-71d3ee2e-25df-44b6-92cf-ec6938a2e3f1 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 42860944-c700-43be-b291-970803fc7955] Deleting contents of the VM from datastore datastore1 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1055.899934] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-71d3ee2e-25df-44b6-92cf-ec6938a2e3f1 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Deleting the datastore file [datastore1] 42860944-c700-43be-b291-970803fc7955 {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1055.900212] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b9b5260b-141a-47af-917f-4c4ff3b1be75 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.906805] env[61957]: DEBUG oslo_vmware.api [None req-71d3ee2e-25df-44b6-92cf-ec6938a2e3f1 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Waiting for the task: (returnval){ [ 1055.906805] env[61957]: value = "task-1278055" [ 1055.906805] env[61957]: _type = "Task" [ 1055.906805] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1055.919572] env[61957]: DEBUG oslo_concurrency.lockutils [None req-f56f6d3c-d57e-4ad4-8496-b310c6e6143c tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Acquiring lock "9b0053fe-420e-4b92-86a8-ed44f6a6ec49" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1055.919821] env[61957]: DEBUG oslo_concurrency.lockutils [None req-f56f6d3c-d57e-4ad4-8496-b310c6e6143c tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Lock "9b0053fe-420e-4b92-86a8-ed44f6a6ec49" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1055.920015] env[61957]: DEBUG oslo_concurrency.lockutils [None req-f56f6d3c-d57e-4ad4-8496-b310c6e6143c tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Acquiring lock "9b0053fe-420e-4b92-86a8-ed44f6a6ec49-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1055.920209] env[61957]: DEBUG oslo_concurrency.lockutils [None req-f56f6d3c-d57e-4ad4-8496-b310c6e6143c tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Lock "9b0053fe-420e-4b92-86a8-ed44f6a6ec49-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1055.920390] env[61957]: DEBUG oslo_concurrency.lockutils [None req-f56f6d3c-d57e-4ad4-8496-b310c6e6143c tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Lock "9b0053fe-420e-4b92-86a8-ed44f6a6ec49-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1055.921935] env[61957]: DEBUG oslo_vmware.api [None req-71d3ee2e-25df-44b6-92cf-ec6938a2e3f1 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1278055, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.924557] env[61957]: INFO nova.compute.manager [None req-f56f6d3c-d57e-4ad4-8496-b310c6e6143c tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] [instance: 9b0053fe-420e-4b92-86a8-ed44f6a6ec49] Terminating instance [ 1055.926328] env[61957]: DEBUG nova.compute.manager [None req-f56f6d3c-d57e-4ad4-8496-b310c6e6143c tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] [instance: 9b0053fe-420e-4b92-86a8-ed44f6a6ec49] Start destroying the instance on the hypervisor. {{(pid=61957) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1055.926524] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-f56f6d3c-d57e-4ad4-8496-b310c6e6143c tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] [instance: 9b0053fe-420e-4b92-86a8-ed44f6a6ec49] Destroying instance {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1055.927301] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ce2c983-3602-4d81-ba26-36feaa63e435 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.933624] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-f56f6d3c-d57e-4ad4-8496-b310c6e6143c tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] [instance: 9b0053fe-420e-4b92-86a8-ed44f6a6ec49] Powering off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1055.933862] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b2c0d93c-7743-4906-853c-35db6be713c2 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.936813] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d32ccaf8-31c4-4a45-945f-e4f771266f64 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.943886] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87a7a32d-102a-4317-8b97-1f8b04fe1c67 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.947973] env[61957]: DEBUG oslo_vmware.api [None req-f56f6d3c-d57e-4ad4-8496-b310c6e6143c tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Waiting for the task: (returnval){ [ 1055.947973] env[61957]: value = "task-1278056" [ 1055.947973] env[61957]: _type = "Task" [ 1055.947973] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1055.979848] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb436ca8-9d75-443b-b2d0-707e4c6f221d {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.985424] env[61957]: DEBUG oslo_vmware.api [None req-f56f6d3c-d57e-4ad4-8496-b310c6e6143c tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Task: {'id': task-1278056, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.995111] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9df7c016-a16a-45fa-bd78-16989cfad15d {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.003729] env[61957]: DEBUG oslo_vmware.api [None req-8a0a01e3-6392-4a5c-b585-5e26a0a02629 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': task-1278052, 'name': ReconfigVM_Task, 'duration_secs': 0.345634} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1056.014616] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-8a0a01e3-6392-4a5c-b585-5e26a0a02629 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 4625d1dc-a621-4a33-9537-9e8cce96c914] Reconfigured VM instance instance-00000065 to attach disk [datastore1] 4625d1dc-a621-4a33-9537-9e8cce96c914/4625d1dc-a621-4a33-9537-9e8cce96c914.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1056.015530] env[61957]: DEBUG nova.compute.provider_tree [None req-bfdcb579-7ae1-4a05-9d08-d1a10d0fec8a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1056.016853] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-035a09ed-796b-456f-bba4-9c0370dc62c4 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.026945] env[61957]: DEBUG oslo_vmware.api [None req-4a6d948f-b9a8-4fc6-bd2c-f6dd4bec3479 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': task-1278053, 'name': Rename_Task, 'duration_secs': 0.34249} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1056.026945] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-4a6d948f-b9a8-4fc6-bd2c-f6dd4bec3479 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: a7419b30-1539-4977-a5d9-93a49d84914a] Powering on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1056.026945] env[61957]: DEBUG oslo_vmware.api [None req-8a0a01e3-6392-4a5c-b585-5e26a0a02629 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Waiting for the task: (returnval){ [ 1056.026945] env[61957]: value = "task-1278057" [ 1056.026945] env[61957]: _type = "Task" [ 1056.026945] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1056.026945] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-139fc066-4c9b-4a2e-b6f3-aa7a2491d932 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.036974] env[61957]: DEBUG oslo_vmware.api [None req-8a0a01e3-6392-4a5c-b585-5e26a0a02629 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': task-1278057, 'name': Rename_Task} progress is 6%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.038388] env[61957]: DEBUG oslo_vmware.api [None req-4a6d948f-b9a8-4fc6-bd2c-f6dd4bec3479 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Waiting for the task: (returnval){ [ 1056.038388] env[61957]: value = "task-1278058" [ 1056.038388] env[61957]: _type = "Task" [ 1056.038388] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1056.047115] env[61957]: DEBUG oslo_vmware.api [None req-4a6d948f-b9a8-4fc6-bd2c-f6dd4bec3479 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': task-1278058, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.418755] env[61957]: DEBUG oslo_vmware.api [None req-71d3ee2e-25df-44b6-92cf-ec6938a2e3f1 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1278055, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.272075} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1056.419051] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-71d3ee2e-25df-44b6-92cf-ec6938a2e3f1 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Deleted the datastore file {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1056.419302] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-71d3ee2e-25df-44b6-92cf-ec6938a2e3f1 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 42860944-c700-43be-b291-970803fc7955] Deleted contents of the VM from datastore datastore1 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1056.419431] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-71d3ee2e-25df-44b6-92cf-ec6938a2e3f1 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 42860944-c700-43be-b291-970803fc7955] Instance destroyed {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1056.419613] env[61957]: INFO nova.compute.manager [None req-71d3ee2e-25df-44b6-92cf-ec6938a2e3f1 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 42860944-c700-43be-b291-970803fc7955] Took 0.62 seconds to destroy the instance on the hypervisor. [ 1056.419854] env[61957]: DEBUG oslo.service.loopingcall [None req-71d3ee2e-25df-44b6-92cf-ec6938a2e3f1 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1056.420211] env[61957]: DEBUG nova.compute.manager [-] [instance: 42860944-c700-43be-b291-970803fc7955] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1056.420211] env[61957]: DEBUG nova.network.neutron [-] [instance: 42860944-c700-43be-b291-970803fc7955] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1056.460150] env[61957]: DEBUG oslo_vmware.api [None req-f56f6d3c-d57e-4ad4-8496-b310c6e6143c tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Task: {'id': task-1278056, 'name': PowerOffVM_Task, 'duration_secs': 0.247994} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1056.460457] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-f56f6d3c-d57e-4ad4-8496-b310c6e6143c tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] [instance: 9b0053fe-420e-4b92-86a8-ed44f6a6ec49] Powered off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1056.460643] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-f56f6d3c-d57e-4ad4-8496-b310c6e6143c tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] [instance: 9b0053fe-420e-4b92-86a8-ed44f6a6ec49] Unregistering the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1056.460904] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-757398aa-9ff8-49cb-a802-21996e4ce2fa {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.520787] env[61957]: DEBUG nova.scheduler.client.report [None req-bfdcb579-7ae1-4a05-9d08-d1a10d0fec8a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1056.527151] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-f56f6d3c-d57e-4ad4-8496-b310c6e6143c tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] [instance: 9b0053fe-420e-4b92-86a8-ed44f6a6ec49] Unregistered the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1056.527151] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-f56f6d3c-d57e-4ad4-8496-b310c6e6143c tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] [instance: 9b0053fe-420e-4b92-86a8-ed44f6a6ec49] Deleting contents of the VM from datastore datastore2 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1056.527151] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-f56f6d3c-d57e-4ad4-8496-b310c6e6143c tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Deleting the datastore file [datastore2] 9b0053fe-420e-4b92-86a8-ed44f6a6ec49 {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1056.527356] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-444999c9-2876-405b-893c-c354094e7390 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.538218] env[61957]: DEBUG oslo_vmware.api [None req-8a0a01e3-6392-4a5c-b585-5e26a0a02629 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': task-1278057, 'name': Rename_Task, 'duration_secs': 0.155731} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1056.540253] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-8a0a01e3-6392-4a5c-b585-5e26a0a02629 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 4625d1dc-a621-4a33-9537-9e8cce96c914] Powering on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1056.540589] env[61957]: DEBUG oslo_vmware.api [None req-f56f6d3c-d57e-4ad4-8496-b310c6e6143c tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Waiting for the task: (returnval){ [ 1056.540589] env[61957]: value = "task-1278060" [ 1056.540589] env[61957]: _type = "Task" [ 1056.540589] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1056.541066] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-48aff4ba-e692-4619-9343-10490fb44c30 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.554305] env[61957]: DEBUG oslo_vmware.api [None req-f56f6d3c-d57e-4ad4-8496-b310c6e6143c tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Task: {'id': task-1278060, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.558032] env[61957]: DEBUG oslo_vmware.api [None req-8a0a01e3-6392-4a5c-b585-5e26a0a02629 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Waiting for the task: (returnval){ [ 1056.558032] env[61957]: value = "task-1278061" [ 1056.558032] env[61957]: _type = "Task" [ 1056.558032] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1056.558262] env[61957]: DEBUG oslo_vmware.api [None req-4a6d948f-b9a8-4fc6-bd2c-f6dd4bec3479 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': task-1278058, 'name': PowerOnVM_Task} progress is 87%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.565876] env[61957]: DEBUG oslo_vmware.api [None req-8a0a01e3-6392-4a5c-b585-5e26a0a02629 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': task-1278061, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.738893] env[61957]: DEBUG nova.compute.manager [req-e2001167-3a09-44e0-92fc-e26cf1ddcc6d req-42d53ce0-b9c5-4208-8ddc-c968f4e99f56 service nova] [instance: 42860944-c700-43be-b291-970803fc7955] Received event network-vif-deleted-0b2a4b91-7550-44b3-b4de-0fccf960ea77 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1056.739057] env[61957]: INFO nova.compute.manager [req-e2001167-3a09-44e0-92fc-e26cf1ddcc6d req-42d53ce0-b9c5-4208-8ddc-c968f4e99f56 service nova] [instance: 42860944-c700-43be-b291-970803fc7955] Neutron deleted interface 0b2a4b91-7550-44b3-b4de-0fccf960ea77; detaching it from the instance and deleting it from the info cache [ 1056.739455] env[61957]: DEBUG nova.network.neutron [req-e2001167-3a09-44e0-92fc-e26cf1ddcc6d req-42d53ce0-b9c5-4208-8ddc-c968f4e99f56 service nova] [instance: 42860944-c700-43be-b291-970803fc7955] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1057.026396] env[61957]: DEBUG oslo_concurrency.lockutils [None req-bfdcb579-7ae1-4a05-9d08-d1a10d0fec8a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.216s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1057.027193] env[61957]: DEBUG nova.compute.manager [None req-bfdcb579-7ae1-4a05-9d08-d1a10d0fec8a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 11d67afa-1d96-4a9b-8439-383d586d8e17] Start building networks asynchronously for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1057.055778] env[61957]: DEBUG oslo_vmware.api [None req-4a6d948f-b9a8-4fc6-bd2c-f6dd4bec3479 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': task-1278058, 'name': PowerOnVM_Task, 'duration_secs': 0.669535} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1057.056683] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-4a6d948f-b9a8-4fc6-bd2c-f6dd4bec3479 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: a7419b30-1539-4977-a5d9-93a49d84914a] Powered on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1057.057034] env[61957]: INFO nova.compute.manager [None req-4a6d948f-b9a8-4fc6-bd2c-f6dd4bec3479 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: a7419b30-1539-4977-a5d9-93a49d84914a] Took 9.47 seconds to spawn the instance on the hypervisor. [ 1057.057345] env[61957]: DEBUG nova.compute.manager [None req-4a6d948f-b9a8-4fc6-bd2c-f6dd4bec3479 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: a7419b30-1539-4977-a5d9-93a49d84914a] Checking state {{(pid=61957) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1057.058596] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1db38754-6b6a-4369-acb2-5657e4427c6a {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.066762] env[61957]: DEBUG oslo_vmware.api [None req-f56f6d3c-d57e-4ad4-8496-b310c6e6143c tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Task: {'id': task-1278060, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.146548} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1057.071806] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-f56f6d3c-d57e-4ad4-8496-b310c6e6143c tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Deleted the datastore file {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1057.072108] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-f56f6d3c-d57e-4ad4-8496-b310c6e6143c tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] [instance: 9b0053fe-420e-4b92-86a8-ed44f6a6ec49] Deleted contents of the VM from datastore datastore2 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1057.072369] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-f56f6d3c-d57e-4ad4-8496-b310c6e6143c tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] [instance: 9b0053fe-420e-4b92-86a8-ed44f6a6ec49] Instance destroyed {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1057.072640] env[61957]: INFO nova.compute.manager [None req-f56f6d3c-d57e-4ad4-8496-b310c6e6143c tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] [instance: 9b0053fe-420e-4b92-86a8-ed44f6a6ec49] Took 1.15 seconds to destroy the instance on the hypervisor. [ 1057.073011] env[61957]: DEBUG oslo.service.loopingcall [None req-f56f6d3c-d57e-4ad4-8496-b310c6e6143c tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1057.076146] env[61957]: DEBUG nova.compute.manager [-] [instance: 9b0053fe-420e-4b92-86a8-ed44f6a6ec49] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1057.076315] env[61957]: DEBUG nova.network.neutron [-] [instance: 9b0053fe-420e-4b92-86a8-ed44f6a6ec49] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1057.088989] env[61957]: DEBUG oslo_vmware.api [None req-8a0a01e3-6392-4a5c-b585-5e26a0a02629 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': task-1278061, 'name': PowerOnVM_Task, 'duration_secs': 0.528575} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1057.089372] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-8a0a01e3-6392-4a5c-b585-5e26a0a02629 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 4625d1dc-a621-4a33-9537-9e8cce96c914] Powered on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1057.089671] env[61957]: INFO nova.compute.manager [None req-8a0a01e3-6392-4a5c-b585-5e26a0a02629 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 4625d1dc-a621-4a33-9537-9e8cce96c914] Took 7.20 seconds to spawn the instance on the hypervisor. [ 1057.089964] env[61957]: DEBUG nova.compute.manager [None req-8a0a01e3-6392-4a5c-b585-5e26a0a02629 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 4625d1dc-a621-4a33-9537-9e8cce96c914] Checking state {{(pid=61957) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1057.091094] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0ae66b9-fb55-47c4-8e9e-198f90855e11 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.217661] env[61957]: DEBUG nova.network.neutron [-] [instance: 42860944-c700-43be-b291-970803fc7955] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1057.241801] env[61957]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d4017d59-61ec-4dc4-9b1a-6c20b1c1585e {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.252207] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73f68b2f-c43e-4582-adbb-ce48caee525b {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.278278] env[61957]: DEBUG nova.compute.manager [req-e2001167-3a09-44e0-92fc-e26cf1ddcc6d req-42d53ce0-b9c5-4208-8ddc-c968f4e99f56 service nova] [instance: 42860944-c700-43be-b291-970803fc7955] Detach interface failed, port_id=0b2a4b91-7550-44b3-b4de-0fccf960ea77, reason: Instance 42860944-c700-43be-b291-970803fc7955 could not be found. {{(pid=61957) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1057.537022] env[61957]: DEBUG nova.compute.utils [None req-bfdcb579-7ae1-4a05-9d08-d1a10d0fec8a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Using /dev/sd instead of None {{(pid=61957) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1057.537022] env[61957]: DEBUG nova.compute.manager [None req-bfdcb579-7ae1-4a05-9d08-d1a10d0fec8a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 11d67afa-1d96-4a9b-8439-383d586d8e17] Allocating IP information in the background. {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1057.537022] env[61957]: DEBUG nova.network.neutron [None req-bfdcb579-7ae1-4a05-9d08-d1a10d0fec8a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 11d67afa-1d96-4a9b-8439-383d586d8e17] allocate_for_instance() {{(pid=61957) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1057.576464] env[61957]: DEBUG nova.policy [None req-bfdcb579-7ae1-4a05-9d08-d1a10d0fec8a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e2f13689624b483d903ce12ef290db8b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '65fd71bcd1bf41238a9cc3a5d6dd4924', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61957) authorize /opt/stack/nova/nova/policy.py:203}} [ 1057.589560] env[61957]: INFO nova.compute.manager [None req-4a6d948f-b9a8-4fc6-bd2c-f6dd4bec3479 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: a7419b30-1539-4977-a5d9-93a49d84914a] Took 21.34 seconds to build instance. [ 1057.613703] env[61957]: INFO nova.compute.manager [None req-8a0a01e3-6392-4a5c-b585-5e26a0a02629 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 4625d1dc-a621-4a33-9537-9e8cce96c914] Took 17.81 seconds to build instance. [ 1057.720161] env[61957]: INFO nova.compute.manager [-] [instance: 42860944-c700-43be-b291-970803fc7955] Took 1.30 seconds to deallocate network for instance. [ 1057.799858] env[61957]: DEBUG nova.network.neutron [-] [instance: 9b0053fe-420e-4b92-86a8-ed44f6a6ec49] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1057.841367] env[61957]: DEBUG nova.network.neutron [None req-bfdcb579-7ae1-4a05-9d08-d1a10d0fec8a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 11d67afa-1d96-4a9b-8439-383d586d8e17] Successfully created port: 10ee2afa-334a-45ec-9b58-02e6ee3af703 {{(pid=61957) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1058.039162] env[61957]: DEBUG nova.compute.manager [None req-bfdcb579-7ae1-4a05-9d08-d1a10d0fec8a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 11d67afa-1d96-4a9b-8439-383d586d8e17] Start building block device mappings for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1058.093020] env[61957]: DEBUG oslo_concurrency.lockutils [None req-4a6d948f-b9a8-4fc6-bd2c-f6dd4bec3479 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Lock "a7419b30-1539-4977-a5d9-93a49d84914a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.857s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1058.115327] env[61957]: DEBUG oslo_concurrency.lockutils [None req-8a0a01e3-6392-4a5c-b585-5e26a0a02629 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Lock "4625d1dc-a621-4a33-9537-9e8cce96c914" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.324s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1058.226681] env[61957]: DEBUG oslo_concurrency.lockutils [None req-71d3ee2e-25df-44b6-92cf-ec6938a2e3f1 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1058.226967] env[61957]: DEBUG oslo_concurrency.lockutils [None req-71d3ee2e-25df-44b6-92cf-ec6938a2e3f1 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1058.227211] env[61957]: DEBUG nova.objects.instance [None req-71d3ee2e-25df-44b6-92cf-ec6938a2e3f1 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Lazy-loading 'resources' on Instance uuid 42860944-c700-43be-b291-970803fc7955 {{(pid=61957) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1058.305630] env[61957]: INFO nova.compute.manager [-] [instance: 9b0053fe-420e-4b92-86a8-ed44f6a6ec49] Took 1.23 seconds to deallocate network for instance. [ 1058.571198] env[61957]: DEBUG nova.compute.manager [req-f9690b99-b0db-433c-b513-e19633b21be8 req-41571b78-3b66-4c14-b003-224a922dec0f service nova] [instance: 4625d1dc-a621-4a33-9537-9e8cce96c914] Received event network-changed-8365f86e-8b14-47de-a6f4-297107be492e {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1058.571438] env[61957]: DEBUG nova.compute.manager [req-f9690b99-b0db-433c-b513-e19633b21be8 req-41571b78-3b66-4c14-b003-224a922dec0f service nova] [instance: 4625d1dc-a621-4a33-9537-9e8cce96c914] Refreshing instance network info cache due to event network-changed-8365f86e-8b14-47de-a6f4-297107be492e. {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1058.571666] env[61957]: DEBUG oslo_concurrency.lockutils [req-f9690b99-b0db-433c-b513-e19633b21be8 req-41571b78-3b66-4c14-b003-224a922dec0f service nova] Acquiring lock "refresh_cache-4625d1dc-a621-4a33-9537-9e8cce96c914" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1058.571815] env[61957]: DEBUG oslo_concurrency.lockutils [req-f9690b99-b0db-433c-b513-e19633b21be8 req-41571b78-3b66-4c14-b003-224a922dec0f service nova] Acquired lock "refresh_cache-4625d1dc-a621-4a33-9537-9e8cce96c914" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1058.572015] env[61957]: DEBUG nova.network.neutron [req-f9690b99-b0db-433c-b513-e19633b21be8 req-41571b78-3b66-4c14-b003-224a922dec0f service nova] [instance: 4625d1dc-a621-4a33-9537-9e8cce96c914] Refreshing network info cache for port 8365f86e-8b14-47de-a6f4-297107be492e {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1058.813637] env[61957]: DEBUG oslo_concurrency.lockutils [None req-f56f6d3c-d57e-4ad4-8496-b310c6e6143c tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1058.825570] env[61957]: DEBUG nova.compute.manager [req-060c24f5-2446-48de-ac51-5a3ea219db50 req-9f51e287-2074-467e-b87c-6a6dafc20425 service nova] [instance: 9b0053fe-420e-4b92-86a8-ed44f6a6ec49] Received event network-vif-deleted-219230f0-0b14-4be9-81bb-67a765d8ced9 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1058.837477] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92b9c42b-8d99-4124-9dd4-ca2f6211d06f {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.845482] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82676e81-dcff-4c26-b989-bd739f583684 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.875952] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2811d773-eb21-441d-b013-9992a69ad8c2 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.883484] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-681afe5b-e040-4678-8585-24ee97debd0e {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.897104] env[61957]: DEBUG nova.compute.provider_tree [None req-71d3ee2e-25df-44b6-92cf-ec6938a2e3f1 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Updating inventory in ProviderTree for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1059.023777] env[61957]: DEBUG oslo_concurrency.lockutils [None req-79a025c2-964c-42b1-abd7-3dafe654ac62 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Acquiring lock "a7419b30-1539-4977-a5d9-93a49d84914a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1059.024283] env[61957]: DEBUG oslo_concurrency.lockutils [None req-79a025c2-964c-42b1-abd7-3dafe654ac62 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Lock "a7419b30-1539-4977-a5d9-93a49d84914a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1059.024369] env[61957]: DEBUG oslo_concurrency.lockutils [None req-79a025c2-964c-42b1-abd7-3dafe654ac62 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Acquiring lock "a7419b30-1539-4977-a5d9-93a49d84914a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1059.024548] env[61957]: DEBUG oslo_concurrency.lockutils [None req-79a025c2-964c-42b1-abd7-3dafe654ac62 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Lock "a7419b30-1539-4977-a5d9-93a49d84914a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1059.024743] env[61957]: DEBUG oslo_concurrency.lockutils [None req-79a025c2-964c-42b1-abd7-3dafe654ac62 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Lock "a7419b30-1539-4977-a5d9-93a49d84914a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1059.026853] env[61957]: INFO nova.compute.manager [None req-79a025c2-964c-42b1-abd7-3dafe654ac62 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: a7419b30-1539-4977-a5d9-93a49d84914a] Terminating instance [ 1059.028984] env[61957]: DEBUG nova.compute.manager [None req-79a025c2-964c-42b1-abd7-3dafe654ac62 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: a7419b30-1539-4977-a5d9-93a49d84914a] Start destroying the instance on the hypervisor. {{(pid=61957) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1059.029138] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-79a025c2-964c-42b1-abd7-3dafe654ac62 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: a7419b30-1539-4977-a5d9-93a49d84914a] Destroying instance {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1059.029972] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78dc0cb3-c840-4c5b-9124-1f42926b6c6c {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.037508] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-79a025c2-964c-42b1-abd7-3dafe654ac62 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: a7419b30-1539-4977-a5d9-93a49d84914a] Powering off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1059.037752] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-42659d73-97f9-414a-ad7d-1d6c96891a2f {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.044609] env[61957]: DEBUG oslo_vmware.api [None req-79a025c2-964c-42b1-abd7-3dafe654ac62 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Waiting for the task: (returnval){ [ 1059.044609] env[61957]: value = "task-1278062" [ 1059.044609] env[61957]: _type = "Task" [ 1059.044609] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.048836] env[61957]: DEBUG nova.compute.manager [None req-bfdcb579-7ae1-4a05-9d08-d1a10d0fec8a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 11d67afa-1d96-4a9b-8439-383d586d8e17] Start spawning the instance on the hypervisor. {{(pid=61957) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1059.055604] env[61957]: DEBUG oslo_vmware.api [None req-79a025c2-964c-42b1-abd7-3dafe654ac62 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': task-1278062, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.070459] env[61957]: DEBUG nova.virt.hardware [None req-bfdcb579-7ae1-4a05-9d08-d1a10d0fec8a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T17:22:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T17:21:50Z,direct_url=,disk_format='vmdk',id=11c76a2c-f705-470a-ba9d-4657858bab38,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='81d3096d0f094373913fc3dc8f5c3c6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T17:21:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1059.071031] env[61957]: DEBUG nova.virt.hardware [None req-bfdcb579-7ae1-4a05-9d08-d1a10d0fec8a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Flavor limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1059.071031] env[61957]: DEBUG nova.virt.hardware [None req-bfdcb579-7ae1-4a05-9d08-d1a10d0fec8a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Image limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1059.071117] env[61957]: DEBUG nova.virt.hardware [None req-bfdcb579-7ae1-4a05-9d08-d1a10d0fec8a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Flavor pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1059.071327] env[61957]: DEBUG nova.virt.hardware [None req-bfdcb579-7ae1-4a05-9d08-d1a10d0fec8a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Image pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1059.071510] env[61957]: DEBUG nova.virt.hardware [None req-bfdcb579-7ae1-4a05-9d08-d1a10d0fec8a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1059.072301] env[61957]: DEBUG nova.virt.hardware [None req-bfdcb579-7ae1-4a05-9d08-d1a10d0fec8a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1059.072301] env[61957]: DEBUG nova.virt.hardware [None req-bfdcb579-7ae1-4a05-9d08-d1a10d0fec8a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1059.072301] env[61957]: DEBUG nova.virt.hardware [None req-bfdcb579-7ae1-4a05-9d08-d1a10d0fec8a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Got 1 possible topologies {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1059.072301] env[61957]: DEBUG nova.virt.hardware [None req-bfdcb579-7ae1-4a05-9d08-d1a10d0fec8a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1059.072516] env[61957]: DEBUG nova.virt.hardware [None req-bfdcb579-7ae1-4a05-9d08-d1a10d0fec8a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1059.073517] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcb406d4-c325-437b-8578-f155798dede0 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.083522] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8a687db-20a9-42f0-9e7b-aaf82e36bb7a {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.343331] env[61957]: DEBUG nova.compute.manager [req-3f34bde4-d680-4844-966d-119ae382ab10 req-01820db2-ef47-4486-9f23-c9ad8854e12a service nova] [instance: 11d67afa-1d96-4a9b-8439-383d586d8e17] Received event network-vif-plugged-10ee2afa-334a-45ec-9b58-02e6ee3af703 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1059.343636] env[61957]: DEBUG oslo_concurrency.lockutils [req-3f34bde4-d680-4844-966d-119ae382ab10 req-01820db2-ef47-4486-9f23-c9ad8854e12a service nova] Acquiring lock "11d67afa-1d96-4a9b-8439-383d586d8e17-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1059.343806] env[61957]: DEBUG oslo_concurrency.lockutils [req-3f34bde4-d680-4844-966d-119ae382ab10 req-01820db2-ef47-4486-9f23-c9ad8854e12a service nova] Lock "11d67afa-1d96-4a9b-8439-383d586d8e17-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1059.343932] env[61957]: DEBUG oslo_concurrency.lockutils [req-3f34bde4-d680-4844-966d-119ae382ab10 req-01820db2-ef47-4486-9f23-c9ad8854e12a service nova] Lock "11d67afa-1d96-4a9b-8439-383d586d8e17-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1059.344693] env[61957]: DEBUG nova.compute.manager [req-3f34bde4-d680-4844-966d-119ae382ab10 req-01820db2-ef47-4486-9f23-c9ad8854e12a service nova] [instance: 11d67afa-1d96-4a9b-8439-383d586d8e17] No waiting events found dispatching network-vif-plugged-10ee2afa-334a-45ec-9b58-02e6ee3af703 {{(pid=61957) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1059.344693] env[61957]: WARNING nova.compute.manager [req-3f34bde4-d680-4844-966d-119ae382ab10 req-01820db2-ef47-4486-9f23-c9ad8854e12a service nova] [instance: 11d67afa-1d96-4a9b-8439-383d586d8e17] Received unexpected event network-vif-plugged-10ee2afa-334a-45ec-9b58-02e6ee3af703 for instance with vm_state building and task_state spawning. [ 1059.400129] env[61957]: DEBUG nova.network.neutron [req-f9690b99-b0db-433c-b513-e19633b21be8 req-41571b78-3b66-4c14-b003-224a922dec0f service nova] [instance: 4625d1dc-a621-4a33-9537-9e8cce96c914] Updated VIF entry in instance network info cache for port 8365f86e-8b14-47de-a6f4-297107be492e. {{(pid=61957) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1059.400129] env[61957]: DEBUG nova.network.neutron [req-f9690b99-b0db-433c-b513-e19633b21be8 req-41571b78-3b66-4c14-b003-224a922dec0f service nova] [instance: 4625d1dc-a621-4a33-9537-9e8cce96c914] Updating instance_info_cache with network_info: [{"id": "8365f86e-8b14-47de-a6f4-297107be492e", "address": "fa:16:3e:e8:59:9d", "network": {"id": "364cabcb-046d-4f91-b1ee-5dd3adfae6d7", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1432396967-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.232", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8713aa35bcb24b86ad0b58ca9fc991ba", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b7a73c01-1bb9-4612-a1a7-16d71b732e81", "external-id": "nsx-vlan-transportzone-711", "segmentation_id": 711, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8365f86e-8b", "ovs_interfaceid": "8365f86e-8b14-47de-a6f4-297107be492e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1059.421932] env[61957]: ERROR nova.scheduler.client.report [None req-71d3ee2e-25df-44b6-92cf-ec6938a2e3f1 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [req-76a3d39d-2cdd-450f-a919-daffee1383b7] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 0ceb6c9e-61c6-496d-8579-9d32627e96da. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-76a3d39d-2cdd-450f-a919-daffee1383b7"}]} [ 1059.440335] env[61957]: DEBUG nova.scheduler.client.report [None req-71d3ee2e-25df-44b6-92cf-ec6938a2e3f1 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Refreshing inventories for resource provider 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 1059.453141] env[61957]: DEBUG nova.scheduler.client.report [None req-71d3ee2e-25df-44b6-92cf-ec6938a2e3f1 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Updating ProviderTree inventory for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 1059.453372] env[61957]: DEBUG nova.compute.provider_tree [None req-71d3ee2e-25df-44b6-92cf-ec6938a2e3f1 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Updating inventory in ProviderTree for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1059.464433] env[61957]: DEBUG nova.scheduler.client.report [None req-71d3ee2e-25df-44b6-92cf-ec6938a2e3f1 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Refreshing aggregate associations for resource provider 0ceb6c9e-61c6-496d-8579-9d32627e96da, aggregates: None {{(pid=61957) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 1059.481988] env[61957]: DEBUG nova.scheduler.client.report [None req-71d3ee2e-25df-44b6-92cf-ec6938a2e3f1 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Refreshing trait associations for resource provider 0ceb6c9e-61c6-496d-8579-9d32627e96da, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO,HW_ARCH_X86_64,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=61957) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 1059.514194] env[61957]: DEBUG oslo_concurrency.lockutils [None req-0629491b-6b09-4d57-8893-e804be404b21 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Acquiring lock "b82af1e9-294a-4ba9-bcad-73b2a2aca86d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1059.514428] env[61957]: DEBUG oslo_concurrency.lockutils [None req-0629491b-6b09-4d57-8893-e804be404b21 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Lock "b82af1e9-294a-4ba9-bcad-73b2a2aca86d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1059.555177] env[61957]: DEBUG oslo_vmware.api [None req-79a025c2-964c-42b1-abd7-3dafe654ac62 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': task-1278062, 'name': PowerOffVM_Task, 'duration_secs': 0.202941} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1059.557448] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-79a025c2-964c-42b1-abd7-3dafe654ac62 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: a7419b30-1539-4977-a5d9-93a49d84914a] Powered off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1059.557681] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-79a025c2-964c-42b1-abd7-3dafe654ac62 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: a7419b30-1539-4977-a5d9-93a49d84914a] Unregistering the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1059.558127] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-24653e93-a12a-40c8-b5f8-4de7cbef75fb {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.587534] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-308608fc-d757-4479-902c-94772284d126 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.595713] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef4b1597-2473-4c55-bec9-0ad7e78b61f6 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.626254] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5db5d2ea-f01d-4c15-a586-1ff9263f53a8 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.629108] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-79a025c2-964c-42b1-abd7-3dafe654ac62 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: a7419b30-1539-4977-a5d9-93a49d84914a] Unregistered the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1059.629337] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-79a025c2-964c-42b1-abd7-3dafe654ac62 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: a7419b30-1539-4977-a5d9-93a49d84914a] Deleting contents of the VM from datastore datastore1 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1059.629521] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-79a025c2-964c-42b1-abd7-3dafe654ac62 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Deleting the datastore file [datastore1] a7419b30-1539-4977-a5d9-93a49d84914a {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1059.629781] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-dc5fb89f-0d4c-4224-a352-75d0d0a7b326 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.639238] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-975ddac5-aa61-4d22-9380-edf1d194bcd8 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.643292] env[61957]: DEBUG oslo_vmware.api [None req-79a025c2-964c-42b1-abd7-3dafe654ac62 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Waiting for the task: (returnval){ [ 1059.643292] env[61957]: value = "task-1278064" [ 1059.643292] env[61957]: _type = "Task" [ 1059.643292] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.654899] env[61957]: DEBUG nova.compute.provider_tree [None req-71d3ee2e-25df-44b6-92cf-ec6938a2e3f1 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Updating inventory in ProviderTree for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1059.661131] env[61957]: DEBUG oslo_vmware.api [None req-79a025c2-964c-42b1-abd7-3dafe654ac62 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': task-1278064, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.904514] env[61957]: DEBUG oslo_concurrency.lockutils [req-f9690b99-b0db-433c-b513-e19633b21be8 req-41571b78-3b66-4c14-b003-224a922dec0f service nova] Releasing lock "refresh_cache-4625d1dc-a621-4a33-9537-9e8cce96c914" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1059.914384] env[61957]: DEBUG nova.network.neutron [None req-bfdcb579-7ae1-4a05-9d08-d1a10d0fec8a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 11d67afa-1d96-4a9b-8439-383d586d8e17] Successfully updated port: 10ee2afa-334a-45ec-9b58-02e6ee3af703 {{(pid=61957) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1060.017395] env[61957]: DEBUG nova.compute.manager [None req-0629491b-6b09-4d57-8893-e804be404b21 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: b82af1e9-294a-4ba9-bcad-73b2a2aca86d] Starting instance... {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1060.153145] env[61957]: DEBUG oslo_vmware.api [None req-79a025c2-964c-42b1-abd7-3dafe654ac62 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Task: {'id': task-1278064, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.161441} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1060.153548] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-79a025c2-964c-42b1-abd7-3dafe654ac62 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Deleted the datastore file {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1060.153629] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-79a025c2-964c-42b1-abd7-3dafe654ac62 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: a7419b30-1539-4977-a5d9-93a49d84914a] Deleted contents of the VM from datastore datastore1 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1060.153815] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-79a025c2-964c-42b1-abd7-3dafe654ac62 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: a7419b30-1539-4977-a5d9-93a49d84914a] Instance destroyed {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1060.153993] env[61957]: INFO nova.compute.manager [None req-79a025c2-964c-42b1-abd7-3dafe654ac62 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] [instance: a7419b30-1539-4977-a5d9-93a49d84914a] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1060.154267] env[61957]: DEBUG oslo.service.loopingcall [None req-79a025c2-964c-42b1-abd7-3dafe654ac62 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1060.154467] env[61957]: DEBUG nova.compute.manager [-] [instance: a7419b30-1539-4977-a5d9-93a49d84914a] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1060.154569] env[61957]: DEBUG nova.network.neutron [-] [instance: a7419b30-1539-4977-a5d9-93a49d84914a] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1060.182871] env[61957]: DEBUG nova.scheduler.client.report [None req-71d3ee2e-25df-44b6-92cf-ec6938a2e3f1 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Updated inventory for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da with generation 124 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 1060.183164] env[61957]: DEBUG nova.compute.provider_tree [None req-71d3ee2e-25df-44b6-92cf-ec6938a2e3f1 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Updating resource provider 0ceb6c9e-61c6-496d-8579-9d32627e96da generation from 124 to 125 during operation: update_inventory {{(pid=61957) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1060.183350] env[61957]: DEBUG nova.compute.provider_tree [None req-71d3ee2e-25df-44b6-92cf-ec6938a2e3f1 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Updating inventory in ProviderTree for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1060.416631] env[61957]: DEBUG oslo_concurrency.lockutils [None req-bfdcb579-7ae1-4a05-9d08-d1a10d0fec8a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Acquiring lock "refresh_cache-11d67afa-1d96-4a9b-8439-383d586d8e17" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1060.416803] env[61957]: DEBUG oslo_concurrency.lockutils [None req-bfdcb579-7ae1-4a05-9d08-d1a10d0fec8a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Acquired lock "refresh_cache-11d67afa-1d96-4a9b-8439-383d586d8e17" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1060.416961] env[61957]: DEBUG nova.network.neutron [None req-bfdcb579-7ae1-4a05-9d08-d1a10d0fec8a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 11d67afa-1d96-4a9b-8439-383d586d8e17] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1060.539341] env[61957]: DEBUG oslo_concurrency.lockutils [None req-0629491b-6b09-4d57-8893-e804be404b21 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1060.687693] env[61957]: DEBUG oslo_concurrency.lockutils [None req-71d3ee2e-25df-44b6-92cf-ec6938a2e3f1 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.461s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1060.691010] env[61957]: DEBUG oslo_concurrency.lockutils [None req-f56f6d3c-d57e-4ad4-8496-b310c6e6143c tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.877s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1060.691279] env[61957]: DEBUG nova.objects.instance [None req-f56f6d3c-d57e-4ad4-8496-b310c6e6143c tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Lazy-loading 'resources' on Instance uuid 9b0053fe-420e-4b92-86a8-ed44f6a6ec49 {{(pid=61957) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1060.707852] env[61957]: INFO nova.scheduler.client.report [None req-71d3ee2e-25df-44b6-92cf-ec6938a2e3f1 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Deleted allocations for instance 42860944-c700-43be-b291-970803fc7955 [ 1060.850914] env[61957]: DEBUG nova.compute.manager [req-78d4589e-92a8-4479-a786-1260381c2a40 req-594dd172-6b8d-4a3a-b9e9-457bb8123c1e service nova] [instance: 11d67afa-1d96-4a9b-8439-383d586d8e17] Received event network-changed-10ee2afa-334a-45ec-9b58-02e6ee3af703 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1060.851041] env[61957]: DEBUG nova.compute.manager [req-78d4589e-92a8-4479-a786-1260381c2a40 req-594dd172-6b8d-4a3a-b9e9-457bb8123c1e service nova] [instance: 11d67afa-1d96-4a9b-8439-383d586d8e17] Refreshing instance network info cache due to event network-changed-10ee2afa-334a-45ec-9b58-02e6ee3af703. {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1060.851248] env[61957]: DEBUG oslo_concurrency.lockutils [req-78d4589e-92a8-4479-a786-1260381c2a40 req-594dd172-6b8d-4a3a-b9e9-457bb8123c1e service nova] Acquiring lock "refresh_cache-11d67afa-1d96-4a9b-8439-383d586d8e17" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1060.888337] env[61957]: DEBUG nova.network.neutron [-] [instance: a7419b30-1539-4977-a5d9-93a49d84914a] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1060.948676] env[61957]: DEBUG nova.network.neutron [None req-bfdcb579-7ae1-4a05-9d08-d1a10d0fec8a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 11d67afa-1d96-4a9b-8439-383d586d8e17] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1061.093374] env[61957]: DEBUG nova.network.neutron [None req-bfdcb579-7ae1-4a05-9d08-d1a10d0fec8a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 11d67afa-1d96-4a9b-8439-383d586d8e17] Updating instance_info_cache with network_info: [{"id": "10ee2afa-334a-45ec-9b58-02e6ee3af703", "address": "fa:16:3e:30:a0:58", "network": {"id": "4f0e57be-8906-48b5-865f-59ff12250a8b", "bridge": "br-int", "label": "tempest-ServersTestJSON-1916182687-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "65fd71bcd1bf41238a9cc3a5d6dd4924", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "939c05b6-8f31-4f3a-95ac-6297e0bd243e", "external-id": "nsx-vlan-transportzone-825", "segmentation_id": 825, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap10ee2afa-33", "ovs_interfaceid": "10ee2afa-334a-45ec-9b58-02e6ee3af703", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1061.219104] env[61957]: DEBUG oslo_concurrency.lockutils [None req-71d3ee2e-25df-44b6-92cf-ec6938a2e3f1 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Lock "42860944-c700-43be-b291-970803fc7955" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.426s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1061.305364] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8349049-de02-4dc6-8f9e-d2c3ac344784 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.314643] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69716d95-e2b3-48c8-937d-640fe23eb929 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.343370] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4717840e-a1b6-442c-bfab-a19d743bb4b2 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.350313] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10b76ba7-07ab-429c-89d9-3094e418949f {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.362855] env[61957]: DEBUG nova.compute.provider_tree [None req-f56f6d3c-d57e-4ad4-8496-b310c6e6143c tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1061.391504] env[61957]: INFO nova.compute.manager [-] [instance: a7419b30-1539-4977-a5d9-93a49d84914a] Took 1.24 seconds to deallocate network for instance. [ 1061.596887] env[61957]: DEBUG oslo_concurrency.lockutils [None req-bfdcb579-7ae1-4a05-9d08-d1a10d0fec8a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Releasing lock "refresh_cache-11d67afa-1d96-4a9b-8439-383d586d8e17" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1061.597256] env[61957]: DEBUG nova.compute.manager [None req-bfdcb579-7ae1-4a05-9d08-d1a10d0fec8a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 11d67afa-1d96-4a9b-8439-383d586d8e17] Instance network_info: |[{"id": "10ee2afa-334a-45ec-9b58-02e6ee3af703", "address": "fa:16:3e:30:a0:58", "network": {"id": "4f0e57be-8906-48b5-865f-59ff12250a8b", "bridge": "br-int", "label": "tempest-ServersTestJSON-1916182687-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "65fd71bcd1bf41238a9cc3a5d6dd4924", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "939c05b6-8f31-4f3a-95ac-6297e0bd243e", "external-id": "nsx-vlan-transportzone-825", "segmentation_id": 825, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap10ee2afa-33", "ovs_interfaceid": "10ee2afa-334a-45ec-9b58-02e6ee3af703", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1061.597581] env[61957]: DEBUG oslo_concurrency.lockutils [req-78d4589e-92a8-4479-a786-1260381c2a40 req-594dd172-6b8d-4a3a-b9e9-457bb8123c1e service nova] Acquired lock "refresh_cache-11d67afa-1d96-4a9b-8439-383d586d8e17" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1061.597768] env[61957]: DEBUG nova.network.neutron [req-78d4589e-92a8-4479-a786-1260381c2a40 req-594dd172-6b8d-4a3a-b9e9-457bb8123c1e service nova] [instance: 11d67afa-1d96-4a9b-8439-383d586d8e17] Refreshing network info cache for port 10ee2afa-334a-45ec-9b58-02e6ee3af703 {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1061.598931] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-bfdcb579-7ae1-4a05-9d08-d1a10d0fec8a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 11d67afa-1d96-4a9b-8439-383d586d8e17] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:30:a0:58', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '939c05b6-8f31-4f3a-95ac-6297e0bd243e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '10ee2afa-334a-45ec-9b58-02e6ee3af703', 'vif_model': 'vmxnet3'}] {{(pid=61957) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1061.606993] env[61957]: DEBUG oslo.service.loopingcall [None req-bfdcb579-7ae1-4a05-9d08-d1a10d0fec8a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1061.610404] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 11d67afa-1d96-4a9b-8439-383d586d8e17] Creating VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1061.610945] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-719abe8d-4355-419d-a20c-3618b0aa3d1e {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.631762] env[61957]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1061.631762] env[61957]: value = "task-1278065" [ 1061.631762] env[61957]: _type = "Task" [ 1061.631762] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1061.639199] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1278065, 'name': CreateVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.844987] env[61957]: DEBUG nova.network.neutron [req-78d4589e-92a8-4479-a786-1260381c2a40 req-594dd172-6b8d-4a3a-b9e9-457bb8123c1e service nova] [instance: 11d67afa-1d96-4a9b-8439-383d586d8e17] Updated VIF entry in instance network info cache for port 10ee2afa-334a-45ec-9b58-02e6ee3af703. {{(pid=61957) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1061.845388] env[61957]: DEBUG nova.network.neutron [req-78d4589e-92a8-4479-a786-1260381c2a40 req-594dd172-6b8d-4a3a-b9e9-457bb8123c1e service nova] [instance: 11d67afa-1d96-4a9b-8439-383d586d8e17] Updating instance_info_cache with network_info: [{"id": "10ee2afa-334a-45ec-9b58-02e6ee3af703", "address": "fa:16:3e:30:a0:58", "network": {"id": "4f0e57be-8906-48b5-865f-59ff12250a8b", "bridge": "br-int", "label": "tempest-ServersTestJSON-1916182687-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "65fd71bcd1bf41238a9cc3a5d6dd4924", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "939c05b6-8f31-4f3a-95ac-6297e0bd243e", "external-id": "nsx-vlan-transportzone-825", "segmentation_id": 825, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap10ee2afa-33", "ovs_interfaceid": "10ee2afa-334a-45ec-9b58-02e6ee3af703", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1061.866053] env[61957]: DEBUG nova.scheduler.client.report [None req-f56f6d3c-d57e-4ad4-8496-b310c6e6143c tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1061.897878] env[61957]: DEBUG oslo_concurrency.lockutils [None req-79a025c2-964c-42b1-abd7-3dafe654ac62 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1062.027421] env[61957]: DEBUG oslo_concurrency.lockutils [None req-92fb2eb5-36b1-4a05-8eec-33f1fe3376d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Acquiring lock "84cd5619-4ac9-41ef-9368-a023ad9ae66d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1062.027666] env[61957]: DEBUG oslo_concurrency.lockutils [None req-92fb2eb5-36b1-4a05-8eec-33f1fe3376d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Lock "84cd5619-4ac9-41ef-9368-a023ad9ae66d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1062.142324] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1278065, 'name': CreateVM_Task, 'duration_secs': 0.296369} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1062.142324] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 11d67afa-1d96-4a9b-8439-383d586d8e17] Created VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1062.142499] env[61957]: DEBUG oslo_concurrency.lockutils [None req-bfdcb579-7ae1-4a05-9d08-d1a10d0fec8a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1062.142753] env[61957]: DEBUG oslo_concurrency.lockutils [None req-bfdcb579-7ae1-4a05-9d08-d1a10d0fec8a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1062.143149] env[61957]: DEBUG oslo_concurrency.lockutils [None req-bfdcb579-7ae1-4a05-9d08-d1a10d0fec8a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1062.143412] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-899fe0fb-9add-4a59-a363-5a6b889e9dbe {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.147830] env[61957]: DEBUG oslo_vmware.api [None req-bfdcb579-7ae1-4a05-9d08-d1a10d0fec8a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Waiting for the task: (returnval){ [ 1062.147830] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]528cdbef-814b-3827-c813-580dcc89f63b" [ 1062.147830] env[61957]: _type = "Task" [ 1062.147830] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1062.155333] env[61957]: DEBUG oslo_vmware.api [None req-bfdcb579-7ae1-4a05-9d08-d1a10d0fec8a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]528cdbef-814b-3827-c813-580dcc89f63b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.348753] env[61957]: DEBUG oslo_concurrency.lockutils [req-78d4589e-92a8-4479-a786-1260381c2a40 req-594dd172-6b8d-4a3a-b9e9-457bb8123c1e service nova] Releasing lock "refresh_cache-11d67afa-1d96-4a9b-8439-383d586d8e17" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1062.349148] env[61957]: DEBUG nova.compute.manager [req-78d4589e-92a8-4479-a786-1260381c2a40 req-594dd172-6b8d-4a3a-b9e9-457bb8123c1e service nova] [instance: a7419b30-1539-4977-a5d9-93a49d84914a] Received event network-vif-deleted-6bfe39fb-7cc2-4933-bdf0-f8b2a121aab2 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1062.349240] env[61957]: INFO nova.compute.manager [req-78d4589e-92a8-4479-a786-1260381c2a40 req-594dd172-6b8d-4a3a-b9e9-457bb8123c1e service nova] [instance: a7419b30-1539-4977-a5d9-93a49d84914a] Neutron deleted interface 6bfe39fb-7cc2-4933-bdf0-f8b2a121aab2; detaching it from the instance and deleting it from the info cache [ 1062.349446] env[61957]: DEBUG nova.network.neutron [req-78d4589e-92a8-4479-a786-1260381c2a40 req-594dd172-6b8d-4a3a-b9e9-457bb8123c1e service nova] [instance: a7419b30-1539-4977-a5d9-93a49d84914a] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1062.370607] env[61957]: DEBUG oslo_concurrency.lockutils [None req-f56f6d3c-d57e-4ad4-8496-b310c6e6143c tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.680s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1062.373044] env[61957]: DEBUG oslo_concurrency.lockutils [None req-0629491b-6b09-4d57-8893-e804be404b21 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.834s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1062.374586] env[61957]: INFO nova.compute.claims [None req-0629491b-6b09-4d57-8893-e804be404b21 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: b82af1e9-294a-4ba9-bcad-73b2a2aca86d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1062.390475] env[61957]: INFO nova.scheduler.client.report [None req-f56f6d3c-d57e-4ad4-8496-b310c6e6143c tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Deleted allocations for instance 9b0053fe-420e-4b92-86a8-ed44f6a6ec49 [ 1062.529915] env[61957]: DEBUG nova.compute.manager [None req-92fb2eb5-36b1-4a05-8eec-33f1fe3376d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 84cd5619-4ac9-41ef-9368-a023ad9ae66d] Starting instance... {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1062.658116] env[61957]: DEBUG oslo_vmware.api [None req-bfdcb579-7ae1-4a05-9d08-d1a10d0fec8a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]528cdbef-814b-3827-c813-580dcc89f63b, 'name': SearchDatastore_Task, 'duration_secs': 0.011346} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1062.658375] env[61957]: DEBUG oslo_concurrency.lockutils [None req-bfdcb579-7ae1-4a05-9d08-d1a10d0fec8a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1062.658614] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-bfdcb579-7ae1-4a05-9d08-d1a10d0fec8a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 11d67afa-1d96-4a9b-8439-383d586d8e17] Processing image 11c76a2c-f705-470a-ba9d-4657858bab38 {{(pid=61957) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1062.658852] env[61957]: DEBUG oslo_concurrency.lockutils [None req-bfdcb579-7ae1-4a05-9d08-d1a10d0fec8a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1062.659012] env[61957]: DEBUG oslo_concurrency.lockutils [None req-bfdcb579-7ae1-4a05-9d08-d1a10d0fec8a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1062.659206] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-bfdcb579-7ae1-4a05-9d08-d1a10d0fec8a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1062.659458] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-82612aff-1776-413f-8512-2887113b4f27 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.667745] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-bfdcb579-7ae1-4a05-9d08-d1a10d0fec8a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1062.667916] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-bfdcb579-7ae1-4a05-9d08-d1a10d0fec8a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61957) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1062.668599] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c233fe6b-8d88-4489-811a-ea9c6b8441e6 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.672952] env[61957]: DEBUG oslo_vmware.api [None req-bfdcb579-7ae1-4a05-9d08-d1a10d0fec8a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Waiting for the task: (returnval){ [ 1062.672952] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52d00e46-af08-c027-5482-7472753d688e" [ 1062.672952] env[61957]: _type = "Task" [ 1062.672952] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1062.679731] env[61957]: DEBUG oslo_vmware.api [None req-bfdcb579-7ae1-4a05-9d08-d1a10d0fec8a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52d00e46-af08-c027-5482-7472753d688e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.851859] env[61957]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b7783c1d-303b-4388-bea4-dd118bac77e5 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.862917] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3710b41-592d-408f-aade-277c7bd360ea {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.887704] env[61957]: DEBUG nova.compute.manager [req-78d4589e-92a8-4479-a786-1260381c2a40 req-594dd172-6b8d-4a3a-b9e9-457bb8123c1e service nova] [instance: a7419b30-1539-4977-a5d9-93a49d84914a] Detach interface failed, port_id=6bfe39fb-7cc2-4933-bdf0-f8b2a121aab2, reason: Instance a7419b30-1539-4977-a5d9-93a49d84914a could not be found. {{(pid=61957) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1062.897820] env[61957]: DEBUG oslo_concurrency.lockutils [None req-f56f6d3c-d57e-4ad4-8496-b310c6e6143c tempest-ServersNegativeTestJSON-622959051 tempest-ServersNegativeTestJSON-622959051-project-member] Lock "9b0053fe-420e-4b92-86a8-ed44f6a6ec49" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.978s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1063.048759] env[61957]: DEBUG oslo_concurrency.lockutils [None req-92fb2eb5-36b1-4a05-8eec-33f1fe3376d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1063.184670] env[61957]: DEBUG oslo_vmware.api [None req-bfdcb579-7ae1-4a05-9d08-d1a10d0fec8a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52d00e46-af08-c027-5482-7472753d688e, 'name': SearchDatastore_Task, 'duration_secs': 0.015759} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1063.185570] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3e30a804-ffba-4172-9017-66174ef3f3f2 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.190683] env[61957]: DEBUG oslo_vmware.api [None req-bfdcb579-7ae1-4a05-9d08-d1a10d0fec8a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Waiting for the task: (returnval){ [ 1063.190683] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52dc5823-9966-a226-fcf6-4d1ff07575bb" [ 1063.190683] env[61957]: _type = "Task" [ 1063.190683] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1063.198122] env[61957]: DEBUG oslo_vmware.api [None req-bfdcb579-7ae1-4a05-9d08-d1a10d0fec8a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52dc5823-9966-a226-fcf6-4d1ff07575bb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.477797] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6b66786-9046-4d90-a3b6-84501e0e4f9f {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.486018] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-899e91b7-0ea5-4ae2-919a-4156aff4abaf {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.514835] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4391bf02-3838-4768-8cb3-2cde8ac88055 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.525836] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53dfc0ad-6690-4731-a1a9-7f18df2cf465 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.535681] env[61957]: DEBUG nova.compute.provider_tree [None req-0629491b-6b09-4d57-8893-e804be404b21 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1063.701621] env[61957]: DEBUG oslo_vmware.api [None req-bfdcb579-7ae1-4a05-9d08-d1a10d0fec8a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52dc5823-9966-a226-fcf6-4d1ff07575bb, 'name': SearchDatastore_Task, 'duration_secs': 0.010802} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1063.701850] env[61957]: DEBUG oslo_concurrency.lockutils [None req-bfdcb579-7ae1-4a05-9d08-d1a10d0fec8a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1063.702122] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-bfdcb579-7ae1-4a05-9d08-d1a10d0fec8a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore2] 11d67afa-1d96-4a9b-8439-383d586d8e17/11d67afa-1d96-4a9b-8439-383d586d8e17.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1063.702380] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-448ce83c-377c-4d3f-a128-788f5e73c31a {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.708192] env[61957]: DEBUG oslo_vmware.api [None req-bfdcb579-7ae1-4a05-9d08-d1a10d0fec8a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Waiting for the task: (returnval){ [ 1063.708192] env[61957]: value = "task-1278066" [ 1063.708192] env[61957]: _type = "Task" [ 1063.708192] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1063.715503] env[61957]: DEBUG oslo_vmware.api [None req-bfdcb579-7ae1-4a05-9d08-d1a10d0fec8a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': task-1278066, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.039724] env[61957]: DEBUG nova.scheduler.client.report [None req-0629491b-6b09-4d57-8893-e804be404b21 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1064.218455] env[61957]: DEBUG oslo_vmware.api [None req-bfdcb579-7ae1-4a05-9d08-d1a10d0fec8a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': task-1278066, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.544958] env[61957]: DEBUG oslo_concurrency.lockutils [None req-0629491b-6b09-4d57-8893-e804be404b21 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.172s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1064.545769] env[61957]: DEBUG nova.compute.manager [None req-0629491b-6b09-4d57-8893-e804be404b21 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: b82af1e9-294a-4ba9-bcad-73b2a2aca86d] Start building networks asynchronously for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1064.548218] env[61957]: DEBUG oslo_concurrency.lockutils [None req-79a025c2-964c-42b1-abd7-3dafe654ac62 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.650s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1064.548887] env[61957]: DEBUG nova.objects.instance [None req-79a025c2-964c-42b1-abd7-3dafe654ac62 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Lazy-loading 'resources' on Instance uuid a7419b30-1539-4977-a5d9-93a49d84914a {{(pid=61957) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1064.720216] env[61957]: DEBUG oslo_vmware.api [None req-bfdcb579-7ae1-4a05-9d08-d1a10d0fec8a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': task-1278066, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.581298} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1064.720539] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-bfdcb579-7ae1-4a05-9d08-d1a10d0fec8a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore2] 11d67afa-1d96-4a9b-8439-383d586d8e17/11d67afa-1d96-4a9b-8439-383d586d8e17.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1064.720810] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-bfdcb579-7ae1-4a05-9d08-d1a10d0fec8a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 11d67afa-1d96-4a9b-8439-383d586d8e17] Extending root virtual disk to 1048576 {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1064.721115] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-519687c4-9317-43a0-b0ea-41b9923c861b {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.727948] env[61957]: DEBUG oslo_vmware.api [None req-bfdcb579-7ae1-4a05-9d08-d1a10d0fec8a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Waiting for the task: (returnval){ [ 1064.727948] env[61957]: value = "task-1278067" [ 1064.727948] env[61957]: _type = "Task" [ 1064.727948] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1064.735996] env[61957]: DEBUG oslo_vmware.api [None req-bfdcb579-7ae1-4a05-9d08-d1a10d0fec8a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': task-1278067, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.051566] env[61957]: DEBUG nova.compute.utils [None req-0629491b-6b09-4d57-8893-e804be404b21 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Using /dev/sd instead of None {{(pid=61957) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1065.055986] env[61957]: DEBUG nova.compute.manager [None req-0629491b-6b09-4d57-8893-e804be404b21 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: b82af1e9-294a-4ba9-bcad-73b2a2aca86d] Allocating IP information in the background. {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1065.056185] env[61957]: DEBUG nova.network.neutron [None req-0629491b-6b09-4d57-8893-e804be404b21 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: b82af1e9-294a-4ba9-bcad-73b2a2aca86d] allocate_for_instance() {{(pid=61957) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1065.106386] env[61957]: DEBUG nova.policy [None req-0629491b-6b09-4d57-8893-e804be404b21 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6fd1c19dc3a44212ada44445e0919106', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8713aa35bcb24b86ad0b58ca9fc991ba', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61957) authorize /opt/stack/nova/nova/policy.py:203}} [ 1065.182333] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8d47df8-7b59-4a5b-8497-7c8ed2bfc229 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.191947] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bed9f03d-a0f3-4810-91fd-da3aab7e1540 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.223861] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-434192ea-da0c-4db2-ba32-197d6d1810a1 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.234432] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b63a2ef4-46e3-4883-8c2c-a617bbbf6c1d {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.249252] env[61957]: DEBUG nova.compute.provider_tree [None req-79a025c2-964c-42b1-abd7-3dafe654ac62 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1065.254064] env[61957]: DEBUG oslo_vmware.api [None req-bfdcb579-7ae1-4a05-9d08-d1a10d0fec8a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': task-1278067, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.166074} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1065.254064] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-bfdcb579-7ae1-4a05-9d08-d1a10d0fec8a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 11d67afa-1d96-4a9b-8439-383d586d8e17] Extended root virtual disk {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1065.254428] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da7ea41f-25b2-4dbb-bb4b-328295bf3a4a {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.278754] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-bfdcb579-7ae1-4a05-9d08-d1a10d0fec8a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 11d67afa-1d96-4a9b-8439-383d586d8e17] Reconfiguring VM instance instance-00000066 to attach disk [datastore2] 11d67afa-1d96-4a9b-8439-383d586d8e17/11d67afa-1d96-4a9b-8439-383d586d8e17.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1065.279427] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-75ee749e-8a41-450d-b91d-fa7a26a0b2a4 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.299192] env[61957]: DEBUG oslo_vmware.api [None req-bfdcb579-7ae1-4a05-9d08-d1a10d0fec8a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Waiting for the task: (returnval){ [ 1065.299192] env[61957]: value = "task-1278068" [ 1065.299192] env[61957]: _type = "Task" [ 1065.299192] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1065.307105] env[61957]: DEBUG oslo_vmware.api [None req-bfdcb579-7ae1-4a05-9d08-d1a10d0fec8a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': task-1278068, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.414282] env[61957]: DEBUG nova.network.neutron [None req-0629491b-6b09-4d57-8893-e804be404b21 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: b82af1e9-294a-4ba9-bcad-73b2a2aca86d] Successfully created port: 58a3f865-3fc6-462a-b685-fb32e01f6c5e {{(pid=61957) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1065.556983] env[61957]: DEBUG nova.compute.manager [None req-0629491b-6b09-4d57-8893-e804be404b21 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: b82af1e9-294a-4ba9-bcad-73b2a2aca86d] Start building block device mappings for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1065.755373] env[61957]: DEBUG nova.scheduler.client.report [None req-79a025c2-964c-42b1-abd7-3dafe654ac62 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1065.809760] env[61957]: DEBUG oslo_vmware.api [None req-bfdcb579-7ae1-4a05-9d08-d1a10d0fec8a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': task-1278068, 'name': ReconfigVM_Task, 'duration_secs': 0.315202} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1065.809988] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-bfdcb579-7ae1-4a05-9d08-d1a10d0fec8a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 11d67afa-1d96-4a9b-8439-383d586d8e17] Reconfigured VM instance instance-00000066 to attach disk [datastore2] 11d67afa-1d96-4a9b-8439-383d586d8e17/11d67afa-1d96-4a9b-8439-383d586d8e17.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1065.810612] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9c930f3f-a215-4232-8e69-d0ff2b13e82d {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.817512] env[61957]: DEBUG oslo_vmware.api [None req-bfdcb579-7ae1-4a05-9d08-d1a10d0fec8a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Waiting for the task: (returnval){ [ 1065.817512] env[61957]: value = "task-1278069" [ 1065.817512] env[61957]: _type = "Task" [ 1065.817512] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1065.824896] env[61957]: DEBUG oslo_vmware.api [None req-bfdcb579-7ae1-4a05-9d08-d1a10d0fec8a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': task-1278069, 'name': Rename_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.259929] env[61957]: DEBUG oslo_concurrency.lockutils [None req-79a025c2-964c-42b1-abd7-3dafe654ac62 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.712s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1066.262248] env[61957]: DEBUG oslo_concurrency.lockutils [None req-92fb2eb5-36b1-4a05-8eec-33f1fe3376d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.214s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1066.263782] env[61957]: INFO nova.compute.claims [None req-92fb2eb5-36b1-4a05-8eec-33f1fe3376d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 84cd5619-4ac9-41ef-9368-a023ad9ae66d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1066.284164] env[61957]: INFO nova.scheduler.client.report [None req-79a025c2-964c-42b1-abd7-3dafe654ac62 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Deleted allocations for instance a7419b30-1539-4977-a5d9-93a49d84914a [ 1066.327496] env[61957]: DEBUG oslo_vmware.api [None req-bfdcb579-7ae1-4a05-9d08-d1a10d0fec8a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': task-1278069, 'name': Rename_Task} progress is 99%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.566845] env[61957]: DEBUG nova.compute.manager [None req-0629491b-6b09-4d57-8893-e804be404b21 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: b82af1e9-294a-4ba9-bcad-73b2a2aca86d] Start spawning the instance on the hypervisor. {{(pid=61957) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1066.591030] env[61957]: DEBUG nova.virt.hardware [None req-0629491b-6b09-4d57-8893-e804be404b21 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T17:22:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T17:21:50Z,direct_url=,disk_format='vmdk',id=11c76a2c-f705-470a-ba9d-4657858bab38,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='81d3096d0f094373913fc3dc8f5c3c6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T17:21:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1066.591301] env[61957]: DEBUG nova.virt.hardware [None req-0629491b-6b09-4d57-8893-e804be404b21 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Flavor limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1066.591466] env[61957]: DEBUG nova.virt.hardware [None req-0629491b-6b09-4d57-8893-e804be404b21 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Image limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1066.591650] env[61957]: DEBUG nova.virt.hardware [None req-0629491b-6b09-4d57-8893-e804be404b21 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Flavor pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1066.591801] env[61957]: DEBUG nova.virt.hardware [None req-0629491b-6b09-4d57-8893-e804be404b21 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Image pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1066.591949] env[61957]: DEBUG nova.virt.hardware [None req-0629491b-6b09-4d57-8893-e804be404b21 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1066.592178] env[61957]: DEBUG nova.virt.hardware [None req-0629491b-6b09-4d57-8893-e804be404b21 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1066.592344] env[61957]: DEBUG nova.virt.hardware [None req-0629491b-6b09-4d57-8893-e804be404b21 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1066.592649] env[61957]: DEBUG nova.virt.hardware [None req-0629491b-6b09-4d57-8893-e804be404b21 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Got 1 possible topologies {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1066.592967] env[61957]: DEBUG nova.virt.hardware [None req-0629491b-6b09-4d57-8893-e804be404b21 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1066.593258] env[61957]: DEBUG nova.virt.hardware [None req-0629491b-6b09-4d57-8893-e804be404b21 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1066.594204] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a050f85-2c10-49b3-b51f-987012d8de66 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.602213] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92fddb23-1110-4201-8a87-1adf44eba433 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.786346] env[61957]: DEBUG nova.compute.manager [req-3c80627f-bd2c-447e-b8e1-3d54ccd55960 req-bd1ff402-1f41-4ce1-af7c-44d8b5d5d8c4 service nova] [instance: b82af1e9-294a-4ba9-bcad-73b2a2aca86d] Received event network-vif-plugged-58a3f865-3fc6-462a-b685-fb32e01f6c5e {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1066.786587] env[61957]: DEBUG oslo_concurrency.lockutils [req-3c80627f-bd2c-447e-b8e1-3d54ccd55960 req-bd1ff402-1f41-4ce1-af7c-44d8b5d5d8c4 service nova] Acquiring lock "b82af1e9-294a-4ba9-bcad-73b2a2aca86d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1066.786814] env[61957]: DEBUG oslo_concurrency.lockutils [req-3c80627f-bd2c-447e-b8e1-3d54ccd55960 req-bd1ff402-1f41-4ce1-af7c-44d8b5d5d8c4 service nova] Lock "b82af1e9-294a-4ba9-bcad-73b2a2aca86d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1066.787036] env[61957]: DEBUG oslo_concurrency.lockutils [req-3c80627f-bd2c-447e-b8e1-3d54ccd55960 req-bd1ff402-1f41-4ce1-af7c-44d8b5d5d8c4 service nova] Lock "b82af1e9-294a-4ba9-bcad-73b2a2aca86d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1066.787170] env[61957]: DEBUG nova.compute.manager [req-3c80627f-bd2c-447e-b8e1-3d54ccd55960 req-bd1ff402-1f41-4ce1-af7c-44d8b5d5d8c4 service nova] [instance: b82af1e9-294a-4ba9-bcad-73b2a2aca86d] No waiting events found dispatching network-vif-plugged-58a3f865-3fc6-462a-b685-fb32e01f6c5e {{(pid=61957) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1066.787341] env[61957]: WARNING nova.compute.manager [req-3c80627f-bd2c-447e-b8e1-3d54ccd55960 req-bd1ff402-1f41-4ce1-af7c-44d8b5d5d8c4 service nova] [instance: b82af1e9-294a-4ba9-bcad-73b2a2aca86d] Received unexpected event network-vif-plugged-58a3f865-3fc6-462a-b685-fb32e01f6c5e for instance with vm_state building and task_state spawning. [ 1066.791153] env[61957]: DEBUG oslo_concurrency.lockutils [None req-79a025c2-964c-42b1-abd7-3dafe654ac62 tempest-ServerDiskConfigTestJSON-1478703897 tempest-ServerDiskConfigTestJSON-1478703897-project-member] Lock "a7419b30-1539-4977-a5d9-93a49d84914a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.767s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1066.828202] env[61957]: DEBUG oslo_vmware.api [None req-bfdcb579-7ae1-4a05-9d08-d1a10d0fec8a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': task-1278069, 'name': Rename_Task} progress is 99%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.332163] env[61957]: DEBUG oslo_vmware.api [None req-bfdcb579-7ae1-4a05-9d08-d1a10d0fec8a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': task-1278069, 'name': Rename_Task, 'duration_secs': 1.155353} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.332163] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-bfdcb579-7ae1-4a05-9d08-d1a10d0fec8a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 11d67afa-1d96-4a9b-8439-383d586d8e17] Powering on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1067.332248] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b1254497-53da-4bbe-b7c9-30c4c229746d {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.340067] env[61957]: DEBUG oslo_vmware.api [None req-bfdcb579-7ae1-4a05-9d08-d1a10d0fec8a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Waiting for the task: (returnval){ [ 1067.340067] env[61957]: value = "task-1278070" [ 1067.340067] env[61957]: _type = "Task" [ 1067.340067] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.349057] env[61957]: DEBUG oslo_vmware.api [None req-bfdcb579-7ae1-4a05-9d08-d1a10d0fec8a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': task-1278070, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.351449] env[61957]: DEBUG nova.network.neutron [None req-0629491b-6b09-4d57-8893-e804be404b21 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: b82af1e9-294a-4ba9-bcad-73b2a2aca86d] Successfully updated port: 58a3f865-3fc6-462a-b685-fb32e01f6c5e {{(pid=61957) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1067.356076] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b1744a2-6c84-409d-ab1f-ff09c163f2cc {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.363712] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7cff0ae-23ea-4da1-9b9d-9bd512e377b2 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.395616] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6dc7f383-c237-4824-9ed4-6e2cbdc548d5 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.399116] env[61957]: DEBUG nova.compute.manager [req-356e92c3-5374-4ce1-8b36-22b96479fbe3 req-08d715ad-a6a0-4390-ac0a-598e63954cd6 service nova] [instance: b82af1e9-294a-4ba9-bcad-73b2a2aca86d] Received event network-changed-58a3f865-3fc6-462a-b685-fb32e01f6c5e {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1067.399314] env[61957]: DEBUG nova.compute.manager [req-356e92c3-5374-4ce1-8b36-22b96479fbe3 req-08d715ad-a6a0-4390-ac0a-598e63954cd6 service nova] [instance: b82af1e9-294a-4ba9-bcad-73b2a2aca86d] Refreshing instance network info cache due to event network-changed-58a3f865-3fc6-462a-b685-fb32e01f6c5e. {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1067.399526] env[61957]: DEBUG oslo_concurrency.lockutils [req-356e92c3-5374-4ce1-8b36-22b96479fbe3 req-08d715ad-a6a0-4390-ac0a-598e63954cd6 service nova] Acquiring lock "refresh_cache-b82af1e9-294a-4ba9-bcad-73b2a2aca86d" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1067.399671] env[61957]: DEBUG oslo_concurrency.lockutils [req-356e92c3-5374-4ce1-8b36-22b96479fbe3 req-08d715ad-a6a0-4390-ac0a-598e63954cd6 service nova] Acquired lock "refresh_cache-b82af1e9-294a-4ba9-bcad-73b2a2aca86d" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1067.399836] env[61957]: DEBUG nova.network.neutron [req-356e92c3-5374-4ce1-8b36-22b96479fbe3 req-08d715ad-a6a0-4390-ac0a-598e63954cd6 service nova] [instance: b82af1e9-294a-4ba9-bcad-73b2a2aca86d] Refreshing network info cache for port 58a3f865-3fc6-462a-b685-fb32e01f6c5e {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1067.406654] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb21002c-11fa-4b62-9ade-815ea753471d {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.422495] env[61957]: DEBUG nova.compute.provider_tree [None req-92fb2eb5-36b1-4a05-8eec-33f1fe3376d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1067.849430] env[61957]: DEBUG oslo_vmware.api [None req-bfdcb579-7ae1-4a05-9d08-d1a10d0fec8a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': task-1278070, 'name': PowerOnVM_Task, 'duration_secs': 0.435222} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.849711] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-bfdcb579-7ae1-4a05-9d08-d1a10d0fec8a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 11d67afa-1d96-4a9b-8439-383d586d8e17] Powered on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1067.849901] env[61957]: INFO nova.compute.manager [None req-bfdcb579-7ae1-4a05-9d08-d1a10d0fec8a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 11d67afa-1d96-4a9b-8439-383d586d8e17] Took 8.80 seconds to spawn the instance on the hypervisor. [ 1067.850104] env[61957]: DEBUG nova.compute.manager [None req-bfdcb579-7ae1-4a05-9d08-d1a10d0fec8a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 11d67afa-1d96-4a9b-8439-383d586d8e17] Checking state {{(pid=61957) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1067.850855] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18f89ba1-4c54-424b-b130-69a560ad5d41 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.853403] env[61957]: DEBUG oslo_concurrency.lockutils [None req-0629491b-6b09-4d57-8893-e804be404b21 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Acquiring lock "refresh_cache-b82af1e9-294a-4ba9-bcad-73b2a2aca86d" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1067.925253] env[61957]: DEBUG nova.scheduler.client.report [None req-92fb2eb5-36b1-4a05-8eec-33f1fe3376d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1067.933996] env[61957]: DEBUG nova.network.neutron [req-356e92c3-5374-4ce1-8b36-22b96479fbe3 req-08d715ad-a6a0-4390-ac0a-598e63954cd6 service nova] [instance: b82af1e9-294a-4ba9-bcad-73b2a2aca86d] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1068.011678] env[61957]: DEBUG nova.network.neutron [req-356e92c3-5374-4ce1-8b36-22b96479fbe3 req-08d715ad-a6a0-4390-ac0a-598e63954cd6 service nova] [instance: b82af1e9-294a-4ba9-bcad-73b2a2aca86d] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1068.368868] env[61957]: INFO nova.compute.manager [None req-bfdcb579-7ae1-4a05-9d08-d1a10d0fec8a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 11d67afa-1d96-4a9b-8439-383d586d8e17] Took 15.26 seconds to build instance. [ 1068.429467] env[61957]: DEBUG oslo_concurrency.lockutils [None req-92fb2eb5-36b1-4a05-8eec-33f1fe3376d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.167s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1068.429997] env[61957]: DEBUG nova.compute.manager [None req-92fb2eb5-36b1-4a05-8eec-33f1fe3376d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 84cd5619-4ac9-41ef-9368-a023ad9ae66d] Start building networks asynchronously for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1068.514636] env[61957]: DEBUG oslo_concurrency.lockutils [req-356e92c3-5374-4ce1-8b36-22b96479fbe3 req-08d715ad-a6a0-4390-ac0a-598e63954cd6 service nova] Releasing lock "refresh_cache-b82af1e9-294a-4ba9-bcad-73b2a2aca86d" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1068.515125] env[61957]: DEBUG oslo_concurrency.lockutils [None req-0629491b-6b09-4d57-8893-e804be404b21 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Acquired lock "refresh_cache-b82af1e9-294a-4ba9-bcad-73b2a2aca86d" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1068.515247] env[61957]: DEBUG nova.network.neutron [None req-0629491b-6b09-4d57-8893-e804be404b21 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: b82af1e9-294a-4ba9-bcad-73b2a2aca86d] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1068.871194] env[61957]: DEBUG oslo_concurrency.lockutils [None req-bfdcb579-7ae1-4a05-9d08-d1a10d0fec8a tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Lock "11d67afa-1d96-4a9b-8439-383d586d8e17" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.792s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1068.935133] env[61957]: DEBUG nova.compute.utils [None req-92fb2eb5-36b1-4a05-8eec-33f1fe3376d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Using /dev/sd instead of None {{(pid=61957) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1068.936518] env[61957]: DEBUG nova.compute.manager [None req-92fb2eb5-36b1-4a05-8eec-33f1fe3376d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 84cd5619-4ac9-41ef-9368-a023ad9ae66d] Allocating IP information in the background. {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1068.936691] env[61957]: DEBUG nova.network.neutron [None req-92fb2eb5-36b1-4a05-8eec-33f1fe3376d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 84cd5619-4ac9-41ef-9368-a023ad9ae66d] allocate_for_instance() {{(pid=61957) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1069.005440] env[61957]: DEBUG nova.policy [None req-92fb2eb5-36b1-4a05-8eec-33f1fe3376d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '85d4f8bb3aa64e7285aac42d14087ff1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e16f6dce3f0e44fb96de516f17d4c6f5', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61957) authorize /opt/stack/nova/nova/policy.py:203}} [ 1069.069300] env[61957]: DEBUG nova.network.neutron [None req-0629491b-6b09-4d57-8893-e804be404b21 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: b82af1e9-294a-4ba9-bcad-73b2a2aca86d] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1069.302190] env[61957]: DEBUG nova.network.neutron [None req-0629491b-6b09-4d57-8893-e804be404b21 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: b82af1e9-294a-4ba9-bcad-73b2a2aca86d] Updating instance_info_cache with network_info: [{"id": "58a3f865-3fc6-462a-b685-fb32e01f6c5e", "address": "fa:16:3e:ae:67:c0", "network": {"id": "364cabcb-046d-4f91-b1ee-5dd3adfae6d7", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1432396967-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8713aa35bcb24b86ad0b58ca9fc991ba", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b7a73c01-1bb9-4612-a1a7-16d71b732e81", "external-id": "nsx-vlan-transportzone-711", "segmentation_id": 711, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap58a3f865-3f", "ovs_interfaceid": "58a3f865-3fc6-462a-b685-fb32e01f6c5e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1069.389388] env[61957]: DEBUG nova.network.neutron [None req-92fb2eb5-36b1-4a05-8eec-33f1fe3376d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 84cd5619-4ac9-41ef-9368-a023ad9ae66d] Successfully created port: b0541069-72e9-4384-b4f4-ccf5bc206dab {{(pid=61957) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1069.440285] env[61957]: DEBUG nova.compute.manager [None req-92fb2eb5-36b1-4a05-8eec-33f1fe3376d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 84cd5619-4ac9-41ef-9368-a023ad9ae66d] Start building block device mappings for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1069.646410] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ed26b16b-8a00-4d91-8fb4-6b369fa27bc4 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Acquiring lock "11d67afa-1d96-4a9b-8439-383d586d8e17" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1069.646917] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ed26b16b-8a00-4d91-8fb4-6b369fa27bc4 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Lock "11d67afa-1d96-4a9b-8439-383d586d8e17" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1069.647232] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ed26b16b-8a00-4d91-8fb4-6b369fa27bc4 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Acquiring lock "11d67afa-1d96-4a9b-8439-383d586d8e17-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1069.647450] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ed26b16b-8a00-4d91-8fb4-6b369fa27bc4 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Lock "11d67afa-1d96-4a9b-8439-383d586d8e17-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1069.647636] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ed26b16b-8a00-4d91-8fb4-6b369fa27bc4 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Lock "11d67afa-1d96-4a9b-8439-383d586d8e17-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1069.649922] env[61957]: INFO nova.compute.manager [None req-ed26b16b-8a00-4d91-8fb4-6b369fa27bc4 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 11d67afa-1d96-4a9b-8439-383d586d8e17] Terminating instance [ 1069.651835] env[61957]: DEBUG nova.compute.manager [None req-ed26b16b-8a00-4d91-8fb4-6b369fa27bc4 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 11d67afa-1d96-4a9b-8439-383d586d8e17] Start destroying the instance on the hypervisor. {{(pid=61957) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1069.652050] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-ed26b16b-8a00-4d91-8fb4-6b369fa27bc4 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 11d67afa-1d96-4a9b-8439-383d586d8e17] Destroying instance {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1069.652959] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2df49036-9d9c-40c4-b196-5457974b1825 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.660351] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-ed26b16b-8a00-4d91-8fb4-6b369fa27bc4 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 11d67afa-1d96-4a9b-8439-383d586d8e17] Powering off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1069.660827] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-37bd7302-c5ed-4d1d-9c19-75eeaee290ea {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.667371] env[61957]: DEBUG oslo_vmware.api [None req-ed26b16b-8a00-4d91-8fb4-6b369fa27bc4 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Waiting for the task: (returnval){ [ 1069.667371] env[61957]: value = "task-1278071" [ 1069.667371] env[61957]: _type = "Task" [ 1069.667371] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1069.675343] env[61957]: DEBUG oslo_vmware.api [None req-ed26b16b-8a00-4d91-8fb4-6b369fa27bc4 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': task-1278071, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.804869] env[61957]: DEBUG oslo_concurrency.lockutils [None req-0629491b-6b09-4d57-8893-e804be404b21 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Releasing lock "refresh_cache-b82af1e9-294a-4ba9-bcad-73b2a2aca86d" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1069.805272] env[61957]: DEBUG nova.compute.manager [None req-0629491b-6b09-4d57-8893-e804be404b21 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: b82af1e9-294a-4ba9-bcad-73b2a2aca86d] Instance network_info: |[{"id": "58a3f865-3fc6-462a-b685-fb32e01f6c5e", "address": "fa:16:3e:ae:67:c0", "network": {"id": "364cabcb-046d-4f91-b1ee-5dd3adfae6d7", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1432396967-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8713aa35bcb24b86ad0b58ca9fc991ba", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b7a73c01-1bb9-4612-a1a7-16d71b732e81", "external-id": "nsx-vlan-transportzone-711", "segmentation_id": 711, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap58a3f865-3f", "ovs_interfaceid": "58a3f865-3fc6-462a-b685-fb32e01f6c5e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1069.805737] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-0629491b-6b09-4d57-8893-e804be404b21 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: b82af1e9-294a-4ba9-bcad-73b2a2aca86d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ae:67:c0', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'b7a73c01-1bb9-4612-a1a7-16d71b732e81', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '58a3f865-3fc6-462a-b685-fb32e01f6c5e', 'vif_model': 'vmxnet3'}] {{(pid=61957) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1069.813576] env[61957]: DEBUG oslo.service.loopingcall [None req-0629491b-6b09-4d57-8893-e804be404b21 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1069.813804] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b82af1e9-294a-4ba9-bcad-73b2a2aca86d] Creating VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1069.814047] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5f5077cd-53d1-41aa-8573-b04b9a389bc3 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.833925] env[61957]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1069.833925] env[61957]: value = "task-1278072" [ 1069.833925] env[61957]: _type = "Task" [ 1069.833925] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1069.843115] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1278072, 'name': CreateVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.176778] env[61957]: DEBUG oslo_vmware.api [None req-ed26b16b-8a00-4d91-8fb4-6b369fa27bc4 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': task-1278071, 'name': PowerOffVM_Task, 'duration_secs': 0.157886} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1070.177074] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-ed26b16b-8a00-4d91-8fb4-6b369fa27bc4 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 11d67afa-1d96-4a9b-8439-383d586d8e17] Powered off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1070.177238] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-ed26b16b-8a00-4d91-8fb4-6b369fa27bc4 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 11d67afa-1d96-4a9b-8439-383d586d8e17] Unregistering the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1070.177489] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8d4b2873-06d3-44db-91d8-9f8add17cdd7 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.246433] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-ed26b16b-8a00-4d91-8fb4-6b369fa27bc4 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 11d67afa-1d96-4a9b-8439-383d586d8e17] Unregistered the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1070.246663] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-ed26b16b-8a00-4d91-8fb4-6b369fa27bc4 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 11d67afa-1d96-4a9b-8439-383d586d8e17] Deleting contents of the VM from datastore datastore2 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1070.246855] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-ed26b16b-8a00-4d91-8fb4-6b369fa27bc4 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Deleting the datastore file [datastore2] 11d67afa-1d96-4a9b-8439-383d586d8e17 {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1070.247174] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-239c6b79-ca76-4524-bd37-91a3faa193c2 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.255278] env[61957]: DEBUG oslo_vmware.api [None req-ed26b16b-8a00-4d91-8fb4-6b369fa27bc4 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Waiting for the task: (returnval){ [ 1070.255278] env[61957]: value = "task-1278074" [ 1070.255278] env[61957]: _type = "Task" [ 1070.255278] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1070.262701] env[61957]: DEBUG oslo_vmware.api [None req-ed26b16b-8a00-4d91-8fb4-6b369fa27bc4 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': task-1278074, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.344783] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1278072, 'name': CreateVM_Task, 'duration_secs': 0.271218} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1070.344996] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b82af1e9-294a-4ba9-bcad-73b2a2aca86d] Created VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1070.345703] env[61957]: DEBUG oslo_concurrency.lockutils [None req-0629491b-6b09-4d57-8893-e804be404b21 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1070.345879] env[61957]: DEBUG oslo_concurrency.lockutils [None req-0629491b-6b09-4d57-8893-e804be404b21 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1070.346221] env[61957]: DEBUG oslo_concurrency.lockutils [None req-0629491b-6b09-4d57-8893-e804be404b21 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1070.346471] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-67da1104-76d2-4257-a567-84f000a95e39 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.350904] env[61957]: DEBUG oslo_vmware.api [None req-0629491b-6b09-4d57-8893-e804be404b21 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Waiting for the task: (returnval){ [ 1070.350904] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]5288033c-499f-9efd-c17c-13ce1f1aaa1d" [ 1070.350904] env[61957]: _type = "Task" [ 1070.350904] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1070.358205] env[61957]: DEBUG oslo_vmware.api [None req-0629491b-6b09-4d57-8893-e804be404b21 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]5288033c-499f-9efd-c17c-13ce1f1aaa1d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.450608] env[61957]: DEBUG nova.compute.manager [None req-92fb2eb5-36b1-4a05-8eec-33f1fe3376d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 84cd5619-4ac9-41ef-9368-a023ad9ae66d] Start spawning the instance on the hypervisor. {{(pid=61957) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1070.475674] env[61957]: DEBUG nova.virt.hardware [None req-92fb2eb5-36b1-4a05-8eec-33f1fe3376d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T17:22:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T17:21:50Z,direct_url=,disk_format='vmdk',id=11c76a2c-f705-470a-ba9d-4657858bab38,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='81d3096d0f094373913fc3dc8f5c3c6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T17:21:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1070.475835] env[61957]: DEBUG nova.virt.hardware [None req-92fb2eb5-36b1-4a05-8eec-33f1fe3376d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Flavor limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1070.475900] env[61957]: DEBUG nova.virt.hardware [None req-92fb2eb5-36b1-4a05-8eec-33f1fe3376d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Image limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1070.476095] env[61957]: DEBUG nova.virt.hardware [None req-92fb2eb5-36b1-4a05-8eec-33f1fe3376d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Flavor pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1070.476256] env[61957]: DEBUG nova.virt.hardware [None req-92fb2eb5-36b1-4a05-8eec-33f1fe3376d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Image pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1070.476407] env[61957]: DEBUG nova.virt.hardware [None req-92fb2eb5-36b1-4a05-8eec-33f1fe3376d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1070.476622] env[61957]: DEBUG nova.virt.hardware [None req-92fb2eb5-36b1-4a05-8eec-33f1fe3376d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1070.477663] env[61957]: DEBUG nova.virt.hardware [None req-92fb2eb5-36b1-4a05-8eec-33f1fe3376d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1070.477663] env[61957]: DEBUG nova.virt.hardware [None req-92fb2eb5-36b1-4a05-8eec-33f1fe3376d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Got 1 possible topologies {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1070.477663] env[61957]: DEBUG nova.virt.hardware [None req-92fb2eb5-36b1-4a05-8eec-33f1fe3376d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1070.477663] env[61957]: DEBUG nova.virt.hardware [None req-92fb2eb5-36b1-4a05-8eec-33f1fe3376d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1070.478154] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dba59467-01d9-42a2-9d93-892c8a4c1d18 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.485724] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe8829bc-88ce-41d9-9447-067efe18bcd9 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.765670] env[61957]: DEBUG oslo_vmware.api [None req-ed26b16b-8a00-4d91-8fb4-6b369fa27bc4 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': task-1278074, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.145466} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1070.765970] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-ed26b16b-8a00-4d91-8fb4-6b369fa27bc4 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Deleted the datastore file {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1070.766181] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-ed26b16b-8a00-4d91-8fb4-6b369fa27bc4 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 11d67afa-1d96-4a9b-8439-383d586d8e17] Deleted contents of the VM from datastore datastore2 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1070.766364] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-ed26b16b-8a00-4d91-8fb4-6b369fa27bc4 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 11d67afa-1d96-4a9b-8439-383d586d8e17] Instance destroyed {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1070.766543] env[61957]: INFO nova.compute.manager [None req-ed26b16b-8a00-4d91-8fb4-6b369fa27bc4 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 11d67afa-1d96-4a9b-8439-383d586d8e17] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1070.766798] env[61957]: DEBUG oslo.service.loopingcall [None req-ed26b16b-8a00-4d91-8fb4-6b369fa27bc4 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1070.767040] env[61957]: DEBUG nova.compute.manager [-] [instance: 11d67afa-1d96-4a9b-8439-383d586d8e17] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1070.767143] env[61957]: DEBUG nova.network.neutron [-] [instance: 11d67afa-1d96-4a9b-8439-383d586d8e17] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1070.792352] env[61957]: DEBUG nova.compute.manager [req-7624c604-2cfc-460e-af1c-ec678c64d5be req-7d96e0b8-b5a5-4b99-befd-c81cafce6e34 service nova] [instance: 84cd5619-4ac9-41ef-9368-a023ad9ae66d] Received event network-vif-plugged-b0541069-72e9-4384-b4f4-ccf5bc206dab {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1070.792420] env[61957]: DEBUG oslo_concurrency.lockutils [req-7624c604-2cfc-460e-af1c-ec678c64d5be req-7d96e0b8-b5a5-4b99-befd-c81cafce6e34 service nova] Acquiring lock "84cd5619-4ac9-41ef-9368-a023ad9ae66d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1070.792715] env[61957]: DEBUG oslo_concurrency.lockutils [req-7624c604-2cfc-460e-af1c-ec678c64d5be req-7d96e0b8-b5a5-4b99-befd-c81cafce6e34 service nova] Lock "84cd5619-4ac9-41ef-9368-a023ad9ae66d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1070.792991] env[61957]: DEBUG oslo_concurrency.lockutils [req-7624c604-2cfc-460e-af1c-ec678c64d5be req-7d96e0b8-b5a5-4b99-befd-c81cafce6e34 service nova] Lock "84cd5619-4ac9-41ef-9368-a023ad9ae66d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1070.793089] env[61957]: DEBUG nova.compute.manager [req-7624c604-2cfc-460e-af1c-ec678c64d5be req-7d96e0b8-b5a5-4b99-befd-c81cafce6e34 service nova] [instance: 84cd5619-4ac9-41ef-9368-a023ad9ae66d] No waiting events found dispatching network-vif-plugged-b0541069-72e9-4384-b4f4-ccf5bc206dab {{(pid=61957) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1070.793257] env[61957]: WARNING nova.compute.manager [req-7624c604-2cfc-460e-af1c-ec678c64d5be req-7d96e0b8-b5a5-4b99-befd-c81cafce6e34 service nova] [instance: 84cd5619-4ac9-41ef-9368-a023ad9ae66d] Received unexpected event network-vif-plugged-b0541069-72e9-4384-b4f4-ccf5bc206dab for instance with vm_state building and task_state spawning. [ 1070.868251] env[61957]: DEBUG oslo_vmware.api [None req-0629491b-6b09-4d57-8893-e804be404b21 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]5288033c-499f-9efd-c17c-13ce1f1aaa1d, 'name': SearchDatastore_Task, 'duration_secs': 0.010879} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1070.868251] env[61957]: DEBUG oslo_concurrency.lockutils [None req-0629491b-6b09-4d57-8893-e804be404b21 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1070.868251] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-0629491b-6b09-4d57-8893-e804be404b21 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: b82af1e9-294a-4ba9-bcad-73b2a2aca86d] Processing image 11c76a2c-f705-470a-ba9d-4657858bab38 {{(pid=61957) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1070.868251] env[61957]: DEBUG oslo_concurrency.lockutils [None req-0629491b-6b09-4d57-8893-e804be404b21 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1070.869029] env[61957]: DEBUG oslo_concurrency.lockutils [None req-0629491b-6b09-4d57-8893-e804be404b21 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1070.869029] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-0629491b-6b09-4d57-8893-e804be404b21 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1070.869029] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-30ebf166-f32f-4a40-bed2-dd3237a934ba {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.878421] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-0629491b-6b09-4d57-8893-e804be404b21 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1070.878421] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-0629491b-6b09-4d57-8893-e804be404b21 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61957) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1070.878560] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ea078bb5-11c2-4ea4-b6fd-7c4e39014477 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.887020] env[61957]: DEBUG oslo_vmware.api [None req-0629491b-6b09-4d57-8893-e804be404b21 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Waiting for the task: (returnval){ [ 1070.887020] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]520d6469-0534-b832-4630-7b85f04ddde7" [ 1070.887020] env[61957]: _type = "Task" [ 1070.887020] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1070.892727] env[61957]: DEBUG oslo_vmware.api [None req-0629491b-6b09-4d57-8893-e804be404b21 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]520d6469-0534-b832-4630-7b85f04ddde7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.999514] env[61957]: DEBUG nova.network.neutron [None req-92fb2eb5-36b1-4a05-8eec-33f1fe3376d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 84cd5619-4ac9-41ef-9368-a023ad9ae66d] Successfully updated port: b0541069-72e9-4384-b4f4-ccf5bc206dab {{(pid=61957) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1071.395741] env[61957]: DEBUG oslo_vmware.api [None req-0629491b-6b09-4d57-8893-e804be404b21 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]520d6469-0534-b832-4630-7b85f04ddde7, 'name': SearchDatastore_Task, 'duration_secs': 0.007868} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1071.396624] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fefc23a8-2bed-4cd9-b371-b45220693ffd {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.402110] env[61957]: DEBUG oslo_vmware.api [None req-0629491b-6b09-4d57-8893-e804be404b21 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Waiting for the task: (returnval){ [ 1071.402110] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52bda413-c54e-687b-2481-119c58aa5f5c" [ 1071.402110] env[61957]: _type = "Task" [ 1071.402110] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1071.409561] env[61957]: DEBUG oslo_vmware.api [None req-0629491b-6b09-4d57-8893-e804be404b21 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52bda413-c54e-687b-2481-119c58aa5f5c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.504508] env[61957]: DEBUG oslo_concurrency.lockutils [None req-92fb2eb5-36b1-4a05-8eec-33f1fe3376d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Acquiring lock "refresh_cache-84cd5619-4ac9-41ef-9368-a023ad9ae66d" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1071.504571] env[61957]: DEBUG oslo_concurrency.lockutils [None req-92fb2eb5-36b1-4a05-8eec-33f1fe3376d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Acquired lock "refresh_cache-84cd5619-4ac9-41ef-9368-a023ad9ae66d" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1071.504714] env[61957]: DEBUG nova.network.neutron [None req-92fb2eb5-36b1-4a05-8eec-33f1fe3376d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 84cd5619-4ac9-41ef-9368-a023ad9ae66d] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1071.601075] env[61957]: DEBUG nova.network.neutron [-] [instance: 11d67afa-1d96-4a9b-8439-383d586d8e17] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1071.912425] env[61957]: DEBUG oslo_vmware.api [None req-0629491b-6b09-4d57-8893-e804be404b21 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52bda413-c54e-687b-2481-119c58aa5f5c, 'name': SearchDatastore_Task, 'duration_secs': 0.008174} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1071.912798] env[61957]: DEBUG oslo_concurrency.lockutils [None req-0629491b-6b09-4d57-8893-e804be404b21 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1071.913095] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-0629491b-6b09-4d57-8893-e804be404b21 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore2] b82af1e9-294a-4ba9-bcad-73b2a2aca86d/b82af1e9-294a-4ba9-bcad-73b2a2aca86d.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1071.913377] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-39f94ad7-701b-47f3-9838-d63edb46b52d {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.919951] env[61957]: DEBUG oslo_vmware.api [None req-0629491b-6b09-4d57-8893-e804be404b21 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Waiting for the task: (returnval){ [ 1071.919951] env[61957]: value = "task-1278075" [ 1071.919951] env[61957]: _type = "Task" [ 1071.919951] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1071.927437] env[61957]: DEBUG oslo_vmware.api [None req-0629491b-6b09-4d57-8893-e804be404b21 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': task-1278075, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.036023] env[61957]: DEBUG nova.network.neutron [None req-92fb2eb5-36b1-4a05-8eec-33f1fe3376d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 84cd5619-4ac9-41ef-9368-a023ad9ae66d] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1072.104095] env[61957]: INFO nova.compute.manager [-] [instance: 11d67afa-1d96-4a9b-8439-383d586d8e17] Took 1.34 seconds to deallocate network for instance. [ 1072.199451] env[61957]: DEBUG nova.network.neutron [None req-92fb2eb5-36b1-4a05-8eec-33f1fe3376d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 84cd5619-4ac9-41ef-9368-a023ad9ae66d] Updating instance_info_cache with network_info: [{"id": "b0541069-72e9-4384-b4f4-ccf5bc206dab", "address": "fa:16:3e:d1:ff:5a", "network": {"id": "efe08823-51fe-4620-a849-5345e262cc71", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1035436819-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e16f6dce3f0e44fb96de516f17d4c6f5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "667a2e97-c1be-421d-9941-6b84c2629b43", "external-id": "nsx-vlan-transportzone-484", "segmentation_id": 484, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb0541069-72", "ovs_interfaceid": "b0541069-72e9-4384-b4f4-ccf5bc206dab", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1072.433320] env[61957]: DEBUG oslo_vmware.api [None req-0629491b-6b09-4d57-8893-e804be404b21 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': task-1278075, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.421413} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1072.433725] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-0629491b-6b09-4d57-8893-e804be404b21 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore2] b82af1e9-294a-4ba9-bcad-73b2a2aca86d/b82af1e9-294a-4ba9-bcad-73b2a2aca86d.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1072.434011] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-0629491b-6b09-4d57-8893-e804be404b21 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: b82af1e9-294a-4ba9-bcad-73b2a2aca86d] Extending root virtual disk to 1048576 {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1072.434301] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a923e974-93a7-4e3f-8fa1-fe90038c6ef9 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.441069] env[61957]: DEBUG oslo_vmware.api [None req-0629491b-6b09-4d57-8893-e804be404b21 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Waiting for the task: (returnval){ [ 1072.441069] env[61957]: value = "task-1278076" [ 1072.441069] env[61957]: _type = "Task" [ 1072.441069] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1072.449991] env[61957]: DEBUG oslo_vmware.api [None req-0629491b-6b09-4d57-8893-e804be404b21 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': task-1278076, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.614329] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ed26b16b-8a00-4d91-8fb4-6b369fa27bc4 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1072.614489] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ed26b16b-8a00-4d91-8fb4-6b369fa27bc4 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1072.614703] env[61957]: DEBUG nova.objects.instance [None req-ed26b16b-8a00-4d91-8fb4-6b369fa27bc4 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Lazy-loading 'resources' on Instance uuid 11d67afa-1d96-4a9b-8439-383d586d8e17 {{(pid=61957) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1072.702624] env[61957]: DEBUG oslo_concurrency.lockutils [None req-92fb2eb5-36b1-4a05-8eec-33f1fe3376d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Releasing lock "refresh_cache-84cd5619-4ac9-41ef-9368-a023ad9ae66d" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1072.702931] env[61957]: DEBUG nova.compute.manager [None req-92fb2eb5-36b1-4a05-8eec-33f1fe3376d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 84cd5619-4ac9-41ef-9368-a023ad9ae66d] Instance network_info: |[{"id": "b0541069-72e9-4384-b4f4-ccf5bc206dab", "address": "fa:16:3e:d1:ff:5a", "network": {"id": "efe08823-51fe-4620-a849-5345e262cc71", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1035436819-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e16f6dce3f0e44fb96de516f17d4c6f5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "667a2e97-c1be-421d-9941-6b84c2629b43", "external-id": "nsx-vlan-transportzone-484", "segmentation_id": 484, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb0541069-72", "ovs_interfaceid": "b0541069-72e9-4384-b4f4-ccf5bc206dab", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1072.703374] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-92fb2eb5-36b1-4a05-8eec-33f1fe3376d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 84cd5619-4ac9-41ef-9368-a023ad9ae66d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d1:ff:5a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '667a2e97-c1be-421d-9941-6b84c2629b43', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b0541069-72e9-4384-b4f4-ccf5bc206dab', 'vif_model': 'vmxnet3'}] {{(pid=61957) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1072.710745] env[61957]: DEBUG oslo.service.loopingcall [None req-92fb2eb5-36b1-4a05-8eec-33f1fe3376d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1072.710953] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 84cd5619-4ac9-41ef-9368-a023ad9ae66d] Creating VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1072.711213] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4cfe362f-3033-4411-beb2-77148dc9139d {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.731866] env[61957]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1072.731866] env[61957]: value = "task-1278077" [ 1072.731866] env[61957]: _type = "Task" [ 1072.731866] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1072.739064] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1278077, 'name': CreateVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.822565] env[61957]: DEBUG nova.compute.manager [req-83992fc9-1266-48a0-a583-29aa97e1fc7a req-0ccf6025-99f7-41da-ad30-444ce44692ae service nova] [instance: 84cd5619-4ac9-41ef-9368-a023ad9ae66d] Received event network-changed-b0541069-72e9-4384-b4f4-ccf5bc206dab {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1072.822764] env[61957]: DEBUG nova.compute.manager [req-83992fc9-1266-48a0-a583-29aa97e1fc7a req-0ccf6025-99f7-41da-ad30-444ce44692ae service nova] [instance: 84cd5619-4ac9-41ef-9368-a023ad9ae66d] Refreshing instance network info cache due to event network-changed-b0541069-72e9-4384-b4f4-ccf5bc206dab. {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1072.823028] env[61957]: DEBUG oslo_concurrency.lockutils [req-83992fc9-1266-48a0-a583-29aa97e1fc7a req-0ccf6025-99f7-41da-ad30-444ce44692ae service nova] Acquiring lock "refresh_cache-84cd5619-4ac9-41ef-9368-a023ad9ae66d" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1072.823187] env[61957]: DEBUG oslo_concurrency.lockutils [req-83992fc9-1266-48a0-a583-29aa97e1fc7a req-0ccf6025-99f7-41da-ad30-444ce44692ae service nova] Acquired lock "refresh_cache-84cd5619-4ac9-41ef-9368-a023ad9ae66d" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1072.823351] env[61957]: DEBUG nova.network.neutron [req-83992fc9-1266-48a0-a583-29aa97e1fc7a req-0ccf6025-99f7-41da-ad30-444ce44692ae service nova] [instance: 84cd5619-4ac9-41ef-9368-a023ad9ae66d] Refreshing network info cache for port b0541069-72e9-4384-b4f4-ccf5bc206dab {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1072.952933] env[61957]: DEBUG oslo_vmware.api [None req-0629491b-6b09-4d57-8893-e804be404b21 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': task-1278076, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.059621} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1072.953255] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-0629491b-6b09-4d57-8893-e804be404b21 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: b82af1e9-294a-4ba9-bcad-73b2a2aca86d] Extended root virtual disk {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1072.954101] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-945d5d93-8db5-4fd2-affc-a0b578270e1d {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.977649] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-0629491b-6b09-4d57-8893-e804be404b21 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: b82af1e9-294a-4ba9-bcad-73b2a2aca86d] Reconfiguring VM instance instance-00000067 to attach disk [datastore2] b82af1e9-294a-4ba9-bcad-73b2a2aca86d/b82af1e9-294a-4ba9-bcad-73b2a2aca86d.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1072.977937] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-584264d9-5929-4177-bd21-056bed1e8c47 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.997175] env[61957]: DEBUG oslo_vmware.api [None req-0629491b-6b09-4d57-8893-e804be404b21 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Waiting for the task: (returnval){ [ 1072.997175] env[61957]: value = "task-1278078" [ 1072.997175] env[61957]: _type = "Task" [ 1072.997175] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1073.007011] env[61957]: DEBUG oslo_vmware.api [None req-0629491b-6b09-4d57-8893-e804be404b21 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': task-1278078, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.210388] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-440aa570-147c-4674-a8cd-e68c40b2e606 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.217763] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-541f827a-50a2-4fcb-83ff-cbb2ee83535e {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.250068] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-156d5aaa-1553-4ff7-bafc-0fe82d8e83d4 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.257008] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1278077, 'name': CreateVM_Task, 'duration_secs': 0.276311} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1073.258792] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 84cd5619-4ac9-41ef-9368-a023ad9ae66d] Created VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1073.259505] env[61957]: DEBUG oslo_concurrency.lockutils [None req-92fb2eb5-36b1-4a05-8eec-33f1fe3376d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1073.259672] env[61957]: DEBUG oslo_concurrency.lockutils [None req-92fb2eb5-36b1-4a05-8eec-33f1fe3376d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1073.259992] env[61957]: DEBUG oslo_concurrency.lockutils [None req-92fb2eb5-36b1-4a05-8eec-33f1fe3376d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1073.261178] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc81d0f6-ebd7-4ce6-b31e-847d1eccfbbe {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.264575] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6b7ac393-dc8d-4af2-81b3-18ff21170f53 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.268834] env[61957]: DEBUG oslo_vmware.api [None req-92fb2eb5-36b1-4a05-8eec-33f1fe3376d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Waiting for the task: (returnval){ [ 1073.268834] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]5227e96f-c519-e5b7-922c-d1e67e86a3d8" [ 1073.268834] env[61957]: _type = "Task" [ 1073.268834] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1073.277524] env[61957]: DEBUG nova.compute.provider_tree [None req-ed26b16b-8a00-4d91-8fb4-6b369fa27bc4 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Updating inventory in ProviderTree for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1073.285988] env[61957]: DEBUG oslo_vmware.api [None req-92fb2eb5-36b1-4a05-8eec-33f1fe3376d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]5227e96f-c519-e5b7-922c-d1e67e86a3d8, 'name': SearchDatastore_Task, 'duration_secs': 0.008208} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1073.286749] env[61957]: DEBUG oslo_concurrency.lockutils [None req-92fb2eb5-36b1-4a05-8eec-33f1fe3376d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1073.286991] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-92fb2eb5-36b1-4a05-8eec-33f1fe3376d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 84cd5619-4ac9-41ef-9368-a023ad9ae66d] Processing image 11c76a2c-f705-470a-ba9d-4657858bab38 {{(pid=61957) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1073.287241] env[61957]: DEBUG oslo_concurrency.lockutils [None req-92fb2eb5-36b1-4a05-8eec-33f1fe3376d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1073.287389] env[61957]: DEBUG oslo_concurrency.lockutils [None req-92fb2eb5-36b1-4a05-8eec-33f1fe3376d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1073.287564] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-92fb2eb5-36b1-4a05-8eec-33f1fe3376d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1073.288050] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5e3b1e38-075b-4af7-8e0b-5b1db867c2df {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.295468] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-92fb2eb5-36b1-4a05-8eec-33f1fe3376d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1073.295691] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-92fb2eb5-36b1-4a05-8eec-33f1fe3376d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61957) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1073.296410] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a91c05ec-c9a2-49f7-9141-60a92d97bc7a {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.300820] env[61957]: DEBUG oslo_vmware.api [None req-92fb2eb5-36b1-4a05-8eec-33f1fe3376d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Waiting for the task: (returnval){ [ 1073.300820] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52243d3b-38e9-7e53-1fc3-073eca5d1064" [ 1073.300820] env[61957]: _type = "Task" [ 1073.300820] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1073.308426] env[61957]: DEBUG oslo_vmware.api [None req-92fb2eb5-36b1-4a05-8eec-33f1fe3376d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52243d3b-38e9-7e53-1fc3-073eca5d1064, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.507368] env[61957]: DEBUG oslo_vmware.api [None req-0629491b-6b09-4d57-8893-e804be404b21 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': task-1278078, 'name': ReconfigVM_Task, 'duration_secs': 0.245373} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1073.508162] env[61957]: DEBUG nova.network.neutron [req-83992fc9-1266-48a0-a583-29aa97e1fc7a req-0ccf6025-99f7-41da-ad30-444ce44692ae service nova] [instance: 84cd5619-4ac9-41ef-9368-a023ad9ae66d] Updated VIF entry in instance network info cache for port b0541069-72e9-4384-b4f4-ccf5bc206dab. {{(pid=61957) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1073.508476] env[61957]: DEBUG nova.network.neutron [req-83992fc9-1266-48a0-a583-29aa97e1fc7a req-0ccf6025-99f7-41da-ad30-444ce44692ae service nova] [instance: 84cd5619-4ac9-41ef-9368-a023ad9ae66d] Updating instance_info_cache with network_info: [{"id": "b0541069-72e9-4384-b4f4-ccf5bc206dab", "address": "fa:16:3e:d1:ff:5a", "network": {"id": "efe08823-51fe-4620-a849-5345e262cc71", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1035436819-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e16f6dce3f0e44fb96de516f17d4c6f5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "667a2e97-c1be-421d-9941-6b84c2629b43", "external-id": "nsx-vlan-transportzone-484", "segmentation_id": 484, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb0541069-72", "ovs_interfaceid": "b0541069-72e9-4384-b4f4-ccf5bc206dab", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1073.509639] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-0629491b-6b09-4d57-8893-e804be404b21 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: b82af1e9-294a-4ba9-bcad-73b2a2aca86d] Reconfigured VM instance instance-00000067 to attach disk [datastore2] b82af1e9-294a-4ba9-bcad-73b2a2aca86d/b82af1e9-294a-4ba9-bcad-73b2a2aca86d.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1073.510403] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3191cd84-2dbb-49b5-a4e7-6c32efc76807 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.516694] env[61957]: DEBUG oslo_vmware.api [None req-0629491b-6b09-4d57-8893-e804be404b21 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Waiting for the task: (returnval){ [ 1073.516694] env[61957]: value = "task-1278079" [ 1073.516694] env[61957]: _type = "Task" [ 1073.516694] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1073.526350] env[61957]: DEBUG oslo_vmware.api [None req-0629491b-6b09-4d57-8893-e804be404b21 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': task-1278079, 'name': Rename_Task} progress is 5%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.799732] env[61957]: ERROR nova.scheduler.client.report [None req-ed26b16b-8a00-4d91-8fb4-6b369fa27bc4 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [req-02de52fd-0db8-4c53-bc90-7b16b7fe7003] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 0ceb6c9e-61c6-496d-8579-9d32627e96da. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-02de52fd-0db8-4c53-bc90-7b16b7fe7003"}]} [ 1073.810996] env[61957]: DEBUG oslo_vmware.api [None req-92fb2eb5-36b1-4a05-8eec-33f1fe3376d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52243d3b-38e9-7e53-1fc3-073eca5d1064, 'name': SearchDatastore_Task, 'duration_secs': 0.007671} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1073.811778] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9f8711bb-1623-4061-8556-ff2f43f15847 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.816685] env[61957]: DEBUG oslo_vmware.api [None req-92fb2eb5-36b1-4a05-8eec-33f1fe3376d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Waiting for the task: (returnval){ [ 1073.816685] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]5274a103-0dd2-af1e-2902-7d248426f00f" [ 1073.816685] env[61957]: _type = "Task" [ 1073.816685] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1073.817580] env[61957]: DEBUG nova.scheduler.client.report [None req-ed26b16b-8a00-4d91-8fb4-6b369fa27bc4 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Refreshing inventories for resource provider 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 1073.827804] env[61957]: DEBUG oslo_vmware.api [None req-92fb2eb5-36b1-4a05-8eec-33f1fe3376d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]5274a103-0dd2-af1e-2902-7d248426f00f, 'name': SearchDatastore_Task, 'duration_secs': 0.008659} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1073.828065] env[61957]: DEBUG oslo_concurrency.lockutils [None req-92fb2eb5-36b1-4a05-8eec-33f1fe3376d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1073.828408] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-92fb2eb5-36b1-4a05-8eec-33f1fe3376d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore2] 84cd5619-4ac9-41ef-9368-a023ad9ae66d/84cd5619-4ac9-41ef-9368-a023ad9ae66d.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1073.828540] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9534862c-6f73-405c-b62f-fdc1d4b42780 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.830783] env[61957]: DEBUG nova.scheduler.client.report [None req-ed26b16b-8a00-4d91-8fb4-6b369fa27bc4 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Updating ProviderTree inventory for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 1073.830984] env[61957]: DEBUG nova.compute.provider_tree [None req-ed26b16b-8a00-4d91-8fb4-6b369fa27bc4 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Updating inventory in ProviderTree for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1073.837094] env[61957]: DEBUG oslo_vmware.api [None req-92fb2eb5-36b1-4a05-8eec-33f1fe3376d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Waiting for the task: (returnval){ [ 1073.837094] env[61957]: value = "task-1278080" [ 1073.837094] env[61957]: _type = "Task" [ 1073.837094] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1073.841027] env[61957]: DEBUG nova.scheduler.client.report [None req-ed26b16b-8a00-4d91-8fb4-6b369fa27bc4 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Refreshing aggregate associations for resource provider 0ceb6c9e-61c6-496d-8579-9d32627e96da, aggregates: None {{(pid=61957) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 1073.845545] env[61957]: DEBUG oslo_vmware.api [None req-92fb2eb5-36b1-4a05-8eec-33f1fe3376d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1278080, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.857512] env[61957]: DEBUG nova.scheduler.client.report [None req-ed26b16b-8a00-4d91-8fb4-6b369fa27bc4 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Refreshing trait associations for resource provider 0ceb6c9e-61c6-496d-8579-9d32627e96da, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO,HW_ARCH_X86_64,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=61957) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 1073.933102] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a6d169a-e757-4080-b186-385ce7cd5e75 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.940385] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ef8c62f-b1b3-4817-b3f2-4bb7207634ae {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.971591] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11e688e5-aefc-48a6-a949-49df0bce0eab {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.978777] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-293eaf1e-c59f-409b-a75f-dfb211beb643 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.991811] env[61957]: DEBUG nova.compute.provider_tree [None req-ed26b16b-8a00-4d91-8fb4-6b369fa27bc4 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Updating inventory in ProviderTree for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1074.011060] env[61957]: DEBUG oslo_concurrency.lockutils [req-83992fc9-1266-48a0-a583-29aa97e1fc7a req-0ccf6025-99f7-41da-ad30-444ce44692ae service nova] Releasing lock "refresh_cache-84cd5619-4ac9-41ef-9368-a023ad9ae66d" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1074.011393] env[61957]: DEBUG nova.compute.manager [req-83992fc9-1266-48a0-a583-29aa97e1fc7a req-0ccf6025-99f7-41da-ad30-444ce44692ae service nova] [instance: 11d67afa-1d96-4a9b-8439-383d586d8e17] Received event network-vif-deleted-10ee2afa-334a-45ec-9b58-02e6ee3af703 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1074.026909] env[61957]: DEBUG oslo_vmware.api [None req-0629491b-6b09-4d57-8893-e804be404b21 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': task-1278079, 'name': Rename_Task, 'duration_secs': 0.155296} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1074.027276] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-0629491b-6b09-4d57-8893-e804be404b21 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: b82af1e9-294a-4ba9-bcad-73b2a2aca86d] Powering on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1074.027540] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-24578b34-92d4-426c-85d6-c1d5df1317bb {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.034769] env[61957]: DEBUG oslo_vmware.api [None req-0629491b-6b09-4d57-8893-e804be404b21 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Waiting for the task: (returnval){ [ 1074.034769] env[61957]: value = "task-1278081" [ 1074.034769] env[61957]: _type = "Task" [ 1074.034769] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1074.044130] env[61957]: DEBUG oslo_vmware.api [None req-0629491b-6b09-4d57-8893-e804be404b21 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': task-1278081, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.347466] env[61957]: DEBUG oslo_vmware.api [None req-92fb2eb5-36b1-4a05-8eec-33f1fe3376d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1278080, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.523118] env[61957]: DEBUG nova.scheduler.client.report [None req-ed26b16b-8a00-4d91-8fb4-6b369fa27bc4 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Updated inventory for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da with generation 126 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 1074.523421] env[61957]: DEBUG nova.compute.provider_tree [None req-ed26b16b-8a00-4d91-8fb4-6b369fa27bc4 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Updating resource provider 0ceb6c9e-61c6-496d-8579-9d32627e96da generation from 126 to 127 during operation: update_inventory {{(pid=61957) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1074.523682] env[61957]: DEBUG nova.compute.provider_tree [None req-ed26b16b-8a00-4d91-8fb4-6b369fa27bc4 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Updating inventory in ProviderTree for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1074.545747] env[61957]: DEBUG oslo_vmware.api [None req-0629491b-6b09-4d57-8893-e804be404b21 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': task-1278081, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.849336] env[61957]: DEBUG oslo_vmware.api [None req-92fb2eb5-36b1-4a05-8eec-33f1fe3376d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1278080, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.580918} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1074.849634] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-92fb2eb5-36b1-4a05-8eec-33f1fe3376d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore2] 84cd5619-4ac9-41ef-9368-a023ad9ae66d/84cd5619-4ac9-41ef-9368-a023ad9ae66d.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1074.849952] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-92fb2eb5-36b1-4a05-8eec-33f1fe3376d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 84cd5619-4ac9-41ef-9368-a023ad9ae66d] Extending root virtual disk to 1048576 {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1074.850241] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ddd41ece-f17a-4dbd-a4eb-bf93bdaa90a4 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.857257] env[61957]: DEBUG oslo_vmware.api [None req-92fb2eb5-36b1-4a05-8eec-33f1fe3376d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Waiting for the task: (returnval){ [ 1074.857257] env[61957]: value = "task-1278082" [ 1074.857257] env[61957]: _type = "Task" [ 1074.857257] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1074.866148] env[61957]: DEBUG oslo_vmware.api [None req-92fb2eb5-36b1-4a05-8eec-33f1fe3376d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1278082, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1075.028836] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ed26b16b-8a00-4d91-8fb4-6b369fa27bc4 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.414s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1075.044755] env[61957]: DEBUG oslo_vmware.api [None req-0629491b-6b09-4d57-8893-e804be404b21 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': task-1278081, 'name': PowerOnVM_Task, 'duration_secs': 0.712994} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1075.045064] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-0629491b-6b09-4d57-8893-e804be404b21 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: b82af1e9-294a-4ba9-bcad-73b2a2aca86d] Powered on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1075.045282] env[61957]: INFO nova.compute.manager [None req-0629491b-6b09-4d57-8893-e804be404b21 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: b82af1e9-294a-4ba9-bcad-73b2a2aca86d] Took 8.48 seconds to spawn the instance on the hypervisor. [ 1075.045470] env[61957]: DEBUG nova.compute.manager [None req-0629491b-6b09-4d57-8893-e804be404b21 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: b82af1e9-294a-4ba9-bcad-73b2a2aca86d] Checking state {{(pid=61957) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1075.046277] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-393a7192-c6c5-4733-bfc1-44ac68e31b18 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.049663] env[61957]: INFO nova.scheduler.client.report [None req-ed26b16b-8a00-4d91-8fb4-6b369fa27bc4 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Deleted allocations for instance 11d67afa-1d96-4a9b-8439-383d586d8e17 [ 1075.367178] env[61957]: DEBUG oslo_vmware.api [None req-92fb2eb5-36b1-4a05-8eec-33f1fe3376d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1278082, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071579} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1075.367527] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-92fb2eb5-36b1-4a05-8eec-33f1fe3376d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 84cd5619-4ac9-41ef-9368-a023ad9ae66d] Extended root virtual disk {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1075.368330] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0df039df-d4ee-4351-aa5c-ddb950864a94 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.391371] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-92fb2eb5-36b1-4a05-8eec-33f1fe3376d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 84cd5619-4ac9-41ef-9368-a023ad9ae66d] Reconfiguring VM instance instance-00000068 to attach disk [datastore2] 84cd5619-4ac9-41ef-9368-a023ad9ae66d/84cd5619-4ac9-41ef-9368-a023ad9ae66d.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1075.391660] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2ee2971c-0de0-444a-a9f5-8229fb5d1376 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.410608] env[61957]: DEBUG oslo_vmware.api [None req-92fb2eb5-36b1-4a05-8eec-33f1fe3376d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Waiting for the task: (returnval){ [ 1075.410608] env[61957]: value = "task-1278083" [ 1075.410608] env[61957]: _type = "Task" [ 1075.410608] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1075.418324] env[61957]: DEBUG oslo_vmware.api [None req-92fb2eb5-36b1-4a05-8eec-33f1fe3376d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1278083, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1075.564552] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ed26b16b-8a00-4d91-8fb4-6b369fa27bc4 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Lock "11d67afa-1d96-4a9b-8439-383d586d8e17" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.918s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1075.567082] env[61957]: INFO nova.compute.manager [None req-0629491b-6b09-4d57-8893-e804be404b21 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: b82af1e9-294a-4ba9-bcad-73b2a2aca86d] Took 15.04 seconds to build instance. [ 1075.920912] env[61957]: DEBUG oslo_vmware.api [None req-92fb2eb5-36b1-4a05-8eec-33f1fe3376d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1278083, 'name': ReconfigVM_Task, 'duration_secs': 0.268597} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1075.921220] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-92fb2eb5-36b1-4a05-8eec-33f1fe3376d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 84cd5619-4ac9-41ef-9368-a023ad9ae66d] Reconfigured VM instance instance-00000068 to attach disk [datastore2] 84cd5619-4ac9-41ef-9368-a023ad9ae66d/84cd5619-4ac9-41ef-9368-a023ad9ae66d.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1075.921872] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-38ed9eec-d824-4b97-b71e-e7fec00e0d8a {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.927866] env[61957]: DEBUG oslo_vmware.api [None req-92fb2eb5-36b1-4a05-8eec-33f1fe3376d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Waiting for the task: (returnval){ [ 1075.927866] env[61957]: value = "task-1278084" [ 1075.927866] env[61957]: _type = "Task" [ 1075.927866] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1075.935476] env[61957]: DEBUG oslo_vmware.api [None req-92fb2eb5-36b1-4a05-8eec-33f1fe3376d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1278084, 'name': Rename_Task} progress is 5%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1076.068797] env[61957]: DEBUG oslo_concurrency.lockutils [None req-0629491b-6b09-4d57-8893-e804be404b21 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Lock "b82af1e9-294a-4ba9-bcad-73b2a2aca86d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.554s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1076.328753] env[61957]: DEBUG nova.compute.manager [req-951d795e-cdc2-41c2-af7f-4ddeba67ba59 req-8e6e5eb0-b72c-4ce8-9082-bbcb40b2b6a2 service nova] [instance: 4625d1dc-a621-4a33-9537-9e8cce96c914] Received event network-changed-8365f86e-8b14-47de-a6f4-297107be492e {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1076.328959] env[61957]: DEBUG nova.compute.manager [req-951d795e-cdc2-41c2-af7f-4ddeba67ba59 req-8e6e5eb0-b72c-4ce8-9082-bbcb40b2b6a2 service nova] [instance: 4625d1dc-a621-4a33-9537-9e8cce96c914] Refreshing instance network info cache due to event network-changed-8365f86e-8b14-47de-a6f4-297107be492e. {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1076.329206] env[61957]: DEBUG oslo_concurrency.lockutils [req-951d795e-cdc2-41c2-af7f-4ddeba67ba59 req-8e6e5eb0-b72c-4ce8-9082-bbcb40b2b6a2 service nova] Acquiring lock "refresh_cache-4625d1dc-a621-4a33-9537-9e8cce96c914" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1076.329925] env[61957]: DEBUG oslo_concurrency.lockutils [req-951d795e-cdc2-41c2-af7f-4ddeba67ba59 req-8e6e5eb0-b72c-4ce8-9082-bbcb40b2b6a2 service nova] Acquired lock "refresh_cache-4625d1dc-a621-4a33-9537-9e8cce96c914" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1076.330158] env[61957]: DEBUG nova.network.neutron [req-951d795e-cdc2-41c2-af7f-4ddeba67ba59 req-8e6e5eb0-b72c-4ce8-9082-bbcb40b2b6a2 service nova] [instance: 4625d1dc-a621-4a33-9537-9e8cce96c914] Refreshing network info cache for port 8365f86e-8b14-47de-a6f4-297107be492e {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1076.440417] env[61957]: DEBUG oslo_vmware.api [None req-92fb2eb5-36b1-4a05-8eec-33f1fe3376d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1278084, 'name': Rename_Task, 'duration_secs': 0.137167} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1076.440417] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-92fb2eb5-36b1-4a05-8eec-33f1fe3376d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 84cd5619-4ac9-41ef-9368-a023ad9ae66d] Powering on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1076.440417] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c86553b4-3c42-47a3-923f-2769c7560255 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.447024] env[61957]: DEBUG oslo_vmware.api [None req-92fb2eb5-36b1-4a05-8eec-33f1fe3376d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Waiting for the task: (returnval){ [ 1076.447024] env[61957]: value = "task-1278085" [ 1076.447024] env[61957]: _type = "Task" [ 1076.447024] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1076.454543] env[61957]: DEBUG oslo_vmware.api [None req-92fb2eb5-36b1-4a05-8eec-33f1fe3376d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1278085, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1076.725194] env[61957]: DEBUG oslo_concurrency.lockutils [None req-81f6269b-a21a-42b6-aa5f-6e2a26a0e628 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Acquiring lock "36769881-ff61-4bf3-b7e4-e7af19275805" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1076.725469] env[61957]: DEBUG oslo_concurrency.lockutils [None req-81f6269b-a21a-42b6-aa5f-6e2a26a0e628 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Lock "36769881-ff61-4bf3-b7e4-e7af19275805" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1076.958293] env[61957]: DEBUG oslo_vmware.api [None req-92fb2eb5-36b1-4a05-8eec-33f1fe3376d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1278085, 'name': PowerOnVM_Task, 'duration_secs': 0.485585} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1076.958894] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-92fb2eb5-36b1-4a05-8eec-33f1fe3376d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 84cd5619-4ac9-41ef-9368-a023ad9ae66d] Powered on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1076.959131] env[61957]: INFO nova.compute.manager [None req-92fb2eb5-36b1-4a05-8eec-33f1fe3376d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 84cd5619-4ac9-41ef-9368-a023ad9ae66d] Took 6.51 seconds to spawn the instance on the hypervisor. [ 1076.959318] env[61957]: DEBUG nova.compute.manager [None req-92fb2eb5-36b1-4a05-8eec-33f1fe3376d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 84cd5619-4ac9-41ef-9368-a023ad9ae66d] Checking state {{(pid=61957) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1076.960080] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6244b0b7-e1e4-4b34-9b0e-2034dcdbb432 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.177114] env[61957]: DEBUG nova.network.neutron [req-951d795e-cdc2-41c2-af7f-4ddeba67ba59 req-8e6e5eb0-b72c-4ce8-9082-bbcb40b2b6a2 service nova] [instance: 4625d1dc-a621-4a33-9537-9e8cce96c914] Updated VIF entry in instance network info cache for port 8365f86e-8b14-47de-a6f4-297107be492e. {{(pid=61957) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1077.177500] env[61957]: DEBUG nova.network.neutron [req-951d795e-cdc2-41c2-af7f-4ddeba67ba59 req-8e6e5eb0-b72c-4ce8-9082-bbcb40b2b6a2 service nova] [instance: 4625d1dc-a621-4a33-9537-9e8cce96c914] Updating instance_info_cache with network_info: [{"id": "8365f86e-8b14-47de-a6f4-297107be492e", "address": "fa:16:3e:e8:59:9d", "network": {"id": "364cabcb-046d-4f91-b1ee-5dd3adfae6d7", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1432396967-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8713aa35bcb24b86ad0b58ca9fc991ba", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b7a73c01-1bb9-4612-a1a7-16d71b732e81", "external-id": "nsx-vlan-transportzone-711", "segmentation_id": 711, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8365f86e-8b", "ovs_interfaceid": "8365f86e-8b14-47de-a6f4-297107be492e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1077.228403] env[61957]: DEBUG nova.compute.manager [None req-81f6269b-a21a-42b6-aa5f-6e2a26a0e628 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 36769881-ff61-4bf3-b7e4-e7af19275805] Starting instance... {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1077.477946] env[61957]: INFO nova.compute.manager [None req-92fb2eb5-36b1-4a05-8eec-33f1fe3376d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 84cd5619-4ac9-41ef-9368-a023ad9ae66d] Took 14.44 seconds to build instance. [ 1077.680064] env[61957]: DEBUG oslo_concurrency.lockutils [req-951d795e-cdc2-41c2-af7f-4ddeba67ba59 req-8e6e5eb0-b72c-4ce8-9082-bbcb40b2b6a2 service nova] Releasing lock "refresh_cache-4625d1dc-a621-4a33-9537-9e8cce96c914" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1077.750364] env[61957]: DEBUG oslo_concurrency.lockutils [None req-81f6269b-a21a-42b6-aa5f-6e2a26a0e628 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1077.750678] env[61957]: DEBUG oslo_concurrency.lockutils [None req-81f6269b-a21a-42b6-aa5f-6e2a26a0e628 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1077.752154] env[61957]: INFO nova.compute.claims [None req-81f6269b-a21a-42b6-aa5f-6e2a26a0e628 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 36769881-ff61-4bf3-b7e4-e7af19275805] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1077.980613] env[61957]: DEBUG oslo_concurrency.lockutils [None req-92fb2eb5-36b1-4a05-8eec-33f1fe3376d9 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Lock "84cd5619-4ac9-41ef-9368-a023ad9ae66d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.953s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1078.356748] env[61957]: DEBUG nova.compute.manager [req-9c589348-895f-4619-ae69-caf6dcb51a05 req-97b698ed-c417-4edb-bf6a-f4efdcb327d3 service nova] [instance: b82af1e9-294a-4ba9-bcad-73b2a2aca86d] Received event network-changed-58a3f865-3fc6-462a-b685-fb32e01f6c5e {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1078.356963] env[61957]: DEBUG nova.compute.manager [req-9c589348-895f-4619-ae69-caf6dcb51a05 req-97b698ed-c417-4edb-bf6a-f4efdcb327d3 service nova] [instance: b82af1e9-294a-4ba9-bcad-73b2a2aca86d] Refreshing instance network info cache due to event network-changed-58a3f865-3fc6-462a-b685-fb32e01f6c5e. {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1078.357252] env[61957]: DEBUG oslo_concurrency.lockutils [req-9c589348-895f-4619-ae69-caf6dcb51a05 req-97b698ed-c417-4edb-bf6a-f4efdcb327d3 service nova] Acquiring lock "refresh_cache-b82af1e9-294a-4ba9-bcad-73b2a2aca86d" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1078.357477] env[61957]: DEBUG oslo_concurrency.lockutils [req-9c589348-895f-4619-ae69-caf6dcb51a05 req-97b698ed-c417-4edb-bf6a-f4efdcb327d3 service nova] Acquired lock "refresh_cache-b82af1e9-294a-4ba9-bcad-73b2a2aca86d" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1078.357510] env[61957]: DEBUG nova.network.neutron [req-9c589348-895f-4619-ae69-caf6dcb51a05 req-97b698ed-c417-4edb-bf6a-f4efdcb327d3 service nova] [instance: b82af1e9-294a-4ba9-bcad-73b2a2aca86d] Refreshing network info cache for port 58a3f865-3fc6-462a-b685-fb32e01f6c5e {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1078.843162] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-186b8ccb-25ad-4253-86bc-21173eb537dd {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.851797] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71a463c1-cad7-42de-bc69-4806343c9fbd {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.885357] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76db9022-8fde-46d9-a853-4af13f1c6675 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.892506] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3396d91-1b9b-47b3-ab4e-45165d617e9a {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.906097] env[61957]: DEBUG nova.compute.provider_tree [None req-81f6269b-a21a-42b6-aa5f-6e2a26a0e628 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1079.072183] env[61957]: DEBUG nova.network.neutron [req-9c589348-895f-4619-ae69-caf6dcb51a05 req-97b698ed-c417-4edb-bf6a-f4efdcb327d3 service nova] [instance: b82af1e9-294a-4ba9-bcad-73b2a2aca86d] Updated VIF entry in instance network info cache for port 58a3f865-3fc6-462a-b685-fb32e01f6c5e. {{(pid=61957) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1079.072539] env[61957]: DEBUG nova.network.neutron [req-9c589348-895f-4619-ae69-caf6dcb51a05 req-97b698ed-c417-4edb-bf6a-f4efdcb327d3 service nova] [instance: b82af1e9-294a-4ba9-bcad-73b2a2aca86d] Updating instance_info_cache with network_info: [{"id": "58a3f865-3fc6-462a-b685-fb32e01f6c5e", "address": "fa:16:3e:ae:67:c0", "network": {"id": "364cabcb-046d-4f91-b1ee-5dd3adfae6d7", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1432396967-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8713aa35bcb24b86ad0b58ca9fc991ba", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b7a73c01-1bb9-4612-a1a7-16d71b732e81", "external-id": "nsx-vlan-transportzone-711", "segmentation_id": 711, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap58a3f865-3f", "ovs_interfaceid": "58a3f865-3fc6-462a-b685-fb32e01f6c5e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1079.346854] env[61957]: DEBUG nova.compute.manager [None req-c68409b5-1674-400e-bc43-8bb6913b0678 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 84cd5619-4ac9-41ef-9368-a023ad9ae66d] Stashing vm_state: active {{(pid=61957) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1079.411269] env[61957]: DEBUG nova.scheduler.client.report [None req-81f6269b-a21a-42b6-aa5f-6e2a26a0e628 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1079.575448] env[61957]: DEBUG oslo_concurrency.lockutils [req-9c589348-895f-4619-ae69-caf6dcb51a05 req-97b698ed-c417-4edb-bf6a-f4efdcb327d3 service nova] Releasing lock "refresh_cache-b82af1e9-294a-4ba9-bcad-73b2a2aca86d" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1079.575848] env[61957]: DEBUG nova.compute.manager [req-9c589348-895f-4619-ae69-caf6dcb51a05 req-97b698ed-c417-4edb-bf6a-f4efdcb327d3 service nova] [instance: b82af1e9-294a-4ba9-bcad-73b2a2aca86d] Received event network-changed-58a3f865-3fc6-462a-b685-fb32e01f6c5e {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1079.576172] env[61957]: DEBUG nova.compute.manager [req-9c589348-895f-4619-ae69-caf6dcb51a05 req-97b698ed-c417-4edb-bf6a-f4efdcb327d3 service nova] [instance: b82af1e9-294a-4ba9-bcad-73b2a2aca86d] Refreshing instance network info cache due to event network-changed-58a3f865-3fc6-462a-b685-fb32e01f6c5e. {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1079.576512] env[61957]: DEBUG oslo_concurrency.lockutils [req-9c589348-895f-4619-ae69-caf6dcb51a05 req-97b698ed-c417-4edb-bf6a-f4efdcb327d3 service nova] Acquiring lock "refresh_cache-b82af1e9-294a-4ba9-bcad-73b2a2aca86d" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1079.576771] env[61957]: DEBUG oslo_concurrency.lockutils [req-9c589348-895f-4619-ae69-caf6dcb51a05 req-97b698ed-c417-4edb-bf6a-f4efdcb327d3 service nova] Acquired lock "refresh_cache-b82af1e9-294a-4ba9-bcad-73b2a2aca86d" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1079.577062] env[61957]: DEBUG nova.network.neutron [req-9c589348-895f-4619-ae69-caf6dcb51a05 req-97b698ed-c417-4edb-bf6a-f4efdcb327d3 service nova] [instance: b82af1e9-294a-4ba9-bcad-73b2a2aca86d] Refreshing network info cache for port 58a3f865-3fc6-462a-b685-fb32e01f6c5e {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1079.870802] env[61957]: DEBUG oslo_concurrency.lockutils [None req-c68409b5-1674-400e-bc43-8bb6913b0678 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1079.917068] env[61957]: DEBUG oslo_concurrency.lockutils [None req-81f6269b-a21a-42b6-aa5f-6e2a26a0e628 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.166s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1079.917606] env[61957]: DEBUG nova.compute.manager [None req-81f6269b-a21a-42b6-aa5f-6e2a26a0e628 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 36769881-ff61-4bf3-b7e4-e7af19275805] Start building networks asynchronously for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1079.920436] env[61957]: DEBUG oslo_concurrency.lockutils [None req-c68409b5-1674-400e-bc43-8bb6913b0678 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.050s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1080.296817] env[61957]: DEBUG nova.network.neutron [req-9c589348-895f-4619-ae69-caf6dcb51a05 req-97b698ed-c417-4edb-bf6a-f4efdcb327d3 service nova] [instance: b82af1e9-294a-4ba9-bcad-73b2a2aca86d] Updated VIF entry in instance network info cache for port 58a3f865-3fc6-462a-b685-fb32e01f6c5e. {{(pid=61957) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1080.297323] env[61957]: DEBUG nova.network.neutron [req-9c589348-895f-4619-ae69-caf6dcb51a05 req-97b698ed-c417-4edb-bf6a-f4efdcb327d3 service nova] [instance: b82af1e9-294a-4ba9-bcad-73b2a2aca86d] Updating instance_info_cache with network_info: [{"id": "58a3f865-3fc6-462a-b685-fb32e01f6c5e", "address": "fa:16:3e:ae:67:c0", "network": {"id": "364cabcb-046d-4f91-b1ee-5dd3adfae6d7", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1432396967-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8713aa35bcb24b86ad0b58ca9fc991ba", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b7a73c01-1bb9-4612-a1a7-16d71b732e81", "external-id": "nsx-vlan-transportzone-711", "segmentation_id": 711, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap58a3f865-3f", "ovs_interfaceid": "58a3f865-3fc6-462a-b685-fb32e01f6c5e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1080.425146] env[61957]: DEBUG nova.compute.utils [None req-81f6269b-a21a-42b6-aa5f-6e2a26a0e628 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Using /dev/sd instead of None {{(pid=61957) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1080.428618] env[61957]: INFO nova.compute.claims [None req-c68409b5-1674-400e-bc43-8bb6913b0678 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 84cd5619-4ac9-41ef-9368-a023ad9ae66d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1080.433869] env[61957]: DEBUG nova.compute.manager [None req-81f6269b-a21a-42b6-aa5f-6e2a26a0e628 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 36769881-ff61-4bf3-b7e4-e7af19275805] Allocating IP information in the background. {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1080.433869] env[61957]: DEBUG nova.network.neutron [None req-81f6269b-a21a-42b6-aa5f-6e2a26a0e628 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 36769881-ff61-4bf3-b7e4-e7af19275805] allocate_for_instance() {{(pid=61957) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1080.476730] env[61957]: DEBUG nova.policy [None req-81f6269b-a21a-42b6-aa5f-6e2a26a0e628 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e2f13689624b483d903ce12ef290db8b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '65fd71bcd1bf41238a9cc3a5d6dd4924', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61957) authorize /opt/stack/nova/nova/policy.py:203}} [ 1080.769920] env[61957]: DEBUG nova.network.neutron [None req-81f6269b-a21a-42b6-aa5f-6e2a26a0e628 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 36769881-ff61-4bf3-b7e4-e7af19275805] Successfully created port: 09f655dd-f483-42c4-b37b-c0985ad5de04 {{(pid=61957) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1080.800101] env[61957]: DEBUG oslo_concurrency.lockutils [req-9c589348-895f-4619-ae69-caf6dcb51a05 req-97b698ed-c417-4edb-bf6a-f4efdcb327d3 service nova] Releasing lock "refresh_cache-b82af1e9-294a-4ba9-bcad-73b2a2aca86d" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1080.800403] env[61957]: DEBUG nova.compute.manager [req-9c589348-895f-4619-ae69-caf6dcb51a05 req-97b698ed-c417-4edb-bf6a-f4efdcb327d3 service nova] [instance: 4625d1dc-a621-4a33-9537-9e8cce96c914] Received event network-changed-8365f86e-8b14-47de-a6f4-297107be492e {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1080.800600] env[61957]: DEBUG nova.compute.manager [req-9c589348-895f-4619-ae69-caf6dcb51a05 req-97b698ed-c417-4edb-bf6a-f4efdcb327d3 service nova] [instance: 4625d1dc-a621-4a33-9537-9e8cce96c914] Refreshing instance network info cache due to event network-changed-8365f86e-8b14-47de-a6f4-297107be492e. {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1080.800862] env[61957]: DEBUG oslo_concurrency.lockutils [req-9c589348-895f-4619-ae69-caf6dcb51a05 req-97b698ed-c417-4edb-bf6a-f4efdcb327d3 service nova] Acquiring lock "refresh_cache-4625d1dc-a621-4a33-9537-9e8cce96c914" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1080.801033] env[61957]: DEBUG oslo_concurrency.lockutils [req-9c589348-895f-4619-ae69-caf6dcb51a05 req-97b698ed-c417-4edb-bf6a-f4efdcb327d3 service nova] Acquired lock "refresh_cache-4625d1dc-a621-4a33-9537-9e8cce96c914" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1080.801209] env[61957]: DEBUG nova.network.neutron [req-9c589348-895f-4619-ae69-caf6dcb51a05 req-97b698ed-c417-4edb-bf6a-f4efdcb327d3 service nova] [instance: 4625d1dc-a621-4a33-9537-9e8cce96c914] Refreshing network info cache for port 8365f86e-8b14-47de-a6f4-297107be492e {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1080.933979] env[61957]: DEBUG nova.compute.manager [None req-81f6269b-a21a-42b6-aa5f-6e2a26a0e628 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 36769881-ff61-4bf3-b7e4-e7af19275805] Start building block device mappings for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1080.939381] env[61957]: INFO nova.compute.resource_tracker [None req-c68409b5-1674-400e-bc43-8bb6913b0678 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 84cd5619-4ac9-41ef-9368-a023ad9ae66d] Updating resource usage from migration 26522749-2bb5-49e0-acaa-5e0766e6c136 [ 1081.043159] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cde7a584-41af-4e7c-8ab7-5ca5c6793f5b {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.050713] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c29ea17-b79b-4c94-bd63-7c10151e4c65 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.079383] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53bcd230-60f2-49b6-8bfa-3fd2f3619d93 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.086745] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f04c934-0cb3-46fe-83b6-a6a994284340 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.100529] env[61957]: DEBUG nova.compute.provider_tree [None req-c68409b5-1674-400e-bc43-8bb6913b0678 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1081.516470] env[61957]: DEBUG nova.network.neutron [req-9c589348-895f-4619-ae69-caf6dcb51a05 req-97b698ed-c417-4edb-bf6a-f4efdcb327d3 service nova] [instance: 4625d1dc-a621-4a33-9537-9e8cce96c914] Updated VIF entry in instance network info cache for port 8365f86e-8b14-47de-a6f4-297107be492e. {{(pid=61957) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1081.516864] env[61957]: DEBUG nova.network.neutron [req-9c589348-895f-4619-ae69-caf6dcb51a05 req-97b698ed-c417-4edb-bf6a-f4efdcb327d3 service nova] [instance: 4625d1dc-a621-4a33-9537-9e8cce96c914] Updating instance_info_cache with network_info: [{"id": "8365f86e-8b14-47de-a6f4-297107be492e", "address": "fa:16:3e:e8:59:9d", "network": {"id": "364cabcb-046d-4f91-b1ee-5dd3adfae6d7", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1432396967-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.232", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8713aa35bcb24b86ad0b58ca9fc991ba", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b7a73c01-1bb9-4612-a1a7-16d71b732e81", "external-id": "nsx-vlan-transportzone-711", "segmentation_id": 711, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8365f86e-8b", "ovs_interfaceid": "8365f86e-8b14-47de-a6f4-297107be492e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1081.604055] env[61957]: DEBUG nova.scheduler.client.report [None req-c68409b5-1674-400e-bc43-8bb6913b0678 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1081.949916] env[61957]: DEBUG nova.compute.manager [None req-81f6269b-a21a-42b6-aa5f-6e2a26a0e628 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 36769881-ff61-4bf3-b7e4-e7af19275805] Start spawning the instance on the hypervisor. {{(pid=61957) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1081.975363] env[61957]: DEBUG nova.virt.hardware [None req-81f6269b-a21a-42b6-aa5f-6e2a26a0e628 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T17:22:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T17:21:50Z,direct_url=,disk_format='vmdk',id=11c76a2c-f705-470a-ba9d-4657858bab38,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='81d3096d0f094373913fc3dc8f5c3c6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T17:21:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1081.975697] env[61957]: DEBUG nova.virt.hardware [None req-81f6269b-a21a-42b6-aa5f-6e2a26a0e628 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Flavor limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1081.975892] env[61957]: DEBUG nova.virt.hardware [None req-81f6269b-a21a-42b6-aa5f-6e2a26a0e628 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Image limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1081.976106] env[61957]: DEBUG nova.virt.hardware [None req-81f6269b-a21a-42b6-aa5f-6e2a26a0e628 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Flavor pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1081.976264] env[61957]: DEBUG nova.virt.hardware [None req-81f6269b-a21a-42b6-aa5f-6e2a26a0e628 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Image pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1081.976419] env[61957]: DEBUG nova.virt.hardware [None req-81f6269b-a21a-42b6-aa5f-6e2a26a0e628 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1081.976642] env[61957]: DEBUG nova.virt.hardware [None req-81f6269b-a21a-42b6-aa5f-6e2a26a0e628 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1081.976809] env[61957]: DEBUG nova.virt.hardware [None req-81f6269b-a21a-42b6-aa5f-6e2a26a0e628 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1081.976978] env[61957]: DEBUG nova.virt.hardware [None req-81f6269b-a21a-42b6-aa5f-6e2a26a0e628 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Got 1 possible topologies {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1081.977164] env[61957]: DEBUG nova.virt.hardware [None req-81f6269b-a21a-42b6-aa5f-6e2a26a0e628 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1081.977343] env[61957]: DEBUG nova.virt.hardware [None req-81f6269b-a21a-42b6-aa5f-6e2a26a0e628 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1081.978267] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5eca1c0-8b15-468e-840d-f15a92fc76cc {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.986353] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58429fd8-a2d1-47cd-84be-5773ad0ba8e2 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.020465] env[61957]: DEBUG oslo_concurrency.lockutils [req-9c589348-895f-4619-ae69-caf6dcb51a05 req-97b698ed-c417-4edb-bf6a-f4efdcb327d3 service nova] Releasing lock "refresh_cache-4625d1dc-a621-4a33-9537-9e8cce96c914" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1082.108875] env[61957]: DEBUG oslo_concurrency.lockutils [None req-c68409b5-1674-400e-bc43-8bb6913b0678 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.188s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1082.108875] env[61957]: INFO nova.compute.manager [None req-c68409b5-1674-400e-bc43-8bb6913b0678 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 84cd5619-4ac9-41ef-9368-a023ad9ae66d] Migrating [ 1082.152124] env[61957]: DEBUG nova.compute.manager [req-0082711d-ee5a-4fa0-8002-8bb7fa2ca5a8 req-cad2f6bc-e985-427f-8ac6-f303ecbfd364 service nova] [instance: 36769881-ff61-4bf3-b7e4-e7af19275805] Received event network-vif-plugged-09f655dd-f483-42c4-b37b-c0985ad5de04 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1082.152364] env[61957]: DEBUG oslo_concurrency.lockutils [req-0082711d-ee5a-4fa0-8002-8bb7fa2ca5a8 req-cad2f6bc-e985-427f-8ac6-f303ecbfd364 service nova] Acquiring lock "36769881-ff61-4bf3-b7e4-e7af19275805-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1082.152578] env[61957]: DEBUG oslo_concurrency.lockutils [req-0082711d-ee5a-4fa0-8002-8bb7fa2ca5a8 req-cad2f6bc-e985-427f-8ac6-f303ecbfd364 service nova] Lock "36769881-ff61-4bf3-b7e4-e7af19275805-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1082.152767] env[61957]: DEBUG oslo_concurrency.lockutils [req-0082711d-ee5a-4fa0-8002-8bb7fa2ca5a8 req-cad2f6bc-e985-427f-8ac6-f303ecbfd364 service nova] Lock "36769881-ff61-4bf3-b7e4-e7af19275805-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1082.152942] env[61957]: DEBUG nova.compute.manager [req-0082711d-ee5a-4fa0-8002-8bb7fa2ca5a8 req-cad2f6bc-e985-427f-8ac6-f303ecbfd364 service nova] [instance: 36769881-ff61-4bf3-b7e4-e7af19275805] No waiting events found dispatching network-vif-plugged-09f655dd-f483-42c4-b37b-c0985ad5de04 {{(pid=61957) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1082.153128] env[61957]: WARNING nova.compute.manager [req-0082711d-ee5a-4fa0-8002-8bb7fa2ca5a8 req-cad2f6bc-e985-427f-8ac6-f303ecbfd364 service nova] [instance: 36769881-ff61-4bf3-b7e4-e7af19275805] Received unexpected event network-vif-plugged-09f655dd-f483-42c4-b37b-c0985ad5de04 for instance with vm_state building and task_state spawning. [ 1082.236058] env[61957]: DEBUG nova.network.neutron [None req-81f6269b-a21a-42b6-aa5f-6e2a26a0e628 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 36769881-ff61-4bf3-b7e4-e7af19275805] Successfully updated port: 09f655dd-f483-42c4-b37b-c0985ad5de04 {{(pid=61957) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1082.623903] env[61957]: DEBUG oslo_concurrency.lockutils [None req-c68409b5-1674-400e-bc43-8bb6913b0678 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Acquiring lock "refresh_cache-84cd5619-4ac9-41ef-9368-a023ad9ae66d" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1082.624218] env[61957]: DEBUG oslo_concurrency.lockutils [None req-c68409b5-1674-400e-bc43-8bb6913b0678 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Acquired lock "refresh_cache-84cd5619-4ac9-41ef-9368-a023ad9ae66d" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1082.624491] env[61957]: DEBUG nova.network.neutron [None req-c68409b5-1674-400e-bc43-8bb6913b0678 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 84cd5619-4ac9-41ef-9368-a023ad9ae66d] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1082.738878] env[61957]: DEBUG oslo_concurrency.lockutils [None req-81f6269b-a21a-42b6-aa5f-6e2a26a0e628 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Acquiring lock "refresh_cache-36769881-ff61-4bf3-b7e4-e7af19275805" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1082.739137] env[61957]: DEBUG oslo_concurrency.lockutils [None req-81f6269b-a21a-42b6-aa5f-6e2a26a0e628 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Acquired lock "refresh_cache-36769881-ff61-4bf3-b7e4-e7af19275805" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1082.739301] env[61957]: DEBUG nova.network.neutron [None req-81f6269b-a21a-42b6-aa5f-6e2a26a0e628 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 36769881-ff61-4bf3-b7e4-e7af19275805] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1083.280134] env[61957]: DEBUG nova.network.neutron [None req-81f6269b-a21a-42b6-aa5f-6e2a26a0e628 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 36769881-ff61-4bf3-b7e4-e7af19275805] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1083.381650] env[61957]: DEBUG nova.network.neutron [None req-c68409b5-1674-400e-bc43-8bb6913b0678 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 84cd5619-4ac9-41ef-9368-a023ad9ae66d] Updating instance_info_cache with network_info: [{"id": "b0541069-72e9-4384-b4f4-ccf5bc206dab", "address": "fa:16:3e:d1:ff:5a", "network": {"id": "efe08823-51fe-4620-a849-5345e262cc71", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1035436819-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e16f6dce3f0e44fb96de516f17d4c6f5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "667a2e97-c1be-421d-9941-6b84c2629b43", "external-id": "nsx-vlan-transportzone-484", "segmentation_id": 484, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb0541069-72", "ovs_interfaceid": "b0541069-72e9-4384-b4f4-ccf5bc206dab", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1083.438733] env[61957]: DEBUG nova.network.neutron [None req-81f6269b-a21a-42b6-aa5f-6e2a26a0e628 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 36769881-ff61-4bf3-b7e4-e7af19275805] Updating instance_info_cache with network_info: [{"id": "09f655dd-f483-42c4-b37b-c0985ad5de04", "address": "fa:16:3e:77:2a:48", "network": {"id": "4f0e57be-8906-48b5-865f-59ff12250a8b", "bridge": "br-int", "label": "tempest-ServersTestJSON-1916182687-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "65fd71bcd1bf41238a9cc3a5d6dd4924", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "939c05b6-8f31-4f3a-95ac-6297e0bd243e", "external-id": "nsx-vlan-transportzone-825", "segmentation_id": 825, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap09f655dd-f4", "ovs_interfaceid": "09f655dd-f483-42c4-b37b-c0985ad5de04", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1083.884342] env[61957]: DEBUG oslo_concurrency.lockutils [None req-c68409b5-1674-400e-bc43-8bb6913b0678 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Releasing lock "refresh_cache-84cd5619-4ac9-41ef-9368-a023ad9ae66d" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1083.941902] env[61957]: DEBUG oslo_concurrency.lockutils [None req-81f6269b-a21a-42b6-aa5f-6e2a26a0e628 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Releasing lock "refresh_cache-36769881-ff61-4bf3-b7e4-e7af19275805" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1083.942227] env[61957]: DEBUG nova.compute.manager [None req-81f6269b-a21a-42b6-aa5f-6e2a26a0e628 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 36769881-ff61-4bf3-b7e4-e7af19275805] Instance network_info: |[{"id": "09f655dd-f483-42c4-b37b-c0985ad5de04", "address": "fa:16:3e:77:2a:48", "network": {"id": "4f0e57be-8906-48b5-865f-59ff12250a8b", "bridge": "br-int", "label": "tempest-ServersTestJSON-1916182687-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "65fd71bcd1bf41238a9cc3a5d6dd4924", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "939c05b6-8f31-4f3a-95ac-6297e0bd243e", "external-id": "nsx-vlan-transportzone-825", "segmentation_id": 825, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap09f655dd-f4", "ovs_interfaceid": "09f655dd-f483-42c4-b37b-c0985ad5de04", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1083.942710] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-81f6269b-a21a-42b6-aa5f-6e2a26a0e628 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 36769881-ff61-4bf3-b7e4-e7af19275805] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:77:2a:48', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '939c05b6-8f31-4f3a-95ac-6297e0bd243e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '09f655dd-f483-42c4-b37b-c0985ad5de04', 'vif_model': 'vmxnet3'}] {{(pid=61957) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1083.950392] env[61957]: DEBUG oslo.service.loopingcall [None req-81f6269b-a21a-42b6-aa5f-6e2a26a0e628 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1083.950656] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 36769881-ff61-4bf3-b7e4-e7af19275805] Creating VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1083.950927] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ca6d5f55-3930-46eb-af44-82512c65b655 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.971606] env[61957]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1083.971606] env[61957]: value = "task-1278086" [ 1083.971606] env[61957]: _type = "Task" [ 1083.971606] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1083.983974] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1278086, 'name': CreateVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.178240] env[61957]: DEBUG nova.compute.manager [req-c4e2e715-9670-4bd2-b2c0-f87d520b4b4b req-a5845250-69de-4ca1-b89c-e3ceaad072e9 service nova] [instance: 36769881-ff61-4bf3-b7e4-e7af19275805] Received event network-changed-09f655dd-f483-42c4-b37b-c0985ad5de04 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1084.178312] env[61957]: DEBUG nova.compute.manager [req-c4e2e715-9670-4bd2-b2c0-f87d520b4b4b req-a5845250-69de-4ca1-b89c-e3ceaad072e9 service nova] [instance: 36769881-ff61-4bf3-b7e4-e7af19275805] Refreshing instance network info cache due to event network-changed-09f655dd-f483-42c4-b37b-c0985ad5de04. {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1084.178519] env[61957]: DEBUG oslo_concurrency.lockutils [req-c4e2e715-9670-4bd2-b2c0-f87d520b4b4b req-a5845250-69de-4ca1-b89c-e3ceaad072e9 service nova] Acquiring lock "refresh_cache-36769881-ff61-4bf3-b7e4-e7af19275805" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1084.178671] env[61957]: DEBUG oslo_concurrency.lockutils [req-c4e2e715-9670-4bd2-b2c0-f87d520b4b4b req-a5845250-69de-4ca1-b89c-e3ceaad072e9 service nova] Acquired lock "refresh_cache-36769881-ff61-4bf3-b7e4-e7af19275805" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1084.178910] env[61957]: DEBUG nova.network.neutron [req-c4e2e715-9670-4bd2-b2c0-f87d520b4b4b req-a5845250-69de-4ca1-b89c-e3ceaad072e9 service nova] [instance: 36769881-ff61-4bf3-b7e4-e7af19275805] Refreshing network info cache for port 09f655dd-f483-42c4-b37b-c0985ad5de04 {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1084.481905] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1278086, 'name': CreateVM_Task, 'duration_secs': 0.32823} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1084.482282] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 36769881-ff61-4bf3-b7e4-e7af19275805] Created VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1084.482859] env[61957]: DEBUG oslo_concurrency.lockutils [None req-81f6269b-a21a-42b6-aa5f-6e2a26a0e628 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1084.483063] env[61957]: DEBUG oslo_concurrency.lockutils [None req-81f6269b-a21a-42b6-aa5f-6e2a26a0e628 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Acquired lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1084.483374] env[61957]: DEBUG oslo_concurrency.lockutils [None req-81f6269b-a21a-42b6-aa5f-6e2a26a0e628 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1084.483642] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7e33cb00-d4f0-4556-8958-bc534004d4f2 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.488343] env[61957]: DEBUG oslo_vmware.api [None req-81f6269b-a21a-42b6-aa5f-6e2a26a0e628 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Waiting for the task: (returnval){ [ 1084.488343] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52187e78-83be-cbf8-5263-20788631133a" [ 1084.488343] env[61957]: _type = "Task" [ 1084.488343] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1084.495592] env[61957]: DEBUG oslo_vmware.api [None req-81f6269b-a21a-42b6-aa5f-6e2a26a0e628 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52187e78-83be-cbf8-5263-20788631133a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.869394] env[61957]: DEBUG nova.network.neutron [req-c4e2e715-9670-4bd2-b2c0-f87d520b4b4b req-a5845250-69de-4ca1-b89c-e3ceaad072e9 service nova] [instance: 36769881-ff61-4bf3-b7e4-e7af19275805] Updated VIF entry in instance network info cache for port 09f655dd-f483-42c4-b37b-c0985ad5de04. {{(pid=61957) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1084.869906] env[61957]: DEBUG nova.network.neutron [req-c4e2e715-9670-4bd2-b2c0-f87d520b4b4b req-a5845250-69de-4ca1-b89c-e3ceaad072e9 service nova] [instance: 36769881-ff61-4bf3-b7e4-e7af19275805] Updating instance_info_cache with network_info: [{"id": "09f655dd-f483-42c4-b37b-c0985ad5de04", "address": "fa:16:3e:77:2a:48", "network": {"id": "4f0e57be-8906-48b5-865f-59ff12250a8b", "bridge": "br-int", "label": "tempest-ServersTestJSON-1916182687-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "65fd71bcd1bf41238a9cc3a5d6dd4924", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "939c05b6-8f31-4f3a-95ac-6297e0bd243e", "external-id": "nsx-vlan-transportzone-825", "segmentation_id": 825, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap09f655dd-f4", "ovs_interfaceid": "09f655dd-f483-42c4-b37b-c0985ad5de04", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1085.000359] env[61957]: DEBUG oslo_vmware.api [None req-81f6269b-a21a-42b6-aa5f-6e2a26a0e628 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52187e78-83be-cbf8-5263-20788631133a, 'name': SearchDatastore_Task, 'duration_secs': 0.009476} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1085.000755] env[61957]: DEBUG oslo_concurrency.lockutils [None req-81f6269b-a21a-42b6-aa5f-6e2a26a0e628 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Releasing lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1085.001077] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-81f6269b-a21a-42b6-aa5f-6e2a26a0e628 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 36769881-ff61-4bf3-b7e4-e7af19275805] Processing image 11c76a2c-f705-470a-ba9d-4657858bab38 {{(pid=61957) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1085.001392] env[61957]: DEBUG oslo_concurrency.lockutils [None req-81f6269b-a21a-42b6-aa5f-6e2a26a0e628 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1085.001603] env[61957]: DEBUG oslo_concurrency.lockutils [None req-81f6269b-a21a-42b6-aa5f-6e2a26a0e628 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Acquired lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1085.001855] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-81f6269b-a21a-42b6-aa5f-6e2a26a0e628 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1085.002186] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ee596a55-e9ae-4a33-a58f-0f0ff06effcb {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.010856] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-81f6269b-a21a-42b6-aa5f-6e2a26a0e628 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1085.011122] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-81f6269b-a21a-42b6-aa5f-6e2a26a0e628 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61957) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1085.011900] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f0c7aafa-3318-4f54-81da-b2fa81cd15cc {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.016978] env[61957]: DEBUG oslo_vmware.api [None req-81f6269b-a21a-42b6-aa5f-6e2a26a0e628 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Waiting for the task: (returnval){ [ 1085.016978] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52ab4a46-d1c5-28fb-373e-8e719c9c892d" [ 1085.016978] env[61957]: _type = "Task" [ 1085.016978] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1085.024266] env[61957]: DEBUG oslo_vmware.api [None req-81f6269b-a21a-42b6-aa5f-6e2a26a0e628 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52ab4a46-d1c5-28fb-373e-8e719c9c892d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.373026] env[61957]: DEBUG oslo_concurrency.lockutils [req-c4e2e715-9670-4bd2-b2c0-f87d520b4b4b req-a5845250-69de-4ca1-b89c-e3ceaad072e9 service nova] Releasing lock "refresh_cache-36769881-ff61-4bf3-b7e4-e7af19275805" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1085.398058] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-323fcf20-dbe9-4c9f-b1e9-975e3931dcd8 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.416281] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-c68409b5-1674-400e-bc43-8bb6913b0678 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 84cd5619-4ac9-41ef-9368-a023ad9ae66d] Updating instance '84cd5619-4ac9-41ef-9368-a023ad9ae66d' progress to 0 {{(pid=61957) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1085.527498] env[61957]: DEBUG oslo_vmware.api [None req-81f6269b-a21a-42b6-aa5f-6e2a26a0e628 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52ab4a46-d1c5-28fb-373e-8e719c9c892d, 'name': SearchDatastore_Task, 'duration_secs': 0.008199} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1085.528276] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c800cb94-4943-40e6-8f71-9e38fa090665 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.533209] env[61957]: DEBUG oslo_vmware.api [None req-81f6269b-a21a-42b6-aa5f-6e2a26a0e628 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Waiting for the task: (returnval){ [ 1085.533209] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]525e6bbd-8654-c21a-621b-92a30c869c73" [ 1085.533209] env[61957]: _type = "Task" [ 1085.533209] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1085.540068] env[61957]: DEBUG oslo_vmware.api [None req-81f6269b-a21a-42b6-aa5f-6e2a26a0e628 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]525e6bbd-8654-c21a-621b-92a30c869c73, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.922783] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-c68409b5-1674-400e-bc43-8bb6913b0678 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 84cd5619-4ac9-41ef-9368-a023ad9ae66d] Powering off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1085.922898] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b309bf4b-3553-4687-9885-2567416f6e3e {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.930596] env[61957]: DEBUG oslo_vmware.api [None req-c68409b5-1674-400e-bc43-8bb6913b0678 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Waiting for the task: (returnval){ [ 1085.930596] env[61957]: value = "task-1278087" [ 1085.930596] env[61957]: _type = "Task" [ 1085.930596] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1085.938572] env[61957]: DEBUG oslo_vmware.api [None req-c68409b5-1674-400e-bc43-8bb6913b0678 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1278087, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.043363] env[61957]: DEBUG oslo_vmware.api [None req-81f6269b-a21a-42b6-aa5f-6e2a26a0e628 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]525e6bbd-8654-c21a-621b-92a30c869c73, 'name': SearchDatastore_Task, 'duration_secs': 0.009224} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1086.043659] env[61957]: DEBUG oslo_concurrency.lockutils [None req-81f6269b-a21a-42b6-aa5f-6e2a26a0e628 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Releasing lock "[datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1086.043916] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-81f6269b-a21a-42b6-aa5f-6e2a26a0e628 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore1] 36769881-ff61-4bf3-b7e4-e7af19275805/36769881-ff61-4bf3-b7e4-e7af19275805.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1086.044204] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f42cb86e-1112-49d7-a3a9-37b760474cc0 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.050970] env[61957]: DEBUG oslo_vmware.api [None req-81f6269b-a21a-42b6-aa5f-6e2a26a0e628 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Waiting for the task: (returnval){ [ 1086.050970] env[61957]: value = "task-1278088" [ 1086.050970] env[61957]: _type = "Task" [ 1086.050970] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1086.060287] env[61957]: DEBUG oslo_vmware.api [None req-81f6269b-a21a-42b6-aa5f-6e2a26a0e628 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': task-1278088, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.441229] env[61957]: DEBUG oslo_vmware.api [None req-c68409b5-1674-400e-bc43-8bb6913b0678 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1278087, 'name': PowerOffVM_Task, 'duration_secs': 0.183875} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1086.441477] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-c68409b5-1674-400e-bc43-8bb6913b0678 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 84cd5619-4ac9-41ef-9368-a023ad9ae66d] Powered off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1086.441649] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-c68409b5-1674-400e-bc43-8bb6913b0678 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 84cd5619-4ac9-41ef-9368-a023ad9ae66d] Updating instance '84cd5619-4ac9-41ef-9368-a023ad9ae66d' progress to 17 {{(pid=61957) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1086.560613] env[61957]: DEBUG oslo_vmware.api [None req-81f6269b-a21a-42b6-aa5f-6e2a26a0e628 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': task-1278088, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.417751} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1086.561179] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-81f6269b-a21a-42b6-aa5f-6e2a26a0e628 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore1] 36769881-ff61-4bf3-b7e4-e7af19275805/36769881-ff61-4bf3-b7e4-e7af19275805.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1086.561179] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-81f6269b-a21a-42b6-aa5f-6e2a26a0e628 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 36769881-ff61-4bf3-b7e4-e7af19275805] Extending root virtual disk to 1048576 {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1086.561400] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6707782b-1974-4e6e-b173-e18d212397c5 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.568808] env[61957]: DEBUG oslo_vmware.api [None req-81f6269b-a21a-42b6-aa5f-6e2a26a0e628 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Waiting for the task: (returnval){ [ 1086.568808] env[61957]: value = "task-1278089" [ 1086.568808] env[61957]: _type = "Task" [ 1086.568808] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1086.576925] env[61957]: DEBUG oslo_vmware.api [None req-81f6269b-a21a-42b6-aa5f-6e2a26a0e628 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': task-1278089, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.948140] env[61957]: DEBUG nova.virt.hardware [None req-c68409b5-1674-400e-bc43-8bb6913b0678 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T17:22:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=11c76a2c-f705-470a-ba9d-4657858bab38,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1086.948413] env[61957]: DEBUG nova.virt.hardware [None req-c68409b5-1674-400e-bc43-8bb6913b0678 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Flavor limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1086.948603] env[61957]: DEBUG nova.virt.hardware [None req-c68409b5-1674-400e-bc43-8bb6913b0678 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Image limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1086.948819] env[61957]: DEBUG nova.virt.hardware [None req-c68409b5-1674-400e-bc43-8bb6913b0678 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Flavor pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1086.948976] env[61957]: DEBUG nova.virt.hardware [None req-c68409b5-1674-400e-bc43-8bb6913b0678 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Image pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1086.949144] env[61957]: DEBUG nova.virt.hardware [None req-c68409b5-1674-400e-bc43-8bb6913b0678 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1086.949359] env[61957]: DEBUG nova.virt.hardware [None req-c68409b5-1674-400e-bc43-8bb6913b0678 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1086.949526] env[61957]: DEBUG nova.virt.hardware [None req-c68409b5-1674-400e-bc43-8bb6913b0678 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1086.949697] env[61957]: DEBUG nova.virt.hardware [None req-c68409b5-1674-400e-bc43-8bb6913b0678 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Got 1 possible topologies {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1086.949874] env[61957]: DEBUG nova.virt.hardware [None req-c68409b5-1674-400e-bc43-8bb6913b0678 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1086.950057] env[61957]: DEBUG nova.virt.hardware [None req-c68409b5-1674-400e-bc43-8bb6913b0678 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1086.955069] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9531a501-55cb-4f17-ba74-2b79fc54dbd7 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.971043] env[61957]: DEBUG oslo_vmware.api [None req-c68409b5-1674-400e-bc43-8bb6913b0678 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Waiting for the task: (returnval){ [ 1086.971043] env[61957]: value = "task-1278090" [ 1086.971043] env[61957]: _type = "Task" [ 1086.971043] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1086.978989] env[61957]: DEBUG oslo_vmware.api [None req-c68409b5-1674-400e-bc43-8bb6913b0678 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1278090, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.077902] env[61957]: DEBUG oslo_vmware.api [None req-81f6269b-a21a-42b6-aa5f-6e2a26a0e628 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': task-1278089, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.057389} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1087.078202] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-81f6269b-a21a-42b6-aa5f-6e2a26a0e628 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 36769881-ff61-4bf3-b7e4-e7af19275805] Extended root virtual disk {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1087.078938] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77fe3aca-faea-4839-ba84-2b4a8aa65904 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.101672] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-81f6269b-a21a-42b6-aa5f-6e2a26a0e628 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 36769881-ff61-4bf3-b7e4-e7af19275805] Reconfiguring VM instance instance-00000069 to attach disk [datastore1] 36769881-ff61-4bf3-b7e4-e7af19275805/36769881-ff61-4bf3-b7e4-e7af19275805.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1087.103089] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7c3dceaa-1393-414f-ae96-8b70d4b3177e {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.119543] env[61957]: DEBUG oslo_concurrency.lockutils [None req-091a8e64-7857-4c73-85c1-c32e51c082b4 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Acquiring lock "039a6b01-a745-478c-9ff7-c8c37afe2ce4" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1087.119763] env[61957]: DEBUG oslo_concurrency.lockutils [None req-091a8e64-7857-4c73-85c1-c32e51c082b4 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Lock "039a6b01-a745-478c-9ff7-c8c37afe2ce4" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1087.123499] env[61957]: DEBUG oslo_vmware.api [None req-81f6269b-a21a-42b6-aa5f-6e2a26a0e628 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Waiting for the task: (returnval){ [ 1087.123499] env[61957]: value = "task-1278091" [ 1087.123499] env[61957]: _type = "Task" [ 1087.123499] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1087.133487] env[61957]: DEBUG oslo_vmware.api [None req-81f6269b-a21a-42b6-aa5f-6e2a26a0e628 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': task-1278091, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.480884] env[61957]: DEBUG oslo_vmware.api [None req-c68409b5-1674-400e-bc43-8bb6913b0678 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1278090, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.623622] env[61957]: DEBUG nova.compute.utils [None req-091a8e64-7857-4c73-85c1-c32e51c082b4 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Using /dev/sd instead of None {{(pid=61957) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1087.635229] env[61957]: DEBUG oslo_vmware.api [None req-81f6269b-a21a-42b6-aa5f-6e2a26a0e628 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': task-1278091, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.983014] env[61957]: DEBUG oslo_vmware.api [None req-c68409b5-1674-400e-bc43-8bb6913b0678 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1278090, 'name': ReconfigVM_Task, 'duration_secs': 0.77859} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1087.983459] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-c68409b5-1674-400e-bc43-8bb6913b0678 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 84cd5619-4ac9-41ef-9368-a023ad9ae66d] Updating instance '84cd5619-4ac9-41ef-9368-a023ad9ae66d' progress to 33 {{(pid=61957) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1088.129799] env[61957]: DEBUG oslo_concurrency.lockutils [None req-091a8e64-7857-4c73-85c1-c32e51c082b4 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Lock "039a6b01-a745-478c-9ff7-c8c37afe2ce4" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.010s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1088.135657] env[61957]: DEBUG oslo_vmware.api [None req-81f6269b-a21a-42b6-aa5f-6e2a26a0e628 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': task-1278091, 'name': ReconfigVM_Task, 'duration_secs': 0.756556} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1088.136223] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-81f6269b-a21a-42b6-aa5f-6e2a26a0e628 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 36769881-ff61-4bf3-b7e4-e7af19275805] Reconfigured VM instance instance-00000069 to attach disk [datastore1] 36769881-ff61-4bf3-b7e4-e7af19275805/36769881-ff61-4bf3-b7e4-e7af19275805.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1088.136540] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-13992d6c-8468-4aaa-8734-cd154d8a4a02 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.142743] env[61957]: DEBUG oslo_vmware.api [None req-81f6269b-a21a-42b6-aa5f-6e2a26a0e628 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Waiting for the task: (returnval){ [ 1088.142743] env[61957]: value = "task-1278092" [ 1088.142743] env[61957]: _type = "Task" [ 1088.142743] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1088.150692] env[61957]: DEBUG oslo_vmware.api [None req-81f6269b-a21a-42b6-aa5f-6e2a26a0e628 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': task-1278092, 'name': Rename_Task} progress is 5%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.490609] env[61957]: DEBUG nova.virt.hardware [None req-c68409b5-1674-400e-bc43-8bb6913b0678 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T17:22:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=11c76a2c-f705-470a-ba9d-4657858bab38,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1088.490912] env[61957]: DEBUG nova.virt.hardware [None req-c68409b5-1674-400e-bc43-8bb6913b0678 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Flavor limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1088.491093] env[61957]: DEBUG nova.virt.hardware [None req-c68409b5-1674-400e-bc43-8bb6913b0678 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Image limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1088.491289] env[61957]: DEBUG nova.virt.hardware [None req-c68409b5-1674-400e-bc43-8bb6913b0678 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Flavor pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1088.491443] env[61957]: DEBUG nova.virt.hardware [None req-c68409b5-1674-400e-bc43-8bb6913b0678 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Image pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1088.491598] env[61957]: DEBUG nova.virt.hardware [None req-c68409b5-1674-400e-bc43-8bb6913b0678 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1088.491809] env[61957]: DEBUG nova.virt.hardware [None req-c68409b5-1674-400e-bc43-8bb6913b0678 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1088.491977] env[61957]: DEBUG nova.virt.hardware [None req-c68409b5-1674-400e-bc43-8bb6913b0678 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1088.492170] env[61957]: DEBUG nova.virt.hardware [None req-c68409b5-1674-400e-bc43-8bb6913b0678 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Got 1 possible topologies {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1088.492343] env[61957]: DEBUG nova.virt.hardware [None req-c68409b5-1674-400e-bc43-8bb6913b0678 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1088.492521] env[61957]: DEBUG nova.virt.hardware [None req-c68409b5-1674-400e-bc43-8bb6913b0678 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1088.497872] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-c68409b5-1674-400e-bc43-8bb6913b0678 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 84cd5619-4ac9-41ef-9368-a023ad9ae66d] Reconfiguring VM instance instance-00000068 to detach disk 2000 {{(pid=61957) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1088.498180] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-23cfc83b-ded5-44bd-96dc-8f6086b2fa59 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.516504] env[61957]: DEBUG oslo_vmware.api [None req-c68409b5-1674-400e-bc43-8bb6913b0678 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Waiting for the task: (returnval){ [ 1088.516504] env[61957]: value = "task-1278093" [ 1088.516504] env[61957]: _type = "Task" [ 1088.516504] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1088.524795] env[61957]: DEBUG oslo_vmware.api [None req-c68409b5-1674-400e-bc43-8bb6913b0678 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1278093, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.652975] env[61957]: DEBUG oslo_vmware.api [None req-81f6269b-a21a-42b6-aa5f-6e2a26a0e628 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': task-1278092, 'name': Rename_Task, 'duration_secs': 0.12701} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1088.653266] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-81f6269b-a21a-42b6-aa5f-6e2a26a0e628 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 36769881-ff61-4bf3-b7e4-e7af19275805] Powering on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1088.653520] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6eb9ead9-fa8a-4fa5-b551-be3f65431536 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.659822] env[61957]: DEBUG oslo_vmware.api [None req-81f6269b-a21a-42b6-aa5f-6e2a26a0e628 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Waiting for the task: (returnval){ [ 1088.659822] env[61957]: value = "task-1278094" [ 1088.659822] env[61957]: _type = "Task" [ 1088.659822] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1088.667228] env[61957]: DEBUG oslo_vmware.api [None req-81f6269b-a21a-42b6-aa5f-6e2a26a0e628 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': task-1278094, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.026500] env[61957]: DEBUG oslo_vmware.api [None req-c68409b5-1674-400e-bc43-8bb6913b0678 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1278093, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.169704] env[61957]: DEBUG oslo_vmware.api [None req-81f6269b-a21a-42b6-aa5f-6e2a26a0e628 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': task-1278094, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.257010] env[61957]: DEBUG oslo_concurrency.lockutils [None req-091a8e64-7857-4c73-85c1-c32e51c082b4 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Acquiring lock "039a6b01-a745-478c-9ff7-c8c37afe2ce4" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1089.257299] env[61957]: DEBUG oslo_concurrency.lockutils [None req-091a8e64-7857-4c73-85c1-c32e51c082b4 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Lock "039a6b01-a745-478c-9ff7-c8c37afe2ce4" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1089.257547] env[61957]: INFO nova.compute.manager [None req-091a8e64-7857-4c73-85c1-c32e51c082b4 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] [instance: 039a6b01-a745-478c-9ff7-c8c37afe2ce4] Attaching volume d29b3bbe-886e-4384-9591-178f5fe64eea to /dev/sdb [ 1089.287332] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1242d01-388e-4c07-b69e-0d53a3f21106 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.295691] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdd6f0da-41aa-4e49-8eca-66c69dbb8907 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.308848] env[61957]: DEBUG nova.virt.block_device [None req-091a8e64-7857-4c73-85c1-c32e51c082b4 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] [instance: 039a6b01-a745-478c-9ff7-c8c37afe2ce4] Updating existing volume attachment record: 44b61501-4a05-450f-a78b-f204312c5b82 {{(pid=61957) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1089.527814] env[61957]: DEBUG oslo_vmware.api [None req-c68409b5-1674-400e-bc43-8bb6913b0678 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1278093, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.671407] env[61957]: DEBUG oslo_vmware.api [None req-81f6269b-a21a-42b6-aa5f-6e2a26a0e628 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': task-1278094, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.782427] env[61957]: DEBUG oslo_concurrency.lockutils [None req-fc650206-5e80-43a6-8c84-e8e89da4deab tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Acquiring lock "interface-4625d1dc-a621-4a33-9537-9e8cce96c914-b3373cf5-1370-4b75-b08c-f76c081ec78f" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1089.782656] env[61957]: DEBUG oslo_concurrency.lockutils [None req-fc650206-5e80-43a6-8c84-e8e89da4deab tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Lock "interface-4625d1dc-a621-4a33-9537-9e8cce96c914-b3373cf5-1370-4b75-b08c-f76c081ec78f" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1089.783112] env[61957]: DEBUG nova.objects.instance [None req-fc650206-5e80-43a6-8c84-e8e89da4deab tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Lazy-loading 'flavor' on Instance uuid 4625d1dc-a621-4a33-9537-9e8cce96c914 {{(pid=61957) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1090.029030] env[61957]: DEBUG oslo_vmware.api [None req-c68409b5-1674-400e-bc43-8bb6913b0678 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1278093, 'name': ReconfigVM_Task, 'duration_secs': 1.171018} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1090.029417] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-c68409b5-1674-400e-bc43-8bb6913b0678 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 84cd5619-4ac9-41ef-9368-a023ad9ae66d] Reconfigured VM instance instance-00000068 to detach disk 2000 {{(pid=61957) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1090.030350] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-215638a0-5eec-4837-ac31-88e232e514cd {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.054204] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-c68409b5-1674-400e-bc43-8bb6913b0678 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 84cd5619-4ac9-41ef-9368-a023ad9ae66d] Reconfiguring VM instance instance-00000068 to attach disk [datastore2] 84cd5619-4ac9-41ef-9368-a023ad9ae66d/84cd5619-4ac9-41ef-9368-a023ad9ae66d.vmdk or device None with type thin {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1090.054493] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bc543192-c9ea-4dc0-a2cf-39e6fc06dcac {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.074533] env[61957]: DEBUG oslo_vmware.api [None req-c68409b5-1674-400e-bc43-8bb6913b0678 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Waiting for the task: (returnval){ [ 1090.074533] env[61957]: value = "task-1278096" [ 1090.074533] env[61957]: _type = "Task" [ 1090.074533] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1090.082837] env[61957]: DEBUG oslo_vmware.api [None req-c68409b5-1674-400e-bc43-8bb6913b0678 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1278096, 'name': ReconfigVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.170941] env[61957]: DEBUG oslo_vmware.api [None req-81f6269b-a21a-42b6-aa5f-6e2a26a0e628 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': task-1278094, 'name': PowerOnVM_Task, 'duration_secs': 1.436452} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1090.171231] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-81f6269b-a21a-42b6-aa5f-6e2a26a0e628 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 36769881-ff61-4bf3-b7e4-e7af19275805] Powered on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1090.171459] env[61957]: INFO nova.compute.manager [None req-81f6269b-a21a-42b6-aa5f-6e2a26a0e628 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 36769881-ff61-4bf3-b7e4-e7af19275805] Took 8.22 seconds to spawn the instance on the hypervisor. [ 1090.171656] env[61957]: DEBUG nova.compute.manager [None req-81f6269b-a21a-42b6-aa5f-6e2a26a0e628 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 36769881-ff61-4bf3-b7e4-e7af19275805] Checking state {{(pid=61957) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1090.172465] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6e2f8b1-aeaf-4cd7-be24-ecaf42487be1 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.400068] env[61957]: DEBUG nova.objects.instance [None req-fc650206-5e80-43a6-8c84-e8e89da4deab tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Lazy-loading 'pci_requests' on Instance uuid 4625d1dc-a621-4a33-9537-9e8cce96c914 {{(pid=61957) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1090.584551] env[61957]: DEBUG oslo_vmware.api [None req-c68409b5-1674-400e-bc43-8bb6913b0678 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1278096, 'name': ReconfigVM_Task, 'duration_secs': 0.334964} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1090.584853] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-c68409b5-1674-400e-bc43-8bb6913b0678 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 84cd5619-4ac9-41ef-9368-a023ad9ae66d] Reconfigured VM instance instance-00000068 to attach disk [datastore2] 84cd5619-4ac9-41ef-9368-a023ad9ae66d/84cd5619-4ac9-41ef-9368-a023ad9ae66d.vmdk or device None with type thin {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1090.585149] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-c68409b5-1674-400e-bc43-8bb6913b0678 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 84cd5619-4ac9-41ef-9368-a023ad9ae66d] Updating instance '84cd5619-4ac9-41ef-9368-a023ad9ae66d' progress to 50 {{(pid=61957) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1090.689392] env[61957]: INFO nova.compute.manager [None req-81f6269b-a21a-42b6-aa5f-6e2a26a0e628 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 36769881-ff61-4bf3-b7e4-e7af19275805] Took 12.96 seconds to build instance. [ 1090.902305] env[61957]: DEBUG nova.objects.base [None req-fc650206-5e80-43a6-8c84-e8e89da4deab tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Object Instance<4625d1dc-a621-4a33-9537-9e8cce96c914> lazy-loaded attributes: flavor,pci_requests {{(pid=61957) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1090.902540] env[61957]: DEBUG nova.network.neutron [None req-fc650206-5e80-43a6-8c84-e8e89da4deab tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 4625d1dc-a621-4a33-9537-9e8cce96c914] allocate_for_instance() {{(pid=61957) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1090.965801] env[61957]: DEBUG nova.policy [None req-fc650206-5e80-43a6-8c84-e8e89da4deab tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6fd1c19dc3a44212ada44445e0919106', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8713aa35bcb24b86ad0b58ca9fc991ba', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61957) authorize /opt/stack/nova/nova/policy.py:203}} [ 1091.092457] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1009281-ecb5-4a01-8b6b-d32cc43c5e81 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.112282] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41813479-9464-49d6-8bd1-83a7364e130a {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.129191] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-c68409b5-1674-400e-bc43-8bb6913b0678 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 84cd5619-4ac9-41ef-9368-a023ad9ae66d] Updating instance '84cd5619-4ac9-41ef-9368-a023ad9ae66d' progress to 67 {{(pid=61957) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1091.191419] env[61957]: DEBUG oslo_concurrency.lockutils [None req-81f6269b-a21a-42b6-aa5f-6e2a26a0e628 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Lock "36769881-ff61-4bf3-b7e4-e7af19275805" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.466s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1091.284862] env[61957]: DEBUG oslo_concurrency.lockutils [None req-89a09176-fe06-41ee-b8ce-68d0c545aa8e tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Acquiring lock "36769881-ff61-4bf3-b7e4-e7af19275805" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1091.285170] env[61957]: DEBUG oslo_concurrency.lockutils [None req-89a09176-fe06-41ee-b8ce-68d0c545aa8e tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Lock "36769881-ff61-4bf3-b7e4-e7af19275805" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1091.285361] env[61957]: DEBUG nova.compute.manager [None req-89a09176-fe06-41ee-b8ce-68d0c545aa8e tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 36769881-ff61-4bf3-b7e4-e7af19275805] Checking state {{(pid=61957) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1091.286259] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb245899-3ce7-4bad-bf09-1701b8b42de7 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.292953] env[61957]: DEBUG nova.compute.manager [None req-89a09176-fe06-41ee-b8ce-68d0c545aa8e tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 36769881-ff61-4bf3-b7e4-e7af19275805] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61957) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1091.293512] env[61957]: DEBUG nova.objects.instance [None req-89a09176-fe06-41ee-b8ce-68d0c545aa8e tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Lazy-loading 'flavor' on Instance uuid 36769881-ff61-4bf3-b7e4-e7af19275805 {{(pid=61957) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1091.668388] env[61957]: DEBUG nova.network.neutron [None req-c68409b5-1674-400e-bc43-8bb6913b0678 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 84cd5619-4ac9-41ef-9368-a023ad9ae66d] Port b0541069-72e9-4384-b4f4-ccf5bc206dab binding to destination host cpu-1 is already ACTIVE {{(pid=61957) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 1091.797536] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-89a09176-fe06-41ee-b8ce-68d0c545aa8e tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 36769881-ff61-4bf3-b7e4-e7af19275805] Powering off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1091.797844] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f4cea793-dbde-4712-81f0-d4bca525064a {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.805385] env[61957]: DEBUG oslo_vmware.api [None req-89a09176-fe06-41ee-b8ce-68d0c545aa8e tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Waiting for the task: (returnval){ [ 1091.805385] env[61957]: value = "task-1278098" [ 1091.805385] env[61957]: _type = "Task" [ 1091.805385] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1091.813504] env[61957]: DEBUG oslo_vmware.api [None req-89a09176-fe06-41ee-b8ce-68d0c545aa8e tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': task-1278098, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1092.316153] env[61957]: DEBUG oslo_vmware.api [None req-89a09176-fe06-41ee-b8ce-68d0c545aa8e tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': task-1278098, 'name': PowerOffVM_Task, 'duration_secs': 0.155278} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1092.316334] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-89a09176-fe06-41ee-b8ce-68d0c545aa8e tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 36769881-ff61-4bf3-b7e4-e7af19275805] Powered off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1092.316687] env[61957]: DEBUG nova.compute.manager [None req-89a09176-fe06-41ee-b8ce-68d0c545aa8e tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 36769881-ff61-4bf3-b7e4-e7af19275805] Checking state {{(pid=61957) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1092.317559] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2074e152-bfe8-42d4-9701-0c999bd3e8f4 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.334662] env[61957]: DEBUG nova.compute.manager [req-6677e843-cd7e-4a0d-91d3-f1886fe3a5d6 req-7c90646e-e158-4e4d-80ee-4322eb7d90a4 service nova] [instance: 4625d1dc-a621-4a33-9537-9e8cce96c914] Received event network-vif-plugged-b3373cf5-1370-4b75-b08c-f76c081ec78f {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1092.334885] env[61957]: DEBUG oslo_concurrency.lockutils [req-6677e843-cd7e-4a0d-91d3-f1886fe3a5d6 req-7c90646e-e158-4e4d-80ee-4322eb7d90a4 service nova] Acquiring lock "4625d1dc-a621-4a33-9537-9e8cce96c914-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1092.335109] env[61957]: DEBUG oslo_concurrency.lockutils [req-6677e843-cd7e-4a0d-91d3-f1886fe3a5d6 req-7c90646e-e158-4e4d-80ee-4322eb7d90a4 service nova] Lock "4625d1dc-a621-4a33-9537-9e8cce96c914-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1092.335281] env[61957]: DEBUG oslo_concurrency.lockutils [req-6677e843-cd7e-4a0d-91d3-f1886fe3a5d6 req-7c90646e-e158-4e4d-80ee-4322eb7d90a4 service nova] Lock "4625d1dc-a621-4a33-9537-9e8cce96c914-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1092.335443] env[61957]: DEBUG nova.compute.manager [req-6677e843-cd7e-4a0d-91d3-f1886fe3a5d6 req-7c90646e-e158-4e4d-80ee-4322eb7d90a4 service nova] [instance: 4625d1dc-a621-4a33-9537-9e8cce96c914] No waiting events found dispatching network-vif-plugged-b3373cf5-1370-4b75-b08c-f76c081ec78f {{(pid=61957) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1092.335611] env[61957]: WARNING nova.compute.manager [req-6677e843-cd7e-4a0d-91d3-f1886fe3a5d6 req-7c90646e-e158-4e4d-80ee-4322eb7d90a4 service nova] [instance: 4625d1dc-a621-4a33-9537-9e8cce96c914] Received unexpected event network-vif-plugged-b3373cf5-1370-4b75-b08c-f76c081ec78f for instance with vm_state active and task_state None. [ 1092.426706] env[61957]: DEBUG nova.network.neutron [None req-fc650206-5e80-43a6-8c84-e8e89da4deab tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 4625d1dc-a621-4a33-9537-9e8cce96c914] Successfully updated port: b3373cf5-1370-4b75-b08c-f76c081ec78f {{(pid=61957) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1092.687200] env[61957]: DEBUG oslo_concurrency.lockutils [None req-c68409b5-1674-400e-bc43-8bb6913b0678 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Acquiring lock "84cd5619-4ac9-41ef-9368-a023ad9ae66d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1092.687449] env[61957]: DEBUG oslo_concurrency.lockutils [None req-c68409b5-1674-400e-bc43-8bb6913b0678 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Lock "84cd5619-4ac9-41ef-9368-a023ad9ae66d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1092.687626] env[61957]: DEBUG oslo_concurrency.lockutils [None req-c68409b5-1674-400e-bc43-8bb6913b0678 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Lock "84cd5619-4ac9-41ef-9368-a023ad9ae66d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1092.829268] env[61957]: DEBUG oslo_concurrency.lockutils [None req-89a09176-fe06-41ee-b8ce-68d0c545aa8e tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Lock "36769881-ff61-4bf3-b7e4-e7af19275805" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.544s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1092.930085] env[61957]: DEBUG oslo_concurrency.lockutils [None req-fc650206-5e80-43a6-8c84-e8e89da4deab tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Acquiring lock "refresh_cache-4625d1dc-a621-4a33-9537-9e8cce96c914" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1092.930293] env[61957]: DEBUG oslo_concurrency.lockutils [None req-fc650206-5e80-43a6-8c84-e8e89da4deab tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Acquired lock "refresh_cache-4625d1dc-a621-4a33-9537-9e8cce96c914" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1092.930493] env[61957]: DEBUG nova.network.neutron [None req-fc650206-5e80-43a6-8c84-e8e89da4deab tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 4625d1dc-a621-4a33-9537-9e8cce96c914] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1093.469681] env[61957]: WARNING nova.network.neutron [None req-fc650206-5e80-43a6-8c84-e8e89da4deab tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 4625d1dc-a621-4a33-9537-9e8cce96c914] 364cabcb-046d-4f91-b1ee-5dd3adfae6d7 already exists in list: networks containing: ['364cabcb-046d-4f91-b1ee-5dd3adfae6d7']. ignoring it [ 1093.739823] env[61957]: DEBUG oslo_concurrency.lockutils [None req-c68409b5-1674-400e-bc43-8bb6913b0678 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Acquiring lock "refresh_cache-84cd5619-4ac9-41ef-9368-a023ad9ae66d" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1093.740034] env[61957]: DEBUG oslo_concurrency.lockutils [None req-c68409b5-1674-400e-bc43-8bb6913b0678 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Acquired lock "refresh_cache-84cd5619-4ac9-41ef-9368-a023ad9ae66d" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1093.740210] env[61957]: DEBUG nova.network.neutron [None req-c68409b5-1674-400e-bc43-8bb6913b0678 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 84cd5619-4ac9-41ef-9368-a023ad9ae66d] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1093.755824] env[61957]: DEBUG nova.network.neutron [None req-fc650206-5e80-43a6-8c84-e8e89da4deab tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 4625d1dc-a621-4a33-9537-9e8cce96c914] Updating instance_info_cache with network_info: [{"id": "8365f86e-8b14-47de-a6f4-297107be492e", "address": "fa:16:3e:e8:59:9d", "network": {"id": "364cabcb-046d-4f91-b1ee-5dd3adfae6d7", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1432396967-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.232", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8713aa35bcb24b86ad0b58ca9fc991ba", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b7a73c01-1bb9-4612-a1a7-16d71b732e81", "external-id": "nsx-vlan-transportzone-711", "segmentation_id": 711, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8365f86e-8b", "ovs_interfaceid": "8365f86e-8b14-47de-a6f4-297107be492e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "b3373cf5-1370-4b75-b08c-f76c081ec78f", "address": "fa:16:3e:13:49:46", "network": {"id": "364cabcb-046d-4f91-b1ee-5dd3adfae6d7", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1432396967-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8713aa35bcb24b86ad0b58ca9fc991ba", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b7a73c01-1bb9-4612-a1a7-16d71b732e81", "external-id": "nsx-vlan-transportzone-711", "segmentation_id": 711, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb3373cf5-13", "ovs_interfaceid": "b3373cf5-1370-4b75-b08c-f76c081ec78f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1093.851078] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-091a8e64-7857-4c73-85c1-c32e51c082b4 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] [instance: 039a6b01-a745-478c-9ff7-c8c37afe2ce4] Volume attach. Driver type: vmdk {{(pid=61957) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1093.851429] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-091a8e64-7857-4c73-85c1-c32e51c082b4 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] [instance: 039a6b01-a745-478c-9ff7-c8c37afe2ce4] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-274608', 'volume_id': 'd29b3bbe-886e-4384-9591-178f5fe64eea', 'name': 'volume-d29b3bbe-886e-4384-9591-178f5fe64eea', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '039a6b01-a745-478c-9ff7-c8c37afe2ce4', 'attached_at': '', 'detached_at': '', 'volume_id': 'd29b3bbe-886e-4384-9591-178f5fe64eea', 'serial': 'd29b3bbe-886e-4384-9591-178f5fe64eea'} {{(pid=61957) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1093.852390] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da9996ba-c2c9-4442-9826-87977e98330b {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.868814] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2e8df3a-1590-4b4c-98f5-9d51eed9a9d4 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.892495] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-091a8e64-7857-4c73-85c1-c32e51c082b4 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] [instance: 039a6b01-a745-478c-9ff7-c8c37afe2ce4] Reconfiguring VM instance instance-00000062 to attach disk [datastore1] volume-d29b3bbe-886e-4384-9591-178f5fe64eea/volume-d29b3bbe-886e-4384-9591-178f5fe64eea.vmdk or device None with type thin {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1093.892786] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e9c44b91-938e-44e0-ba6d-7eefa9859ff0 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.908130] env[61957]: DEBUG oslo_concurrency.lockutils [None req-709a43cc-5ecb-4f90-bb40-dfd82c188b82 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Acquiring lock "36769881-ff61-4bf3-b7e4-e7af19275805" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1093.908359] env[61957]: DEBUG oslo_concurrency.lockutils [None req-709a43cc-5ecb-4f90-bb40-dfd82c188b82 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Lock "36769881-ff61-4bf3-b7e4-e7af19275805" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1093.908561] env[61957]: DEBUG oslo_concurrency.lockutils [None req-709a43cc-5ecb-4f90-bb40-dfd82c188b82 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Acquiring lock "36769881-ff61-4bf3-b7e4-e7af19275805-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1093.908747] env[61957]: DEBUG oslo_concurrency.lockutils [None req-709a43cc-5ecb-4f90-bb40-dfd82c188b82 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Lock "36769881-ff61-4bf3-b7e4-e7af19275805-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1093.908918] env[61957]: DEBUG oslo_concurrency.lockutils [None req-709a43cc-5ecb-4f90-bb40-dfd82c188b82 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Lock "36769881-ff61-4bf3-b7e4-e7af19275805-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1093.911526] env[61957]: DEBUG oslo_vmware.api [None req-091a8e64-7857-4c73-85c1-c32e51c082b4 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Waiting for the task: (returnval){ [ 1093.911526] env[61957]: value = "task-1278099" [ 1093.911526] env[61957]: _type = "Task" [ 1093.911526] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1093.911948] env[61957]: INFO nova.compute.manager [None req-709a43cc-5ecb-4f90-bb40-dfd82c188b82 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 36769881-ff61-4bf3-b7e4-e7af19275805] Terminating instance [ 1093.917963] env[61957]: DEBUG nova.compute.manager [None req-709a43cc-5ecb-4f90-bb40-dfd82c188b82 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 36769881-ff61-4bf3-b7e4-e7af19275805] Start destroying the instance on the hypervisor. {{(pid=61957) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1093.918179] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-709a43cc-5ecb-4f90-bb40-dfd82c188b82 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 36769881-ff61-4bf3-b7e4-e7af19275805] Destroying instance {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1093.918860] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42d88969-9f95-4b58-8216-5bfb699a2ec6 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.926358] env[61957]: DEBUG oslo_vmware.api [None req-091a8e64-7857-4c73-85c1-c32e51c082b4 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Task: {'id': task-1278099, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1093.928494] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-709a43cc-5ecb-4f90-bb40-dfd82c188b82 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 36769881-ff61-4bf3-b7e4-e7af19275805] Unregistering the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1093.928710] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9df1ffd0-a2be-4390-968b-de1418cd43a4 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.983754] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-709a43cc-5ecb-4f90-bb40-dfd82c188b82 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 36769881-ff61-4bf3-b7e4-e7af19275805] Unregistered the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1093.984010] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-709a43cc-5ecb-4f90-bb40-dfd82c188b82 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 36769881-ff61-4bf3-b7e4-e7af19275805] Deleting contents of the VM from datastore datastore1 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1093.984278] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-709a43cc-5ecb-4f90-bb40-dfd82c188b82 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Deleting the datastore file [datastore1] 36769881-ff61-4bf3-b7e4-e7af19275805 {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1093.984479] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3149fb4e-e1e3-4d7f-926d-02c1bb59d1ff {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.990385] env[61957]: DEBUG oslo_vmware.api [None req-709a43cc-5ecb-4f90-bb40-dfd82c188b82 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Waiting for the task: (returnval){ [ 1093.990385] env[61957]: value = "task-1278101" [ 1093.990385] env[61957]: _type = "Task" [ 1093.990385] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1093.997775] env[61957]: DEBUG oslo_vmware.api [None req-709a43cc-5ecb-4f90-bb40-dfd82c188b82 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': task-1278101, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1094.258507] env[61957]: DEBUG oslo_concurrency.lockutils [None req-fc650206-5e80-43a6-8c84-e8e89da4deab tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Releasing lock "refresh_cache-4625d1dc-a621-4a33-9537-9e8cce96c914" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1094.259191] env[61957]: DEBUG oslo_concurrency.lockutils [None req-fc650206-5e80-43a6-8c84-e8e89da4deab tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Acquiring lock "4625d1dc-a621-4a33-9537-9e8cce96c914" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1094.259351] env[61957]: DEBUG oslo_concurrency.lockutils [None req-fc650206-5e80-43a6-8c84-e8e89da4deab tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Acquired lock "4625d1dc-a621-4a33-9537-9e8cce96c914" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1094.260267] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-796451e6-d7d5-48fd-b7e8-93ed47d4eb41 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.277212] env[61957]: DEBUG nova.virt.hardware [None req-fc650206-5e80-43a6-8c84-e8e89da4deab tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T17:22:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1094.277461] env[61957]: DEBUG nova.virt.hardware [None req-fc650206-5e80-43a6-8c84-e8e89da4deab tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Flavor limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1094.277665] env[61957]: DEBUG nova.virt.hardware [None req-fc650206-5e80-43a6-8c84-e8e89da4deab tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Image limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1094.277806] env[61957]: DEBUG nova.virt.hardware [None req-fc650206-5e80-43a6-8c84-e8e89da4deab tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Flavor pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1094.278009] env[61957]: DEBUG nova.virt.hardware [None req-fc650206-5e80-43a6-8c84-e8e89da4deab tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Image pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1094.278200] env[61957]: DEBUG nova.virt.hardware [None req-fc650206-5e80-43a6-8c84-e8e89da4deab tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1094.278411] env[61957]: DEBUG nova.virt.hardware [None req-fc650206-5e80-43a6-8c84-e8e89da4deab tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1094.278580] env[61957]: DEBUG nova.virt.hardware [None req-fc650206-5e80-43a6-8c84-e8e89da4deab tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1094.278752] env[61957]: DEBUG nova.virt.hardware [None req-fc650206-5e80-43a6-8c84-e8e89da4deab tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Got 1 possible topologies {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1094.278912] env[61957]: DEBUG nova.virt.hardware [None req-fc650206-5e80-43a6-8c84-e8e89da4deab tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1094.279106] env[61957]: DEBUG nova.virt.hardware [None req-fc650206-5e80-43a6-8c84-e8e89da4deab tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1094.285493] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-fc650206-5e80-43a6-8c84-e8e89da4deab tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 4625d1dc-a621-4a33-9537-9e8cce96c914] Reconfiguring VM to attach interface {{(pid=61957) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1929}} [ 1094.287907] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0098fcd9-b62d-4925-b16e-010aa4b86301 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.305282] env[61957]: DEBUG oslo_vmware.api [None req-fc650206-5e80-43a6-8c84-e8e89da4deab tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Waiting for the task: (returnval){ [ 1094.305282] env[61957]: value = "task-1278102" [ 1094.305282] env[61957]: _type = "Task" [ 1094.305282] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1094.312818] env[61957]: DEBUG oslo_vmware.api [None req-fc650206-5e80-43a6-8c84-e8e89da4deab tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': task-1278102, 'name': ReconfigVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1094.363837] env[61957]: DEBUG nova.compute.manager [req-17815a0e-59b4-4e60-944d-b87a1c2c3cc7 req-0fbe87ae-2f94-4967-ba97-9fcc8bdcb358 service nova] [instance: 4625d1dc-a621-4a33-9537-9e8cce96c914] Received event network-changed-b3373cf5-1370-4b75-b08c-f76c081ec78f {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1094.364150] env[61957]: DEBUG nova.compute.manager [req-17815a0e-59b4-4e60-944d-b87a1c2c3cc7 req-0fbe87ae-2f94-4967-ba97-9fcc8bdcb358 service nova] [instance: 4625d1dc-a621-4a33-9537-9e8cce96c914] Refreshing instance network info cache due to event network-changed-b3373cf5-1370-4b75-b08c-f76c081ec78f. {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1094.364388] env[61957]: DEBUG oslo_concurrency.lockutils [req-17815a0e-59b4-4e60-944d-b87a1c2c3cc7 req-0fbe87ae-2f94-4967-ba97-9fcc8bdcb358 service nova] Acquiring lock "refresh_cache-4625d1dc-a621-4a33-9537-9e8cce96c914" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1094.364536] env[61957]: DEBUG oslo_concurrency.lockutils [req-17815a0e-59b4-4e60-944d-b87a1c2c3cc7 req-0fbe87ae-2f94-4967-ba97-9fcc8bdcb358 service nova] Acquired lock "refresh_cache-4625d1dc-a621-4a33-9537-9e8cce96c914" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1094.364699] env[61957]: DEBUG nova.network.neutron [req-17815a0e-59b4-4e60-944d-b87a1c2c3cc7 req-0fbe87ae-2f94-4967-ba97-9fcc8bdcb358 service nova] [instance: 4625d1dc-a621-4a33-9537-9e8cce96c914] Refreshing network info cache for port b3373cf5-1370-4b75-b08c-f76c081ec78f {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1094.421930] env[61957]: DEBUG oslo_vmware.api [None req-091a8e64-7857-4c73-85c1-c32e51c082b4 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Task: {'id': task-1278099, 'name': ReconfigVM_Task, 'duration_secs': 0.330097} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1094.422232] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-091a8e64-7857-4c73-85c1-c32e51c082b4 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] [instance: 039a6b01-a745-478c-9ff7-c8c37afe2ce4] Reconfigured VM instance instance-00000062 to attach disk [datastore1] volume-d29b3bbe-886e-4384-9591-178f5fe64eea/volume-d29b3bbe-886e-4384-9591-178f5fe64eea.vmdk or device None with type thin {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1094.426960] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3ceef3cc-fa51-4df8-8051-10d889b58015 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.444956] env[61957]: DEBUG oslo_vmware.api [None req-091a8e64-7857-4c73-85c1-c32e51c082b4 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Waiting for the task: (returnval){ [ 1094.444956] env[61957]: value = "task-1278103" [ 1094.444956] env[61957]: _type = "Task" [ 1094.444956] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1094.452471] env[61957]: DEBUG oslo_vmware.api [None req-091a8e64-7857-4c73-85c1-c32e51c082b4 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Task: {'id': task-1278103, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1094.470699] env[61957]: DEBUG nova.network.neutron [None req-c68409b5-1674-400e-bc43-8bb6913b0678 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 84cd5619-4ac9-41ef-9368-a023ad9ae66d] Updating instance_info_cache with network_info: [{"id": "b0541069-72e9-4384-b4f4-ccf5bc206dab", "address": "fa:16:3e:d1:ff:5a", "network": {"id": "efe08823-51fe-4620-a849-5345e262cc71", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1035436819-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e16f6dce3f0e44fb96de516f17d4c6f5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "667a2e97-c1be-421d-9941-6b84c2629b43", "external-id": "nsx-vlan-transportzone-484", "segmentation_id": 484, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb0541069-72", "ovs_interfaceid": "b0541069-72e9-4384-b4f4-ccf5bc206dab", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1094.499740] env[61957]: DEBUG oslo_vmware.api [None req-709a43cc-5ecb-4f90-bb40-dfd82c188b82 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': task-1278101, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.125556} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1094.500017] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-709a43cc-5ecb-4f90-bb40-dfd82c188b82 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Deleted the datastore file {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1094.500224] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-709a43cc-5ecb-4f90-bb40-dfd82c188b82 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 36769881-ff61-4bf3-b7e4-e7af19275805] Deleted contents of the VM from datastore datastore1 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1094.500409] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-709a43cc-5ecb-4f90-bb40-dfd82c188b82 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 36769881-ff61-4bf3-b7e4-e7af19275805] Instance destroyed {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1094.500586] env[61957]: INFO nova.compute.manager [None req-709a43cc-5ecb-4f90-bb40-dfd82c188b82 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 36769881-ff61-4bf3-b7e4-e7af19275805] Took 0.58 seconds to destroy the instance on the hypervisor. [ 1094.500866] env[61957]: DEBUG oslo.service.loopingcall [None req-709a43cc-5ecb-4f90-bb40-dfd82c188b82 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1094.501091] env[61957]: DEBUG nova.compute.manager [-] [instance: 36769881-ff61-4bf3-b7e4-e7af19275805] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1094.501190] env[61957]: DEBUG nova.network.neutron [-] [instance: 36769881-ff61-4bf3-b7e4-e7af19275805] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1094.814594] env[61957]: DEBUG oslo_vmware.api [None req-fc650206-5e80-43a6-8c84-e8e89da4deab tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': task-1278102, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1094.957869] env[61957]: DEBUG oslo_vmware.api [None req-091a8e64-7857-4c73-85c1-c32e51c082b4 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Task: {'id': task-1278103, 'name': ReconfigVM_Task, 'duration_secs': 0.13094} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1094.958295] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-091a8e64-7857-4c73-85c1-c32e51c082b4 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] [instance: 039a6b01-a745-478c-9ff7-c8c37afe2ce4] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-274608', 'volume_id': 'd29b3bbe-886e-4384-9591-178f5fe64eea', 'name': 'volume-d29b3bbe-886e-4384-9591-178f5fe64eea', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '039a6b01-a745-478c-9ff7-c8c37afe2ce4', 'attached_at': '', 'detached_at': '', 'volume_id': 'd29b3bbe-886e-4384-9591-178f5fe64eea', 'serial': 'd29b3bbe-886e-4384-9591-178f5fe64eea'} {{(pid=61957) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1094.973953] env[61957]: DEBUG oslo_concurrency.lockutils [None req-c68409b5-1674-400e-bc43-8bb6913b0678 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Releasing lock "refresh_cache-84cd5619-4ac9-41ef-9368-a023ad9ae66d" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1095.073726] env[61957]: DEBUG nova.network.neutron [req-17815a0e-59b4-4e60-944d-b87a1c2c3cc7 req-0fbe87ae-2f94-4967-ba97-9fcc8bdcb358 service nova] [instance: 4625d1dc-a621-4a33-9537-9e8cce96c914] Updated VIF entry in instance network info cache for port b3373cf5-1370-4b75-b08c-f76c081ec78f. {{(pid=61957) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1095.074201] env[61957]: DEBUG nova.network.neutron [req-17815a0e-59b4-4e60-944d-b87a1c2c3cc7 req-0fbe87ae-2f94-4967-ba97-9fcc8bdcb358 service nova] [instance: 4625d1dc-a621-4a33-9537-9e8cce96c914] Updating instance_info_cache with network_info: [{"id": "8365f86e-8b14-47de-a6f4-297107be492e", "address": "fa:16:3e:e8:59:9d", "network": {"id": "364cabcb-046d-4f91-b1ee-5dd3adfae6d7", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1432396967-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.232", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8713aa35bcb24b86ad0b58ca9fc991ba", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b7a73c01-1bb9-4612-a1a7-16d71b732e81", "external-id": "nsx-vlan-transportzone-711", "segmentation_id": 711, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8365f86e-8b", "ovs_interfaceid": "8365f86e-8b14-47de-a6f4-297107be492e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "b3373cf5-1370-4b75-b08c-f76c081ec78f", "address": "fa:16:3e:13:49:46", "network": {"id": "364cabcb-046d-4f91-b1ee-5dd3adfae6d7", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1432396967-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8713aa35bcb24b86ad0b58ca9fc991ba", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b7a73c01-1bb9-4612-a1a7-16d71b732e81", "external-id": "nsx-vlan-transportzone-711", "segmentation_id": 711, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb3373cf5-13", "ovs_interfaceid": "b3373cf5-1370-4b75-b08c-f76c081ec78f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1095.234147] env[61957]: DEBUG nova.network.neutron [-] [instance: 36769881-ff61-4bf3-b7e4-e7af19275805] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1095.315995] env[61957]: DEBUG oslo_vmware.api [None req-fc650206-5e80-43a6-8c84-e8e89da4deab tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': task-1278102, 'name': ReconfigVM_Task, 'duration_secs': 0.520962} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1095.316530] env[61957]: DEBUG oslo_concurrency.lockutils [None req-fc650206-5e80-43a6-8c84-e8e89da4deab tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Releasing lock "4625d1dc-a621-4a33-9537-9e8cce96c914" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1095.316755] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-fc650206-5e80-43a6-8c84-e8e89da4deab tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 4625d1dc-a621-4a33-9537-9e8cce96c914] Reconfigured VM to attach interface {{(pid=61957) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1943}} [ 1095.501112] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c74771e1-b170-4c68-ac34-8aea272ab867 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.520865] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ffa050a-8545-402c-aa0e-8b7c1d3db8c8 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.527713] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-c68409b5-1674-400e-bc43-8bb6913b0678 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 84cd5619-4ac9-41ef-9368-a023ad9ae66d] Updating instance '84cd5619-4ac9-41ef-9368-a023ad9ae66d' progress to 83 {{(pid=61957) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1095.576467] env[61957]: DEBUG oslo_concurrency.lockutils [req-17815a0e-59b4-4e60-944d-b87a1c2c3cc7 req-0fbe87ae-2f94-4967-ba97-9fcc8bdcb358 service nova] Releasing lock "refresh_cache-4625d1dc-a621-4a33-9537-9e8cce96c914" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1095.736357] env[61957]: INFO nova.compute.manager [-] [instance: 36769881-ff61-4bf3-b7e4-e7af19275805] Took 1.23 seconds to deallocate network for instance. [ 1095.821631] env[61957]: DEBUG oslo_concurrency.lockutils [None req-fc650206-5e80-43a6-8c84-e8e89da4deab tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Lock "interface-4625d1dc-a621-4a33-9537-9e8cce96c914-b3373cf5-1370-4b75-b08c-f76c081ec78f" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 6.039s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1095.994953] env[61957]: DEBUG nova.objects.instance [None req-091a8e64-7857-4c73-85c1-c32e51c082b4 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Lazy-loading 'flavor' on Instance uuid 039a6b01-a745-478c-9ff7-c8c37afe2ce4 {{(pid=61957) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1096.033377] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-c68409b5-1674-400e-bc43-8bb6913b0678 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 84cd5619-4ac9-41ef-9368-a023ad9ae66d] Powering on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1096.033689] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b57e48ee-17c6-4592-9d19-7638ef29bcd1 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.044550] env[61957]: DEBUG oslo_vmware.api [None req-c68409b5-1674-400e-bc43-8bb6913b0678 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Waiting for the task: (returnval){ [ 1096.044550] env[61957]: value = "task-1278104" [ 1096.044550] env[61957]: _type = "Task" [ 1096.044550] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1096.052379] env[61957]: DEBUG oslo_vmware.api [None req-c68409b5-1674-400e-bc43-8bb6913b0678 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1278104, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1096.242709] env[61957]: DEBUG oslo_concurrency.lockutils [None req-709a43cc-5ecb-4f90-bb40-dfd82c188b82 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1096.243056] env[61957]: DEBUG oslo_concurrency.lockutils [None req-709a43cc-5ecb-4f90-bb40-dfd82c188b82 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1096.243382] env[61957]: DEBUG nova.objects.instance [None req-709a43cc-5ecb-4f90-bb40-dfd82c188b82 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Lazy-loading 'resources' on Instance uuid 36769881-ff61-4bf3-b7e4-e7af19275805 {{(pid=61957) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1096.389928] env[61957]: DEBUG nova.compute.manager [req-ac0fe2e6-e807-41ab-8209-917b306dcbe7 req-f40069d4-7bf4-47fd-9e23-96eae96364e5 service nova] [instance: 36769881-ff61-4bf3-b7e4-e7af19275805] Received event network-vif-deleted-09f655dd-f483-42c4-b37b-c0985ad5de04 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1096.501014] env[61957]: DEBUG oslo_concurrency.lockutils [None req-091a8e64-7857-4c73-85c1-c32e51c082b4 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Lock "039a6b01-a745-478c-9ff7-c8c37afe2ce4" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.243s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1096.554513] env[61957]: DEBUG oslo_vmware.api [None req-c68409b5-1674-400e-bc43-8bb6913b0678 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1278104, 'name': PowerOnVM_Task, 'duration_secs': 0.343839} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1096.554787] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-c68409b5-1674-400e-bc43-8bb6913b0678 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 84cd5619-4ac9-41ef-9368-a023ad9ae66d] Powered on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1096.555047] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-c68409b5-1674-400e-bc43-8bb6913b0678 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 84cd5619-4ac9-41ef-9368-a023ad9ae66d] Updating instance '84cd5619-4ac9-41ef-9368-a023ad9ae66d' progress to 100 {{(pid=61957) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1096.839039] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d594b222-3e2f-4714-9339-5468540bed9f {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.846607] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08459c0e-8e41-4874-a254-f642c83f4ad9 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.877431] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa338d45-b88f-4670-a6ed-657b84331656 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.884734] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21cede49-9c3b-4e76-aa36-a957abe5ff9a {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.898295] env[61957]: DEBUG nova.compute.provider_tree [None req-709a43cc-5ecb-4f90-bb40-dfd82c188b82 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1097.402920] env[61957]: DEBUG nova.scheduler.client.report [None req-709a43cc-5ecb-4f90-bb40-dfd82c188b82 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1097.416458] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ac6e1710-ee9b-4755-89eb-bd5d0557f455 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Acquiring lock "039a6b01-a745-478c-9ff7-c8c37afe2ce4" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1097.416458] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ac6e1710-ee9b-4755-89eb-bd5d0557f455 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Lock "039a6b01-a745-478c-9ff7-c8c37afe2ce4" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1097.466783] env[61957]: DEBUG oslo_concurrency.lockutils [None req-0cc2a3a7-4047-410e-a46d-b5ab6f7151bc tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Acquiring lock "interface-4625d1dc-a621-4a33-9537-9e8cce96c914-b3373cf5-1370-4b75-b08c-f76c081ec78f" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1097.467046] env[61957]: DEBUG oslo_concurrency.lockutils [None req-0cc2a3a7-4047-410e-a46d-b5ab6f7151bc tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Lock "interface-4625d1dc-a621-4a33-9537-9e8cce96c914-b3373cf5-1370-4b75-b08c-f76c081ec78f" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1097.909553] env[61957]: DEBUG oslo_concurrency.lockutils [None req-709a43cc-5ecb-4f90-bb40-dfd82c188b82 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.666s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1097.920118] env[61957]: DEBUG nova.compute.utils [None req-ac6e1710-ee9b-4755-89eb-bd5d0557f455 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Using /dev/sd instead of None {{(pid=61957) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1097.928106] env[61957]: INFO nova.scheduler.client.report [None req-709a43cc-5ecb-4f90-bb40-dfd82c188b82 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Deleted allocations for instance 36769881-ff61-4bf3-b7e4-e7af19275805 [ 1097.969872] env[61957]: DEBUG oslo_concurrency.lockutils [None req-0cc2a3a7-4047-410e-a46d-b5ab6f7151bc tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Acquiring lock "4625d1dc-a621-4a33-9537-9e8cce96c914" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1097.970121] env[61957]: DEBUG oslo_concurrency.lockutils [None req-0cc2a3a7-4047-410e-a46d-b5ab6f7151bc tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Acquired lock "4625d1dc-a621-4a33-9537-9e8cce96c914" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1097.971503] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2117994d-1bd5-46d5-bb3d-799ff630cc9b {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.988739] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6967a189-a5fa-4849-9940-3ad2358a1f20 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.013041] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-0cc2a3a7-4047-410e-a46d-b5ab6f7151bc tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 4625d1dc-a621-4a33-9537-9e8cce96c914] Reconfiguring VM to detach interface {{(pid=61957) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1974}} [ 1098.013242] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cd67111c-c8c1-4a43-9858-3bfb3ee3f2f8 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.031116] env[61957]: DEBUG oslo_vmware.api [None req-0cc2a3a7-4047-410e-a46d-b5ab6f7151bc tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Waiting for the task: (returnval){ [ 1098.031116] env[61957]: value = "task-1278105" [ 1098.031116] env[61957]: _type = "Task" [ 1098.031116] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1098.038692] env[61957]: DEBUG oslo_vmware.api [None req-0cc2a3a7-4047-410e-a46d-b5ab6f7151bc tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': task-1278105, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.423696] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ac6e1710-ee9b-4755-89eb-bd5d0557f455 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Lock "039a6b01-a745-478c-9ff7-c8c37afe2ce4" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.007s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1098.436482] env[61957]: DEBUG oslo_concurrency.lockutils [None req-709a43cc-5ecb-4f90-bb40-dfd82c188b82 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Lock "36769881-ff61-4bf3-b7e4-e7af19275805" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.528s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1098.540823] env[61957]: DEBUG oslo_vmware.api [None req-0cc2a3a7-4047-410e-a46d-b5ab6f7151bc tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': task-1278105, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.983008] env[61957]: DEBUG oslo_concurrency.lockutils [None req-017d3b28-7c60-435c-8f0d-fb9db5de2a93 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Acquiring lock "84cd5619-4ac9-41ef-9368-a023ad9ae66d" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1098.983314] env[61957]: DEBUG oslo_concurrency.lockutils [None req-017d3b28-7c60-435c-8f0d-fb9db5de2a93 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Lock "84cd5619-4ac9-41ef-9368-a023ad9ae66d" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.001s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1098.983513] env[61957]: DEBUG nova.compute.manager [None req-017d3b28-7c60-435c-8f0d-fb9db5de2a93 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 84cd5619-4ac9-41ef-9368-a023ad9ae66d] Going to confirm migration 3 {{(pid=61957) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 1099.042234] env[61957]: DEBUG oslo_vmware.api [None req-0cc2a3a7-4047-410e-a46d-b5ab6f7151bc tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': task-1278105, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1099.492319] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ac6e1710-ee9b-4755-89eb-bd5d0557f455 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Acquiring lock "039a6b01-a745-478c-9ff7-c8c37afe2ce4" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1099.494027] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ac6e1710-ee9b-4755-89eb-bd5d0557f455 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Lock "039a6b01-a745-478c-9ff7-c8c37afe2ce4" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1099.494027] env[61957]: INFO nova.compute.manager [None req-ac6e1710-ee9b-4755-89eb-bd5d0557f455 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] [instance: 039a6b01-a745-478c-9ff7-c8c37afe2ce4] Attaching volume 95dd15cb-394c-419f-bbc2-aa25920febe7 to /dev/sdc [ 1099.524330] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7cd0996-a44c-4066-a53c-2adebaf129b4 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.531888] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0aa6995f-ce40-406a-a683-c71052447065 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.541541] env[61957]: DEBUG oslo_vmware.api [None req-0cc2a3a7-4047-410e-a46d-b5ab6f7151bc tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': task-1278105, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1099.546714] env[61957]: DEBUG nova.virt.block_device [None req-ac6e1710-ee9b-4755-89eb-bd5d0557f455 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] [instance: 039a6b01-a745-478c-9ff7-c8c37afe2ce4] Updating existing volume attachment record: 9d49701b-21b8-4859-9e7a-16642cee87bb {{(pid=61957) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1099.561409] env[61957]: DEBUG oslo_concurrency.lockutils [None req-017d3b28-7c60-435c-8f0d-fb9db5de2a93 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Acquiring lock "refresh_cache-84cd5619-4ac9-41ef-9368-a023ad9ae66d" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1099.561594] env[61957]: DEBUG oslo_concurrency.lockutils [None req-017d3b28-7c60-435c-8f0d-fb9db5de2a93 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Acquired lock "refresh_cache-84cd5619-4ac9-41ef-9368-a023ad9ae66d" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1099.561793] env[61957]: DEBUG nova.network.neutron [None req-017d3b28-7c60-435c-8f0d-fb9db5de2a93 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 84cd5619-4ac9-41ef-9368-a023ad9ae66d] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1099.561963] env[61957]: DEBUG nova.objects.instance [None req-017d3b28-7c60-435c-8f0d-fb9db5de2a93 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Lazy-loading 'info_cache' on Instance uuid 84cd5619-4ac9-41ef-9368-a023ad9ae66d {{(pid=61957) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1099.788825] env[61957]: DEBUG oslo_concurrency.lockutils [None req-77f64c0f-751d-4c0e-b32e-91391a5af017 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Acquiring lock "1a0327c2-5671-4970-9db7-c7cc912d8678" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1099.789189] env[61957]: DEBUG oslo_concurrency.lockutils [None req-77f64c0f-751d-4c0e-b32e-91391a5af017 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Lock "1a0327c2-5671-4970-9db7-c7cc912d8678" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1099.789462] env[61957]: DEBUG oslo_concurrency.lockutils [None req-77f64c0f-751d-4c0e-b32e-91391a5af017 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Acquiring lock "1a0327c2-5671-4970-9db7-c7cc912d8678-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1099.789730] env[61957]: DEBUG oslo_concurrency.lockutils [None req-77f64c0f-751d-4c0e-b32e-91391a5af017 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Lock "1a0327c2-5671-4970-9db7-c7cc912d8678-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1099.789926] env[61957]: DEBUG oslo_concurrency.lockutils [None req-77f64c0f-751d-4c0e-b32e-91391a5af017 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Lock "1a0327c2-5671-4970-9db7-c7cc912d8678-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1099.792107] env[61957]: INFO nova.compute.manager [None req-77f64c0f-751d-4c0e-b32e-91391a5af017 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 1a0327c2-5671-4970-9db7-c7cc912d8678] Terminating instance [ 1099.794024] env[61957]: DEBUG nova.compute.manager [None req-77f64c0f-751d-4c0e-b32e-91391a5af017 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 1a0327c2-5671-4970-9db7-c7cc912d8678] Start destroying the instance on the hypervisor. {{(pid=61957) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1099.794239] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-77f64c0f-751d-4c0e-b32e-91391a5af017 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 1a0327c2-5671-4970-9db7-c7cc912d8678] Destroying instance {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1099.795098] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86ff3ac9-78e2-48c2-ad0f-c1831e463ac7 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.804390] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-77f64c0f-751d-4c0e-b32e-91391a5af017 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 1a0327c2-5671-4970-9db7-c7cc912d8678] Powering off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1099.804630] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c17d2221-9d54-4fb9-b509-cb1c174287fd {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.811017] env[61957]: DEBUG oslo_vmware.api [None req-77f64c0f-751d-4c0e-b32e-91391a5af017 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Waiting for the task: (returnval){ [ 1099.811017] env[61957]: value = "task-1278107" [ 1099.811017] env[61957]: _type = "Task" [ 1099.811017] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1099.821127] env[61957]: DEBUG oslo_vmware.api [None req-77f64c0f-751d-4c0e-b32e-91391a5af017 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': task-1278107, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.044350] env[61957]: DEBUG oslo_vmware.api [None req-0cc2a3a7-4047-410e-a46d-b5ab6f7151bc tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': task-1278105, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.321066] env[61957]: DEBUG oslo_vmware.api [None req-77f64c0f-751d-4c0e-b32e-91391a5af017 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': task-1278107, 'name': PowerOffVM_Task} progress is 100%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.543910] env[61957]: DEBUG oslo_vmware.api [None req-0cc2a3a7-4047-410e-a46d-b5ab6f7151bc tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': task-1278105, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.821103] env[61957]: DEBUG oslo_vmware.api [None req-77f64c0f-751d-4c0e-b32e-91391a5af017 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': task-1278107, 'name': PowerOffVM_Task, 'duration_secs': 0.802156} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1100.821385] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-77f64c0f-751d-4c0e-b32e-91391a5af017 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 1a0327c2-5671-4970-9db7-c7cc912d8678] Powered off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1100.821556] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-77f64c0f-751d-4c0e-b32e-91391a5af017 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 1a0327c2-5671-4970-9db7-c7cc912d8678] Unregistering the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1100.821813] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7aa73e47-bb9e-4c05-9beb-3e232336d7d8 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.897668] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-77f64c0f-751d-4c0e-b32e-91391a5af017 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 1a0327c2-5671-4970-9db7-c7cc912d8678] Unregistered the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1100.897898] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-77f64c0f-751d-4c0e-b32e-91391a5af017 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 1a0327c2-5671-4970-9db7-c7cc912d8678] Deleting contents of the VM from datastore datastore2 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1100.898106] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-77f64c0f-751d-4c0e-b32e-91391a5af017 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Deleting the datastore file [datastore2] 1a0327c2-5671-4970-9db7-c7cc912d8678 {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1100.898381] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5bc8e7b5-9356-4f7b-b4a6-9e2a82b4cdb4 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.904189] env[61957]: DEBUG oslo_vmware.api [None req-77f64c0f-751d-4c0e-b32e-91391a5af017 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Waiting for the task: (returnval){ [ 1100.904189] env[61957]: value = "task-1278109" [ 1100.904189] env[61957]: _type = "Task" [ 1100.904189] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1100.911961] env[61957]: DEBUG oslo_vmware.api [None req-77f64c0f-751d-4c0e-b32e-91391a5af017 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': task-1278109, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.924235] env[61957]: DEBUG nova.network.neutron [None req-017d3b28-7c60-435c-8f0d-fb9db5de2a93 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 84cd5619-4ac9-41ef-9368-a023ad9ae66d] Updating instance_info_cache with network_info: [{"id": "b0541069-72e9-4384-b4f4-ccf5bc206dab", "address": "fa:16:3e:d1:ff:5a", "network": {"id": "efe08823-51fe-4620-a849-5345e262cc71", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1035436819-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e16f6dce3f0e44fb96de516f17d4c6f5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "667a2e97-c1be-421d-9941-6b84c2629b43", "external-id": "nsx-vlan-transportzone-484", "segmentation_id": 484, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb0541069-72", "ovs_interfaceid": "b0541069-72e9-4384-b4f4-ccf5bc206dab", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1101.044424] env[61957]: DEBUG oslo_vmware.api [None req-0cc2a3a7-4047-410e-a46d-b5ab6f7151bc tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': task-1278105, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1101.414029] env[61957]: DEBUG oslo_vmware.api [None req-77f64c0f-751d-4c0e-b32e-91391a5af017 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Task: {'id': task-1278109, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.12971} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1101.414251] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-77f64c0f-751d-4c0e-b32e-91391a5af017 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Deleted the datastore file {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1101.414437] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-77f64c0f-751d-4c0e-b32e-91391a5af017 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 1a0327c2-5671-4970-9db7-c7cc912d8678] Deleted contents of the VM from datastore datastore2 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1101.414615] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-77f64c0f-751d-4c0e-b32e-91391a5af017 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 1a0327c2-5671-4970-9db7-c7cc912d8678] Instance destroyed {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1101.414792] env[61957]: INFO nova.compute.manager [None req-77f64c0f-751d-4c0e-b32e-91391a5af017 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] [instance: 1a0327c2-5671-4970-9db7-c7cc912d8678] Took 1.62 seconds to destroy the instance on the hypervisor. [ 1101.415044] env[61957]: DEBUG oslo.service.loopingcall [None req-77f64c0f-751d-4c0e-b32e-91391a5af017 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1101.415270] env[61957]: DEBUG nova.compute.manager [-] [instance: 1a0327c2-5671-4970-9db7-c7cc912d8678] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1101.415367] env[61957]: DEBUG nova.network.neutron [-] [instance: 1a0327c2-5671-4970-9db7-c7cc912d8678] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1101.426458] env[61957]: DEBUG oslo_concurrency.lockutils [None req-017d3b28-7c60-435c-8f0d-fb9db5de2a93 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Releasing lock "refresh_cache-84cd5619-4ac9-41ef-9368-a023ad9ae66d" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1101.426688] env[61957]: DEBUG nova.objects.instance [None req-017d3b28-7c60-435c-8f0d-fb9db5de2a93 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Lazy-loading 'migration_context' on Instance uuid 84cd5619-4ac9-41ef-9368-a023ad9ae66d {{(pid=61957) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1101.546843] env[61957]: DEBUG oslo_vmware.api [None req-0cc2a3a7-4047-410e-a46d-b5ab6f7151bc tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': task-1278105, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1101.668843] env[61957]: DEBUG nova.compute.manager [req-245902eb-e92a-444d-9080-bab524edfde7 req-e3c86782-e04a-41b6-aeea-82c4e7545543 service nova] [instance: 1a0327c2-5671-4970-9db7-c7cc912d8678] Received event network-vif-deleted-e84fe58e-f595-4b03-9e8a-88238d003168 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1101.669110] env[61957]: INFO nova.compute.manager [req-245902eb-e92a-444d-9080-bab524edfde7 req-e3c86782-e04a-41b6-aeea-82c4e7545543 service nova] [instance: 1a0327c2-5671-4970-9db7-c7cc912d8678] Neutron deleted interface e84fe58e-f595-4b03-9e8a-88238d003168; detaching it from the instance and deleting it from the info cache [ 1101.669364] env[61957]: DEBUG nova.network.neutron [req-245902eb-e92a-444d-9080-bab524edfde7 req-e3c86782-e04a-41b6-aeea-82c4e7545543 service nova] [instance: 1a0327c2-5671-4970-9db7-c7cc912d8678] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1101.931440] env[61957]: DEBUG nova.objects.base [None req-017d3b28-7c60-435c-8f0d-fb9db5de2a93 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Object Instance<84cd5619-4ac9-41ef-9368-a023ad9ae66d> lazy-loaded attributes: info_cache,migration_context {{(pid=61957) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1101.932345] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32719420-e9b0-4206-918d-c17b8bdf1a58 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.953130] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7bc61caa-0e58-4d71-8594-d476aa312ad7 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.958529] env[61957]: DEBUG oslo_vmware.api [None req-017d3b28-7c60-435c-8f0d-fb9db5de2a93 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Waiting for the task: (returnval){ [ 1101.958529] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52be5131-fd0c-b594-d021-89b31e389f6f" [ 1101.958529] env[61957]: _type = "Task" [ 1101.958529] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1101.966257] env[61957]: DEBUG oslo_vmware.api [None req-017d3b28-7c60-435c-8f0d-fb9db5de2a93 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52be5131-fd0c-b594-d021-89b31e389f6f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1102.045259] env[61957]: DEBUG oslo_vmware.api [None req-0cc2a3a7-4047-410e-a46d-b5ab6f7151bc tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': task-1278105, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1102.149130] env[61957]: DEBUG nova.network.neutron [-] [instance: 1a0327c2-5671-4970-9db7-c7cc912d8678] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1102.172427] env[61957]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ef74d1d1-9eee-4d9f-9944-cfbced4c803e {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.181706] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-672f1af9-7122-4ffe-afcd-c8c22877eba9 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.207113] env[61957]: DEBUG nova.compute.manager [req-245902eb-e92a-444d-9080-bab524edfde7 req-e3c86782-e04a-41b6-aeea-82c4e7545543 service nova] [instance: 1a0327c2-5671-4970-9db7-c7cc912d8678] Detach interface failed, port_id=e84fe58e-f595-4b03-9e8a-88238d003168, reason: Instance 1a0327c2-5671-4970-9db7-c7cc912d8678 could not be found. {{(pid=61957) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1102.468736] env[61957]: DEBUG oslo_vmware.api [None req-017d3b28-7c60-435c-8f0d-fb9db5de2a93 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52be5131-fd0c-b594-d021-89b31e389f6f, 'name': SearchDatastore_Task, 'duration_secs': 0.006489} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1102.468988] env[61957]: DEBUG oslo_concurrency.lockutils [None req-017d3b28-7c60-435c-8f0d-fb9db5de2a93 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1102.469242] env[61957]: DEBUG oslo_concurrency.lockutils [None req-017d3b28-7c60-435c-8f0d-fb9db5de2a93 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1102.546405] env[61957]: DEBUG oslo_vmware.api [None req-0cc2a3a7-4047-410e-a46d-b5ab6f7151bc tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': task-1278105, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1102.651598] env[61957]: INFO nova.compute.manager [-] [instance: 1a0327c2-5671-4970-9db7-c7cc912d8678] Took 1.24 seconds to deallocate network for instance. [ 1103.047794] env[61957]: DEBUG oslo_vmware.api [None req-0cc2a3a7-4047-410e-a46d-b5ab6f7151bc tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': task-1278105, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1103.050681] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f8084df-2455-4d58-81b1-d750251dacd8 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.056780] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffad856b-17a2-47f7-840d-7f47f86dfb77 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.089847] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8375f100-f1b4-46f2-a607-dfcf7f508f59 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.096798] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-356f1c77-3a3c-4231-8c11-9ea520f5f765 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.109347] env[61957]: DEBUG nova.compute.provider_tree [None req-017d3b28-7c60-435c-8f0d-fb9db5de2a93 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Updating inventory in ProviderTree for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1103.158432] env[61957]: DEBUG oslo_concurrency.lockutils [None req-77f64c0f-751d-4c0e-b32e-91391a5af017 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1103.547777] env[61957]: DEBUG oslo_vmware.api [None req-0cc2a3a7-4047-410e-a46d-b5ab6f7151bc tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': task-1278105, 'name': ReconfigVM_Task} progress is 18%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1103.630158] env[61957]: ERROR nova.scheduler.client.report [None req-017d3b28-7c60-435c-8f0d-fb9db5de2a93 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [req-0aab9001-8b75-48c3-92f9-ffedf247f19a] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 0ceb6c9e-61c6-496d-8579-9d32627e96da. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-0aab9001-8b75-48c3-92f9-ffedf247f19a"}]} [ 1103.647358] env[61957]: DEBUG nova.scheduler.client.report [None req-017d3b28-7c60-435c-8f0d-fb9db5de2a93 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Refreshing inventories for resource provider 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 1103.661505] env[61957]: DEBUG nova.scheduler.client.report [None req-017d3b28-7c60-435c-8f0d-fb9db5de2a93 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Updating ProviderTree inventory for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 1103.661807] env[61957]: DEBUG nova.compute.provider_tree [None req-017d3b28-7c60-435c-8f0d-fb9db5de2a93 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Updating inventory in ProviderTree for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1103.672153] env[61957]: DEBUG nova.scheduler.client.report [None req-017d3b28-7c60-435c-8f0d-fb9db5de2a93 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Refreshing aggregate associations for resource provider 0ceb6c9e-61c6-496d-8579-9d32627e96da, aggregates: None {{(pid=61957) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 1103.688583] env[61957]: DEBUG nova.scheduler.client.report [None req-017d3b28-7c60-435c-8f0d-fb9db5de2a93 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Refreshing trait associations for resource provider 0ceb6c9e-61c6-496d-8579-9d32627e96da, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO,HW_ARCH_X86_64,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=61957) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 1103.764749] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b265c367-e0e1-4828-ac00-f0c7dc955988 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.772216] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03b32413-ee80-4365-8917-7f577398a1ba {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.801084] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f73986d-6865-49a7-976d-68afe34320e2 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.807926] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e49c8297-4bd7-4bb8-8189-ec3404ed3b3d {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.821962] env[61957]: DEBUG nova.compute.provider_tree [None req-017d3b28-7c60-435c-8f0d-fb9db5de2a93 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Updating inventory in ProviderTree for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1104.048965] env[61957]: DEBUG oslo_vmware.api [None req-0cc2a3a7-4047-410e-a46d-b5ab6f7151bc tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': task-1278105, 'name': ReconfigVM_Task, 'duration_secs': 5.77821} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1104.049311] env[61957]: DEBUG oslo_concurrency.lockutils [None req-0cc2a3a7-4047-410e-a46d-b5ab6f7151bc tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Releasing lock "4625d1dc-a621-4a33-9537-9e8cce96c914" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1104.049570] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-0cc2a3a7-4047-410e-a46d-b5ab6f7151bc tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 4625d1dc-a621-4a33-9537-9e8cce96c914] Reconfigured VM to detach interface {{(pid=61957) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1984}} [ 1104.090219] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-ac6e1710-ee9b-4755-89eb-bd5d0557f455 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] [instance: 039a6b01-a745-478c-9ff7-c8c37afe2ce4] Volume attach. Driver type: vmdk {{(pid=61957) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1104.090482] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-ac6e1710-ee9b-4755-89eb-bd5d0557f455 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] [instance: 039a6b01-a745-478c-9ff7-c8c37afe2ce4] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-274609', 'volume_id': '95dd15cb-394c-419f-bbc2-aa25920febe7', 'name': 'volume-95dd15cb-394c-419f-bbc2-aa25920febe7', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '039a6b01-a745-478c-9ff7-c8c37afe2ce4', 'attached_at': '', 'detached_at': '', 'volume_id': '95dd15cb-394c-419f-bbc2-aa25920febe7', 'serial': '95dd15cb-394c-419f-bbc2-aa25920febe7'} {{(pid=61957) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1104.091477] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d5deaae-cfab-42ea-8a48-daa51c98bdae {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.108903] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5102474-e7b3-4648-a6ca-fe83c857684d {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.138728] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-ac6e1710-ee9b-4755-89eb-bd5d0557f455 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] [instance: 039a6b01-a745-478c-9ff7-c8c37afe2ce4] Reconfiguring VM instance instance-00000062 to attach disk [datastore2] volume-95dd15cb-394c-419f-bbc2-aa25920febe7/volume-95dd15cb-394c-419f-bbc2-aa25920febe7.vmdk or device None with type thin {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1104.140758] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-628e5eb3-9327-48c9-a55f-19669a5b6d50 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.161662] env[61957]: DEBUG oslo_vmware.api [None req-ac6e1710-ee9b-4755-89eb-bd5d0557f455 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Waiting for the task: (returnval){ [ 1104.161662] env[61957]: value = "task-1278111" [ 1104.161662] env[61957]: _type = "Task" [ 1104.161662] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1104.171198] env[61957]: DEBUG oslo_vmware.api [None req-ac6e1710-ee9b-4755-89eb-bd5d0557f455 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Task: {'id': task-1278111, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1104.354999] env[61957]: DEBUG nova.scheduler.client.report [None req-017d3b28-7c60-435c-8f0d-fb9db5de2a93 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Updated inventory for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da with generation 130 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 1104.355288] env[61957]: DEBUG nova.compute.provider_tree [None req-017d3b28-7c60-435c-8f0d-fb9db5de2a93 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Updating resource provider 0ceb6c9e-61c6-496d-8579-9d32627e96da generation from 130 to 131 during operation: update_inventory {{(pid=61957) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1104.355510] env[61957]: DEBUG nova.compute.provider_tree [None req-017d3b28-7c60-435c-8f0d-fb9db5de2a93 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Updating inventory in ProviderTree for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1104.671422] env[61957]: DEBUG oslo_vmware.api [None req-ac6e1710-ee9b-4755-89eb-bd5d0557f455 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Task: {'id': task-1278111, 'name': ReconfigVM_Task, 'duration_secs': 0.354188} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1104.671785] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-ac6e1710-ee9b-4755-89eb-bd5d0557f455 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] [instance: 039a6b01-a745-478c-9ff7-c8c37afe2ce4] Reconfigured VM instance instance-00000062 to attach disk [datastore2] volume-95dd15cb-394c-419f-bbc2-aa25920febe7/volume-95dd15cb-394c-419f-bbc2-aa25920febe7.vmdk or device None with type thin {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1104.676673] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f2e6c369-6421-4934-bcc6-bb6ed25fa8de {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.692167] env[61957]: DEBUG oslo_vmware.api [None req-ac6e1710-ee9b-4755-89eb-bd5d0557f455 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Waiting for the task: (returnval){ [ 1104.692167] env[61957]: value = "task-1278112" [ 1104.692167] env[61957]: _type = "Task" [ 1104.692167] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1104.699684] env[61957]: DEBUG oslo_vmware.api [None req-ac6e1710-ee9b-4755-89eb-bd5d0557f455 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Task: {'id': task-1278112, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1105.204423] env[61957]: DEBUG oslo_vmware.api [None req-ac6e1710-ee9b-4755-89eb-bd5d0557f455 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Task: {'id': task-1278112, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1105.366240] env[61957]: DEBUG oslo_concurrency.lockutils [None req-017d3b28-7c60-435c-8f0d-fb9db5de2a93 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.897s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1105.369179] env[61957]: DEBUG oslo_concurrency.lockutils [None req-77f64c0f-751d-4c0e-b32e-91391a5af017 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.211s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1105.369451] env[61957]: DEBUG nova.objects.instance [None req-77f64c0f-751d-4c0e-b32e-91391a5af017 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Lazy-loading 'resources' on Instance uuid 1a0327c2-5671-4970-9db7-c7cc912d8678 {{(pid=61957) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1105.387463] env[61957]: DEBUG oslo_concurrency.lockutils [None req-0cc2a3a7-4047-410e-a46d-b5ab6f7151bc tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Acquiring lock "refresh_cache-4625d1dc-a621-4a33-9537-9e8cce96c914" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1105.388707] env[61957]: DEBUG oslo_concurrency.lockutils [None req-0cc2a3a7-4047-410e-a46d-b5ab6f7151bc tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Acquired lock "refresh_cache-4625d1dc-a621-4a33-9537-9e8cce96c914" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1105.388707] env[61957]: DEBUG nova.network.neutron [None req-0cc2a3a7-4047-410e-a46d-b5ab6f7151bc tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 4625d1dc-a621-4a33-9537-9e8cce96c914] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1105.703549] env[61957]: DEBUG oslo_vmware.api [None req-ac6e1710-ee9b-4755-89eb-bd5d0557f455 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Task: {'id': task-1278112, 'name': ReconfigVM_Task, 'duration_secs': 0.928701} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1105.703549] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-ac6e1710-ee9b-4755-89eb-bd5d0557f455 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] [instance: 039a6b01-a745-478c-9ff7-c8c37afe2ce4] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-274609', 'volume_id': '95dd15cb-394c-419f-bbc2-aa25920febe7', 'name': 'volume-95dd15cb-394c-419f-bbc2-aa25920febe7', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '039a6b01-a745-478c-9ff7-c8c37afe2ce4', 'attached_at': '', 'detached_at': '', 'volume_id': '95dd15cb-394c-419f-bbc2-aa25920febe7', 'serial': '95dd15cb-394c-419f-bbc2-aa25920febe7'} {{(pid=61957) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1105.936384] env[61957]: INFO nova.scheduler.client.report [None req-017d3b28-7c60-435c-8f0d-fb9db5de2a93 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Deleted allocation for migration 26522749-2bb5-49e0-acaa-5e0766e6c136 [ 1105.975759] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd4b4028-e51a-4dc7-90c9-af9b00c20673 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.984762] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df9c2e3a-afaa-4db7-921b-a145924a61b8 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.017396] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31941c79-9f7d-41c6-a3a5-85415489a800 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.027326] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61722f11-9f67-482f-8284-573b0be1f154 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.042849] env[61957]: DEBUG nova.compute.provider_tree [None req-77f64c0f-751d-4c0e-b32e-91391a5af017 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1106.075707] env[61957]: DEBUG nova.compute.manager [req-a7181231-474b-4451-9f7b-96dbcc322050 req-1a678c64-9159-4782-84c2-3133111afa65 service nova] [instance: 4625d1dc-a621-4a33-9537-9e8cce96c914] Received event network-changed-8365f86e-8b14-47de-a6f4-297107be492e {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1106.075997] env[61957]: DEBUG nova.compute.manager [req-a7181231-474b-4451-9f7b-96dbcc322050 req-1a678c64-9159-4782-84c2-3133111afa65 service nova] [instance: 4625d1dc-a621-4a33-9537-9e8cce96c914] Refreshing instance network info cache due to event network-changed-8365f86e-8b14-47de-a6f4-297107be492e. {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1106.076122] env[61957]: DEBUG oslo_concurrency.lockutils [req-a7181231-474b-4451-9f7b-96dbcc322050 req-1a678c64-9159-4782-84c2-3133111afa65 service nova] Acquiring lock "refresh_cache-4625d1dc-a621-4a33-9537-9e8cce96c914" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1106.152025] env[61957]: INFO nova.network.neutron [None req-0cc2a3a7-4047-410e-a46d-b5ab6f7151bc tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 4625d1dc-a621-4a33-9537-9e8cce96c914] Port b3373cf5-1370-4b75-b08c-f76c081ec78f from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 1106.153028] env[61957]: DEBUG nova.network.neutron [None req-0cc2a3a7-4047-410e-a46d-b5ab6f7151bc tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 4625d1dc-a621-4a33-9537-9e8cce96c914] Updating instance_info_cache with network_info: [{"id": "8365f86e-8b14-47de-a6f4-297107be492e", "address": "fa:16:3e:e8:59:9d", "network": {"id": "364cabcb-046d-4f91-b1ee-5dd3adfae6d7", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1432396967-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8713aa35bcb24b86ad0b58ca9fc991ba", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b7a73c01-1bb9-4612-a1a7-16d71b732e81", "external-id": "nsx-vlan-transportzone-711", "segmentation_id": 711, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8365f86e-8b", "ovs_interfaceid": "8365f86e-8b14-47de-a6f4-297107be492e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1106.443622] env[61957]: DEBUG oslo_concurrency.lockutils [None req-017d3b28-7c60-435c-8f0d-fb9db5de2a93 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Lock "84cd5619-4ac9-41ef-9368-a023ad9ae66d" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 7.460s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1106.545737] env[61957]: DEBUG nova.scheduler.client.report [None req-77f64c0f-751d-4c0e-b32e-91391a5af017 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1106.654784] env[61957]: DEBUG oslo_concurrency.lockutils [None req-0cc2a3a7-4047-410e-a46d-b5ab6f7151bc tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Releasing lock "refresh_cache-4625d1dc-a621-4a33-9537-9e8cce96c914" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1106.656932] env[61957]: DEBUG oslo_concurrency.lockutils [req-a7181231-474b-4451-9f7b-96dbcc322050 req-1a678c64-9159-4782-84c2-3133111afa65 service nova] Acquired lock "refresh_cache-4625d1dc-a621-4a33-9537-9e8cce96c914" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1106.657157] env[61957]: DEBUG nova.network.neutron [req-a7181231-474b-4451-9f7b-96dbcc322050 req-1a678c64-9159-4782-84c2-3133111afa65 service nova] [instance: 4625d1dc-a621-4a33-9537-9e8cce96c914] Refreshing network info cache for port 8365f86e-8b14-47de-a6f4-297107be492e {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1106.740234] env[61957]: DEBUG nova.objects.instance [None req-ac6e1710-ee9b-4755-89eb-bd5d0557f455 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Lazy-loading 'flavor' on Instance uuid 039a6b01-a745-478c-9ff7-c8c37afe2ce4 {{(pid=61957) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1106.742084] env[61957]: DEBUG oslo_concurrency.lockutils [None req-7729b744-1547-433f-8fa6-259cdc94c8f9 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Acquiring lock "interface-b82af1e9-294a-4ba9-bcad-73b2a2aca86d-b3373cf5-1370-4b75-b08c-f76c081ec78f" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1106.742319] env[61957]: DEBUG oslo_concurrency.lockutils [None req-7729b744-1547-433f-8fa6-259cdc94c8f9 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Lock "interface-b82af1e9-294a-4ba9-bcad-73b2a2aca86d-b3373cf5-1370-4b75-b08c-f76c081ec78f" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1106.742654] env[61957]: DEBUG nova.objects.instance [None req-7729b744-1547-433f-8fa6-259cdc94c8f9 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Lazy-loading 'flavor' on Instance uuid b82af1e9-294a-4ba9-bcad-73b2a2aca86d {{(pid=61957) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1107.050693] env[61957]: DEBUG oslo_concurrency.lockutils [None req-77f64c0f-751d-4c0e-b32e-91391a5af017 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.681s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1107.071266] env[61957]: INFO nova.scheduler.client.report [None req-77f64c0f-751d-4c0e-b32e-91391a5af017 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Deleted allocations for instance 1a0327c2-5671-4970-9db7-c7cc912d8678 [ 1107.163020] env[61957]: DEBUG oslo_concurrency.lockutils [None req-0cc2a3a7-4047-410e-a46d-b5ab6f7151bc tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Lock "interface-4625d1dc-a621-4a33-9537-9e8cce96c914-b3373cf5-1370-4b75-b08c-f76c081ec78f" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 9.693s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1107.249713] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ac6e1710-ee9b-4755-89eb-bd5d0557f455 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Lock "039a6b01-a745-478c-9ff7-c8c37afe2ce4" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.757s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1107.419447] env[61957]: DEBUG nova.objects.instance [None req-7729b744-1547-433f-8fa6-259cdc94c8f9 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Lazy-loading 'pci_requests' on Instance uuid b82af1e9-294a-4ba9-bcad-73b2a2aca86d {{(pid=61957) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1107.435377] env[61957]: DEBUG nova.network.neutron [req-a7181231-474b-4451-9f7b-96dbcc322050 req-1a678c64-9159-4782-84c2-3133111afa65 service nova] [instance: 4625d1dc-a621-4a33-9537-9e8cce96c914] Updated VIF entry in instance network info cache for port 8365f86e-8b14-47de-a6f4-297107be492e. {{(pid=61957) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1107.435741] env[61957]: DEBUG nova.network.neutron [req-a7181231-474b-4451-9f7b-96dbcc322050 req-1a678c64-9159-4782-84c2-3133111afa65 service nova] [instance: 4625d1dc-a621-4a33-9537-9e8cce96c914] Updating instance_info_cache with network_info: [{"id": "8365f86e-8b14-47de-a6f4-297107be492e", "address": "fa:16:3e:e8:59:9d", "network": {"id": "364cabcb-046d-4f91-b1ee-5dd3adfae6d7", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1432396967-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8713aa35bcb24b86ad0b58ca9fc991ba", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b7a73c01-1bb9-4612-a1a7-16d71b732e81", "external-id": "nsx-vlan-transportzone-711", "segmentation_id": 711, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8365f86e-8b", "ovs_interfaceid": "8365f86e-8b14-47de-a6f4-297107be492e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1107.505475] env[61957]: DEBUG oslo_concurrency.lockutils [None req-017d3b28-7c60-435c-8f0d-fb9db5de2a93 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Acquiring lock "84cd5619-4ac9-41ef-9368-a023ad9ae66d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1107.505753] env[61957]: DEBUG oslo_concurrency.lockutils [None req-017d3b28-7c60-435c-8f0d-fb9db5de2a93 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Lock "84cd5619-4ac9-41ef-9368-a023ad9ae66d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1107.505979] env[61957]: DEBUG oslo_concurrency.lockutils [None req-017d3b28-7c60-435c-8f0d-fb9db5de2a93 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Acquiring lock "84cd5619-4ac9-41ef-9368-a023ad9ae66d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1107.506194] env[61957]: DEBUG oslo_concurrency.lockutils [None req-017d3b28-7c60-435c-8f0d-fb9db5de2a93 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Lock "84cd5619-4ac9-41ef-9368-a023ad9ae66d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1107.506465] env[61957]: DEBUG oslo_concurrency.lockutils [None req-017d3b28-7c60-435c-8f0d-fb9db5de2a93 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Lock "84cd5619-4ac9-41ef-9368-a023ad9ae66d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1107.509720] env[61957]: INFO nova.compute.manager [None req-017d3b28-7c60-435c-8f0d-fb9db5de2a93 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 84cd5619-4ac9-41ef-9368-a023ad9ae66d] Terminating instance [ 1107.511639] env[61957]: DEBUG nova.compute.manager [None req-017d3b28-7c60-435c-8f0d-fb9db5de2a93 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 84cd5619-4ac9-41ef-9368-a023ad9ae66d] Start destroying the instance on the hypervisor. {{(pid=61957) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1107.511639] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-017d3b28-7c60-435c-8f0d-fb9db5de2a93 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 84cd5619-4ac9-41ef-9368-a023ad9ae66d] Destroying instance {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1107.512397] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03b36a77-c88f-488e-a6fc-fbf19af5b7dd {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.522113] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-017d3b28-7c60-435c-8f0d-fb9db5de2a93 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 84cd5619-4ac9-41ef-9368-a023ad9ae66d] Powering off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1107.522113] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-325e872c-bbd7-4233-b5fb-dc11208531b4 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.527077] env[61957]: DEBUG oslo_vmware.api [None req-017d3b28-7c60-435c-8f0d-fb9db5de2a93 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Waiting for the task: (returnval){ [ 1107.527077] env[61957]: value = "task-1278113" [ 1107.527077] env[61957]: _type = "Task" [ 1107.527077] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1107.536606] env[61957]: DEBUG oslo_vmware.api [None req-017d3b28-7c60-435c-8f0d-fb9db5de2a93 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1278113, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1107.580207] env[61957]: DEBUG oslo_concurrency.lockutils [None req-77f64c0f-751d-4c0e-b32e-91391a5af017 tempest-ServersTestJSON-1338650625 tempest-ServersTestJSON-1338650625-project-member] Lock "1a0327c2-5671-4970-9db7-c7cc912d8678" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.791s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1107.923123] env[61957]: DEBUG nova.objects.base [None req-7729b744-1547-433f-8fa6-259cdc94c8f9 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=61957) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1107.923475] env[61957]: DEBUG nova.network.neutron [None req-7729b744-1547-433f-8fa6-259cdc94c8f9 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: b82af1e9-294a-4ba9-bcad-73b2a2aca86d] allocate_for_instance() {{(pid=61957) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1107.938874] env[61957]: DEBUG oslo_concurrency.lockutils [req-a7181231-474b-4451-9f7b-96dbcc322050 req-1a678c64-9159-4782-84c2-3133111afa65 service nova] Releasing lock "refresh_cache-4625d1dc-a621-4a33-9537-9e8cce96c914" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1107.999902] env[61957]: DEBUG nova.policy [None req-7729b744-1547-433f-8fa6-259cdc94c8f9 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6fd1c19dc3a44212ada44445e0919106', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8713aa35bcb24b86ad0b58ca9fc991ba', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61957) authorize /opt/stack/nova/nova/policy.py:203}} [ 1108.037052] env[61957]: DEBUG oslo_vmware.api [None req-017d3b28-7c60-435c-8f0d-fb9db5de2a93 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1278113, 'name': PowerOffVM_Task, 'duration_secs': 0.171478} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1108.037347] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-017d3b28-7c60-435c-8f0d-fb9db5de2a93 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 84cd5619-4ac9-41ef-9368-a023ad9ae66d] Powered off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1108.037519] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-017d3b28-7c60-435c-8f0d-fb9db5de2a93 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 84cd5619-4ac9-41ef-9368-a023ad9ae66d] Unregistering the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1108.037805] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-67bb5ff8-57df-4f4c-83a5-07a3703b9fa8 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.072228] env[61957]: DEBUG oslo_concurrency.lockutils [None req-4248be44-37e2-49b2-ac48-7340725065b5 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Acquiring lock "039a6b01-a745-478c-9ff7-c8c37afe2ce4" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1108.072516] env[61957]: DEBUG oslo_concurrency.lockutils [None req-4248be44-37e2-49b2-ac48-7340725065b5 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Lock "039a6b01-a745-478c-9ff7-c8c37afe2ce4" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1108.101122] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-017d3b28-7c60-435c-8f0d-fb9db5de2a93 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 84cd5619-4ac9-41ef-9368-a023ad9ae66d] Unregistered the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1108.101379] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-017d3b28-7c60-435c-8f0d-fb9db5de2a93 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 84cd5619-4ac9-41ef-9368-a023ad9ae66d] Deleting contents of the VM from datastore datastore2 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1108.101569] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-017d3b28-7c60-435c-8f0d-fb9db5de2a93 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Deleting the datastore file [datastore2] 84cd5619-4ac9-41ef-9368-a023ad9ae66d {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1108.101831] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-df5d94d0-f5a1-4e77-a118-9d6574e92254 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.108675] env[61957]: DEBUG oslo_vmware.api [None req-017d3b28-7c60-435c-8f0d-fb9db5de2a93 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Waiting for the task: (returnval){ [ 1108.108675] env[61957]: value = "task-1278115" [ 1108.108675] env[61957]: _type = "Task" [ 1108.108675] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1108.118216] env[61957]: DEBUG oslo_vmware.api [None req-017d3b28-7c60-435c-8f0d-fb9db5de2a93 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1278115, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1108.119395] env[61957]: DEBUG nova.compute.manager [req-72483a27-b66b-4481-99d3-28fdb91aff43 req-cb5f44d0-4b58-4bbf-a129-7c24c01cd55a service nova] [instance: b82af1e9-294a-4ba9-bcad-73b2a2aca86d] Received event network-changed-58a3f865-3fc6-462a-b685-fb32e01f6c5e {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1108.119579] env[61957]: DEBUG nova.compute.manager [req-72483a27-b66b-4481-99d3-28fdb91aff43 req-cb5f44d0-4b58-4bbf-a129-7c24c01cd55a service nova] [instance: b82af1e9-294a-4ba9-bcad-73b2a2aca86d] Refreshing instance network info cache due to event network-changed-58a3f865-3fc6-462a-b685-fb32e01f6c5e. {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1108.119794] env[61957]: DEBUG oslo_concurrency.lockutils [req-72483a27-b66b-4481-99d3-28fdb91aff43 req-cb5f44d0-4b58-4bbf-a129-7c24c01cd55a service nova] Acquiring lock "refresh_cache-b82af1e9-294a-4ba9-bcad-73b2a2aca86d" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1108.119939] env[61957]: DEBUG oslo_concurrency.lockutils [req-72483a27-b66b-4481-99d3-28fdb91aff43 req-cb5f44d0-4b58-4bbf-a129-7c24c01cd55a service nova] Acquired lock "refresh_cache-b82af1e9-294a-4ba9-bcad-73b2a2aca86d" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1108.120115] env[61957]: DEBUG nova.network.neutron [req-72483a27-b66b-4481-99d3-28fdb91aff43 req-cb5f44d0-4b58-4bbf-a129-7c24c01cd55a service nova] [instance: b82af1e9-294a-4ba9-bcad-73b2a2aca86d] Refreshing network info cache for port 58a3f865-3fc6-462a-b685-fb32e01f6c5e {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1108.575798] env[61957]: INFO nova.compute.manager [None req-4248be44-37e2-49b2-ac48-7340725065b5 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] [instance: 039a6b01-a745-478c-9ff7-c8c37afe2ce4] Detaching volume d29b3bbe-886e-4384-9591-178f5fe64eea [ 1108.615321] env[61957]: INFO nova.virt.block_device [None req-4248be44-37e2-49b2-ac48-7340725065b5 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] [instance: 039a6b01-a745-478c-9ff7-c8c37afe2ce4] Attempting to driver detach volume d29b3bbe-886e-4384-9591-178f5fe64eea from mountpoint /dev/sdb [ 1108.615557] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-4248be44-37e2-49b2-ac48-7340725065b5 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] [instance: 039a6b01-a745-478c-9ff7-c8c37afe2ce4] Volume detach. Driver type: vmdk {{(pid=61957) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1108.615739] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-4248be44-37e2-49b2-ac48-7340725065b5 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] [instance: 039a6b01-a745-478c-9ff7-c8c37afe2ce4] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-274608', 'volume_id': 'd29b3bbe-886e-4384-9591-178f5fe64eea', 'name': 'volume-d29b3bbe-886e-4384-9591-178f5fe64eea', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '039a6b01-a745-478c-9ff7-c8c37afe2ce4', 'attached_at': '', 'detached_at': '', 'volume_id': 'd29b3bbe-886e-4384-9591-178f5fe64eea', 'serial': 'd29b3bbe-886e-4384-9591-178f5fe64eea'} {{(pid=61957) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1108.616559] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03741e87-0a20-470a-89d3-384e28e5ad27 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.626122] env[61957]: DEBUG oslo_vmware.api [None req-017d3b28-7c60-435c-8f0d-fb9db5de2a93 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Task: {'id': task-1278115, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.187304} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1108.645674] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-017d3b28-7c60-435c-8f0d-fb9db5de2a93 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Deleted the datastore file {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1108.645889] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-017d3b28-7c60-435c-8f0d-fb9db5de2a93 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 84cd5619-4ac9-41ef-9368-a023ad9ae66d] Deleted contents of the VM from datastore datastore2 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1108.646079] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-017d3b28-7c60-435c-8f0d-fb9db5de2a93 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 84cd5619-4ac9-41ef-9368-a023ad9ae66d] Instance destroyed {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1108.646260] env[61957]: INFO nova.compute.manager [None req-017d3b28-7c60-435c-8f0d-fb9db5de2a93 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] [instance: 84cd5619-4ac9-41ef-9368-a023ad9ae66d] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1108.646509] env[61957]: DEBUG oslo.service.loopingcall [None req-017d3b28-7c60-435c-8f0d-fb9db5de2a93 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1108.647426] env[61957]: DEBUG nova.compute.manager [-] [instance: 84cd5619-4ac9-41ef-9368-a023ad9ae66d] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1108.647529] env[61957]: DEBUG nova.network.neutron [-] [instance: 84cd5619-4ac9-41ef-9368-a023ad9ae66d] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1108.649546] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92460606-a3ef-4a31-bee0-99d2a28b4160 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.656927] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba4f152d-88f2-4082-9c70-d65736b80990 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.678851] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0400fee2-c620-4b36-b8eb-496163a0c4d1 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.693054] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-4248be44-37e2-49b2-ac48-7340725065b5 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] The volume has not been displaced from its original location: [datastore1] volume-d29b3bbe-886e-4384-9591-178f5fe64eea/volume-d29b3bbe-886e-4384-9591-178f5fe64eea.vmdk. No consolidation needed. {{(pid=61957) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1108.698225] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-4248be44-37e2-49b2-ac48-7340725065b5 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] [instance: 039a6b01-a745-478c-9ff7-c8c37afe2ce4] Reconfiguring VM instance instance-00000062 to detach disk 2001 {{(pid=61957) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1108.700614] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8e00c5a8-ad4f-46c9-b7a0-859a6c6cb9ac {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.718817] env[61957]: DEBUG oslo_vmware.api [None req-4248be44-37e2-49b2-ac48-7340725065b5 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Waiting for the task: (returnval){ [ 1108.718817] env[61957]: value = "task-1278116" [ 1108.718817] env[61957]: _type = "Task" [ 1108.718817] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1108.728080] env[61957]: DEBUG oslo_vmware.api [None req-4248be44-37e2-49b2-ac48-7340725065b5 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Task: {'id': task-1278116, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1108.999203] env[61957]: DEBUG nova.network.neutron [req-72483a27-b66b-4481-99d3-28fdb91aff43 req-cb5f44d0-4b58-4bbf-a129-7c24c01cd55a service nova] [instance: b82af1e9-294a-4ba9-bcad-73b2a2aca86d] Updated VIF entry in instance network info cache for port 58a3f865-3fc6-462a-b685-fb32e01f6c5e. {{(pid=61957) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1108.999612] env[61957]: DEBUG nova.network.neutron [req-72483a27-b66b-4481-99d3-28fdb91aff43 req-cb5f44d0-4b58-4bbf-a129-7c24c01cd55a service nova] [instance: b82af1e9-294a-4ba9-bcad-73b2a2aca86d] Updating instance_info_cache with network_info: [{"id": "58a3f865-3fc6-462a-b685-fb32e01f6c5e", "address": "fa:16:3e:ae:67:c0", "network": {"id": "364cabcb-046d-4f91-b1ee-5dd3adfae6d7", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1432396967-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.232", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8713aa35bcb24b86ad0b58ca9fc991ba", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b7a73c01-1bb9-4612-a1a7-16d71b732e81", "external-id": "nsx-vlan-transportzone-711", "segmentation_id": 711, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap58a3f865-3f", "ovs_interfaceid": "58a3f865-3fc6-462a-b685-fb32e01f6c5e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1109.071535] env[61957]: DEBUG nova.compute.manager [req-761a1e5d-8ee6-457b-a9b8-1616250bbbd9 req-650953d4-c0b6-44cc-930a-1be891822975 service nova] [instance: 84cd5619-4ac9-41ef-9368-a023ad9ae66d] Received event network-vif-deleted-b0541069-72e9-4384-b4f4-ccf5bc206dab {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1109.071796] env[61957]: INFO nova.compute.manager [req-761a1e5d-8ee6-457b-a9b8-1616250bbbd9 req-650953d4-c0b6-44cc-930a-1be891822975 service nova] [instance: 84cd5619-4ac9-41ef-9368-a023ad9ae66d] Neutron deleted interface b0541069-72e9-4384-b4f4-ccf5bc206dab; detaching it from the instance and deleting it from the info cache [ 1109.071977] env[61957]: DEBUG nova.network.neutron [req-761a1e5d-8ee6-457b-a9b8-1616250bbbd9 req-650953d4-c0b6-44cc-930a-1be891822975 service nova] [instance: 84cd5619-4ac9-41ef-9368-a023ad9ae66d] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1109.230628] env[61957]: DEBUG oslo_vmware.api [None req-4248be44-37e2-49b2-ac48-7340725065b5 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Task: {'id': task-1278116, 'name': ReconfigVM_Task, 'duration_secs': 0.491421} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1109.230628] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-4248be44-37e2-49b2-ac48-7340725065b5 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] [instance: 039a6b01-a745-478c-9ff7-c8c37afe2ce4] Reconfigured VM instance instance-00000062 to detach disk 2001 {{(pid=61957) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1109.233681] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8a8c2908-bf5d-44e8-9fef-8df3f226f263 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.248852] env[61957]: DEBUG oslo_vmware.api [None req-4248be44-37e2-49b2-ac48-7340725065b5 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Waiting for the task: (returnval){ [ 1109.248852] env[61957]: value = "task-1278117" [ 1109.248852] env[61957]: _type = "Task" [ 1109.248852] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1109.257405] env[61957]: DEBUG oslo_vmware.api [None req-4248be44-37e2-49b2-ac48-7340725065b5 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Task: {'id': task-1278117, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1109.500961] env[61957]: DEBUG nova.network.neutron [None req-7729b744-1547-433f-8fa6-259cdc94c8f9 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: b82af1e9-294a-4ba9-bcad-73b2a2aca86d] Successfully updated port: b3373cf5-1370-4b75-b08c-f76c081ec78f {{(pid=61957) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1109.505799] env[61957]: DEBUG oslo_concurrency.lockutils [req-72483a27-b66b-4481-99d3-28fdb91aff43 req-cb5f44d0-4b58-4bbf-a129-7c24c01cd55a service nova] Releasing lock "refresh_cache-b82af1e9-294a-4ba9-bcad-73b2a2aca86d" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1109.542286] env[61957]: DEBUG nova.network.neutron [-] [instance: 84cd5619-4ac9-41ef-9368-a023ad9ae66d] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1109.574165] env[61957]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e06c290d-07ee-4b45-a252-648a354ede3e {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.582901] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8ef5c26-3a8f-418b-9c11-450f2a29a5bd {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.612168] env[61957]: DEBUG nova.compute.manager [req-761a1e5d-8ee6-457b-a9b8-1616250bbbd9 req-650953d4-c0b6-44cc-930a-1be891822975 service nova] [instance: 84cd5619-4ac9-41ef-9368-a023ad9ae66d] Detach interface failed, port_id=b0541069-72e9-4384-b4f4-ccf5bc206dab, reason: Instance 84cd5619-4ac9-41ef-9368-a023ad9ae66d could not be found. {{(pid=61957) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1109.758332] env[61957]: DEBUG oslo_vmware.api [None req-4248be44-37e2-49b2-ac48-7340725065b5 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Task: {'id': task-1278117, 'name': ReconfigVM_Task, 'duration_secs': 0.140656} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1109.758669] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-4248be44-37e2-49b2-ac48-7340725065b5 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] [instance: 039a6b01-a745-478c-9ff7-c8c37afe2ce4] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-274608', 'volume_id': 'd29b3bbe-886e-4384-9591-178f5fe64eea', 'name': 'volume-d29b3bbe-886e-4384-9591-178f5fe64eea', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '039a6b01-a745-478c-9ff7-c8c37afe2ce4', 'attached_at': '', 'detached_at': '', 'volume_id': 'd29b3bbe-886e-4384-9591-178f5fe64eea', 'serial': 'd29b3bbe-886e-4384-9591-178f5fe64eea'} {{(pid=61957) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1110.004059] env[61957]: DEBUG oslo_concurrency.lockutils [None req-7729b744-1547-433f-8fa6-259cdc94c8f9 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Acquiring lock "refresh_cache-b82af1e9-294a-4ba9-bcad-73b2a2aca86d" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1110.004388] env[61957]: DEBUG oslo_concurrency.lockutils [None req-7729b744-1547-433f-8fa6-259cdc94c8f9 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Acquired lock "refresh_cache-b82af1e9-294a-4ba9-bcad-73b2a2aca86d" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1110.004388] env[61957]: DEBUG nova.network.neutron [None req-7729b744-1547-433f-8fa6-259cdc94c8f9 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: b82af1e9-294a-4ba9-bcad-73b2a2aca86d] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1110.045067] env[61957]: INFO nova.compute.manager [-] [instance: 84cd5619-4ac9-41ef-9368-a023ad9ae66d] Took 1.40 seconds to deallocate network for instance. [ 1110.301812] env[61957]: DEBUG nova.objects.instance [None req-4248be44-37e2-49b2-ac48-7340725065b5 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Lazy-loading 'flavor' on Instance uuid 039a6b01-a745-478c-9ff7-c8c37afe2ce4 {{(pid=61957) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1110.547294] env[61957]: WARNING nova.network.neutron [None req-7729b744-1547-433f-8fa6-259cdc94c8f9 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: b82af1e9-294a-4ba9-bcad-73b2a2aca86d] 364cabcb-046d-4f91-b1ee-5dd3adfae6d7 already exists in list: networks containing: ['364cabcb-046d-4f91-b1ee-5dd3adfae6d7']. ignoring it [ 1110.551548] env[61957]: DEBUG oslo_concurrency.lockutils [None req-017d3b28-7c60-435c-8f0d-fb9db5de2a93 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1110.551811] env[61957]: DEBUG oslo_concurrency.lockutils [None req-017d3b28-7c60-435c-8f0d-fb9db5de2a93 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1110.552022] env[61957]: DEBUG oslo_concurrency.lockutils [None req-017d3b28-7c60-435c-8f0d-fb9db5de2a93 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1110.574772] env[61957]: INFO nova.scheduler.client.report [None req-017d3b28-7c60-435c-8f0d-fb9db5de2a93 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Deleted allocations for instance 84cd5619-4ac9-41ef-9368-a023ad9ae66d [ 1110.827288] env[61957]: DEBUG nova.network.neutron [None req-7729b744-1547-433f-8fa6-259cdc94c8f9 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: b82af1e9-294a-4ba9-bcad-73b2a2aca86d] Updating instance_info_cache with network_info: [{"id": "58a3f865-3fc6-462a-b685-fb32e01f6c5e", "address": "fa:16:3e:ae:67:c0", "network": {"id": "364cabcb-046d-4f91-b1ee-5dd3adfae6d7", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1432396967-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.232", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8713aa35bcb24b86ad0b58ca9fc991ba", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b7a73c01-1bb9-4612-a1a7-16d71b732e81", "external-id": "nsx-vlan-transportzone-711", "segmentation_id": 711, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap58a3f865-3f", "ovs_interfaceid": "58a3f865-3fc6-462a-b685-fb32e01f6c5e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "b3373cf5-1370-4b75-b08c-f76c081ec78f", "address": "fa:16:3e:13:49:46", "network": {"id": "364cabcb-046d-4f91-b1ee-5dd3adfae6d7", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1432396967-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8713aa35bcb24b86ad0b58ca9fc991ba", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b7a73c01-1bb9-4612-a1a7-16d71b732e81", "external-id": "nsx-vlan-transportzone-711", "segmentation_id": 711, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb3373cf5-13", "ovs_interfaceid": "b3373cf5-1370-4b75-b08c-f76c081ec78f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1111.085106] env[61957]: DEBUG oslo_concurrency.lockutils [None req-017d3b28-7c60-435c-8f0d-fb9db5de2a93 tempest-DeleteServersTestJSON-696031498 tempest-DeleteServersTestJSON-696031498-project-member] Lock "84cd5619-4ac9-41ef-9368-a023ad9ae66d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 3.579s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1111.095708] env[61957]: DEBUG nova.compute.manager [req-e3709c85-801f-48b0-93a3-f747bb875c89 req-6406ef9d-5533-4d05-96d9-e7e5bc40cbac service nova] [instance: b82af1e9-294a-4ba9-bcad-73b2a2aca86d] Received event network-vif-plugged-b3373cf5-1370-4b75-b08c-f76c081ec78f {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1111.096502] env[61957]: DEBUG oslo_concurrency.lockutils [req-e3709c85-801f-48b0-93a3-f747bb875c89 req-6406ef9d-5533-4d05-96d9-e7e5bc40cbac service nova] Acquiring lock "b82af1e9-294a-4ba9-bcad-73b2a2aca86d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1111.096740] env[61957]: DEBUG oslo_concurrency.lockutils [req-e3709c85-801f-48b0-93a3-f747bb875c89 req-6406ef9d-5533-4d05-96d9-e7e5bc40cbac service nova] Lock "b82af1e9-294a-4ba9-bcad-73b2a2aca86d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1111.096740] env[61957]: DEBUG oslo_concurrency.lockutils [req-e3709c85-801f-48b0-93a3-f747bb875c89 req-6406ef9d-5533-4d05-96d9-e7e5bc40cbac service nova] Lock "b82af1e9-294a-4ba9-bcad-73b2a2aca86d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1111.096926] env[61957]: DEBUG nova.compute.manager [req-e3709c85-801f-48b0-93a3-f747bb875c89 req-6406ef9d-5533-4d05-96d9-e7e5bc40cbac service nova] [instance: b82af1e9-294a-4ba9-bcad-73b2a2aca86d] No waiting events found dispatching network-vif-plugged-b3373cf5-1370-4b75-b08c-f76c081ec78f {{(pid=61957) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1111.097110] env[61957]: WARNING nova.compute.manager [req-e3709c85-801f-48b0-93a3-f747bb875c89 req-6406ef9d-5533-4d05-96d9-e7e5bc40cbac service nova] [instance: b82af1e9-294a-4ba9-bcad-73b2a2aca86d] Received unexpected event network-vif-plugged-b3373cf5-1370-4b75-b08c-f76c081ec78f for instance with vm_state active and task_state None. [ 1111.097280] env[61957]: DEBUG nova.compute.manager [req-e3709c85-801f-48b0-93a3-f747bb875c89 req-6406ef9d-5533-4d05-96d9-e7e5bc40cbac service nova] [instance: b82af1e9-294a-4ba9-bcad-73b2a2aca86d] Received event network-changed-b3373cf5-1370-4b75-b08c-f76c081ec78f {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1111.097438] env[61957]: DEBUG nova.compute.manager [req-e3709c85-801f-48b0-93a3-f747bb875c89 req-6406ef9d-5533-4d05-96d9-e7e5bc40cbac service nova] [instance: b82af1e9-294a-4ba9-bcad-73b2a2aca86d] Refreshing instance network info cache due to event network-changed-b3373cf5-1370-4b75-b08c-f76c081ec78f. {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1111.097608] env[61957]: DEBUG oslo_concurrency.lockutils [req-e3709c85-801f-48b0-93a3-f747bb875c89 req-6406ef9d-5533-4d05-96d9-e7e5bc40cbac service nova] Acquiring lock "refresh_cache-b82af1e9-294a-4ba9-bcad-73b2a2aca86d" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1111.308977] env[61957]: DEBUG oslo_concurrency.lockutils [None req-4248be44-37e2-49b2-ac48-7340725065b5 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Lock "039a6b01-a745-478c-9ff7-c8c37afe2ce4" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.236s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1111.328981] env[61957]: DEBUG oslo_concurrency.lockutils [None req-7729b744-1547-433f-8fa6-259cdc94c8f9 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Releasing lock "refresh_cache-b82af1e9-294a-4ba9-bcad-73b2a2aca86d" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1111.329668] env[61957]: DEBUG oslo_concurrency.lockutils [None req-7729b744-1547-433f-8fa6-259cdc94c8f9 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Acquiring lock "b82af1e9-294a-4ba9-bcad-73b2a2aca86d" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1111.329878] env[61957]: DEBUG oslo_concurrency.lockutils [None req-7729b744-1547-433f-8fa6-259cdc94c8f9 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Acquired lock "b82af1e9-294a-4ba9-bcad-73b2a2aca86d" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1111.330250] env[61957]: DEBUG oslo_concurrency.lockutils [None req-f475a664-751d-43c4-9965-71dc57136499 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Acquiring lock "039a6b01-a745-478c-9ff7-c8c37afe2ce4" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1111.330452] env[61957]: DEBUG oslo_concurrency.lockutils [None req-f475a664-751d-43c4-9965-71dc57136499 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Lock "039a6b01-a745-478c-9ff7-c8c37afe2ce4" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1111.331637] env[61957]: DEBUG oslo_concurrency.lockutils [req-e3709c85-801f-48b0-93a3-f747bb875c89 req-6406ef9d-5533-4d05-96d9-e7e5bc40cbac service nova] Acquired lock "refresh_cache-b82af1e9-294a-4ba9-bcad-73b2a2aca86d" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1111.331828] env[61957]: DEBUG nova.network.neutron [req-e3709c85-801f-48b0-93a3-f747bb875c89 req-6406ef9d-5533-4d05-96d9-e7e5bc40cbac service nova] [instance: b82af1e9-294a-4ba9-bcad-73b2a2aca86d] Refreshing network info cache for port b3373cf5-1370-4b75-b08c-f76c081ec78f {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1111.335216] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfd4802d-a0e8-49b9-92ac-fd44847a63ea {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.338772] env[61957]: INFO nova.compute.manager [None req-f475a664-751d-43c4-9965-71dc57136499 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] [instance: 039a6b01-a745-478c-9ff7-c8c37afe2ce4] Detaching volume 95dd15cb-394c-419f-bbc2-aa25920febe7 [ 1111.354730] env[61957]: DEBUG nova.virt.hardware [None req-7729b744-1547-433f-8fa6-259cdc94c8f9 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T17:22:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1111.354954] env[61957]: DEBUG nova.virt.hardware [None req-7729b744-1547-433f-8fa6-259cdc94c8f9 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Flavor limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1111.355134] env[61957]: DEBUG nova.virt.hardware [None req-7729b744-1547-433f-8fa6-259cdc94c8f9 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Image limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1111.355323] env[61957]: DEBUG nova.virt.hardware [None req-7729b744-1547-433f-8fa6-259cdc94c8f9 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Flavor pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1111.355472] env[61957]: DEBUG nova.virt.hardware [None req-7729b744-1547-433f-8fa6-259cdc94c8f9 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Image pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1111.355621] env[61957]: DEBUG nova.virt.hardware [None req-7729b744-1547-433f-8fa6-259cdc94c8f9 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1111.355821] env[61957]: DEBUG nova.virt.hardware [None req-7729b744-1547-433f-8fa6-259cdc94c8f9 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1111.355983] env[61957]: DEBUG nova.virt.hardware [None req-7729b744-1547-433f-8fa6-259cdc94c8f9 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1111.356168] env[61957]: DEBUG nova.virt.hardware [None req-7729b744-1547-433f-8fa6-259cdc94c8f9 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Got 1 possible topologies {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1111.356334] env[61957]: DEBUG nova.virt.hardware [None req-7729b744-1547-433f-8fa6-259cdc94c8f9 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1111.356506] env[61957]: DEBUG nova.virt.hardware [None req-7729b744-1547-433f-8fa6-259cdc94c8f9 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1111.362634] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-7729b744-1547-433f-8fa6-259cdc94c8f9 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: b82af1e9-294a-4ba9-bcad-73b2a2aca86d] Reconfiguring VM to attach interface {{(pid=61957) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1929}} [ 1111.365213] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f363c55b-4afb-421c-b946-e5af9f09ed63 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.382192] env[61957]: DEBUG oslo_vmware.api [None req-7729b744-1547-433f-8fa6-259cdc94c8f9 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Waiting for the task: (returnval){ [ 1111.382192] env[61957]: value = "task-1278118" [ 1111.382192] env[61957]: _type = "Task" [ 1111.382192] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1111.389975] env[61957]: DEBUG oslo_vmware.api [None req-7729b744-1547-433f-8fa6-259cdc94c8f9 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': task-1278118, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1111.402177] env[61957]: INFO nova.virt.block_device [None req-f475a664-751d-43c4-9965-71dc57136499 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] [instance: 039a6b01-a745-478c-9ff7-c8c37afe2ce4] Attempting to driver detach volume 95dd15cb-394c-419f-bbc2-aa25920febe7 from mountpoint /dev/sdc [ 1111.402426] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-f475a664-751d-43c4-9965-71dc57136499 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] [instance: 039a6b01-a745-478c-9ff7-c8c37afe2ce4] Volume detach. Driver type: vmdk {{(pid=61957) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1111.402637] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-f475a664-751d-43c4-9965-71dc57136499 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] [instance: 039a6b01-a745-478c-9ff7-c8c37afe2ce4] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-274609', 'volume_id': '95dd15cb-394c-419f-bbc2-aa25920febe7', 'name': 'volume-95dd15cb-394c-419f-bbc2-aa25920febe7', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '039a6b01-a745-478c-9ff7-c8c37afe2ce4', 'attached_at': '', 'detached_at': '', 'volume_id': '95dd15cb-394c-419f-bbc2-aa25920febe7', 'serial': '95dd15cb-394c-419f-bbc2-aa25920febe7'} {{(pid=61957) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1111.403561] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1367ad9f-3391-4e75-ae30-1feac7ef3f37 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.425022] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5589343-ea30-4385-9127-d4e496f0d77a {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.432308] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c14af143-9136-4540-98c3-c1b316f8a8bb {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.451448] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba6902c2-e118-4415-abb0-19190de38d7c {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.465683] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-f475a664-751d-43c4-9965-71dc57136499 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] The volume has not been displaced from its original location: [datastore2] volume-95dd15cb-394c-419f-bbc2-aa25920febe7/volume-95dd15cb-394c-419f-bbc2-aa25920febe7.vmdk. No consolidation needed. {{(pid=61957) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1111.470853] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-f475a664-751d-43c4-9965-71dc57136499 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] [instance: 039a6b01-a745-478c-9ff7-c8c37afe2ce4] Reconfiguring VM instance instance-00000062 to detach disk 2002 {{(pid=61957) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1111.471146] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-542ac441-3c15-4e21-b07a-e769440af4c1 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.488084] env[61957]: DEBUG oslo_vmware.api [None req-f475a664-751d-43c4-9965-71dc57136499 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Waiting for the task: (returnval){ [ 1111.488084] env[61957]: value = "task-1278119" [ 1111.488084] env[61957]: _type = "Task" [ 1111.488084] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1111.495742] env[61957]: DEBUG oslo_vmware.api [None req-f475a664-751d-43c4-9965-71dc57136499 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Task: {'id': task-1278119, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1111.895525] env[61957]: DEBUG oslo_vmware.api [None req-7729b744-1547-433f-8fa6-259cdc94c8f9 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': task-1278118, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1111.999238] env[61957]: DEBUG oslo_vmware.api [None req-f475a664-751d-43c4-9965-71dc57136499 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Task: {'id': task-1278119, 'name': ReconfigVM_Task, 'duration_secs': 0.196801} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1111.999518] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-f475a664-751d-43c4-9965-71dc57136499 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] [instance: 039a6b01-a745-478c-9ff7-c8c37afe2ce4] Reconfigured VM instance instance-00000062 to detach disk 2002 {{(pid=61957) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1112.004091] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-006ea751-4477-4387-86ef-3514d6962ecc {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.018628] env[61957]: DEBUG oslo_vmware.api [None req-f475a664-751d-43c4-9965-71dc57136499 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Waiting for the task: (returnval){ [ 1112.018628] env[61957]: value = "task-1278121" [ 1112.018628] env[61957]: _type = "Task" [ 1112.018628] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1112.026033] env[61957]: DEBUG oslo_vmware.api [None req-f475a664-751d-43c4-9965-71dc57136499 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Task: {'id': task-1278121, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1112.063730] env[61957]: DEBUG nova.network.neutron [req-e3709c85-801f-48b0-93a3-f747bb875c89 req-6406ef9d-5533-4d05-96d9-e7e5bc40cbac service nova] [instance: b82af1e9-294a-4ba9-bcad-73b2a2aca86d] Updated VIF entry in instance network info cache for port b3373cf5-1370-4b75-b08c-f76c081ec78f. {{(pid=61957) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1112.064215] env[61957]: DEBUG nova.network.neutron [req-e3709c85-801f-48b0-93a3-f747bb875c89 req-6406ef9d-5533-4d05-96d9-e7e5bc40cbac service nova] [instance: b82af1e9-294a-4ba9-bcad-73b2a2aca86d] Updating instance_info_cache with network_info: [{"id": "58a3f865-3fc6-462a-b685-fb32e01f6c5e", "address": "fa:16:3e:ae:67:c0", "network": {"id": "364cabcb-046d-4f91-b1ee-5dd3adfae6d7", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1432396967-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.232", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8713aa35bcb24b86ad0b58ca9fc991ba", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b7a73c01-1bb9-4612-a1a7-16d71b732e81", "external-id": "nsx-vlan-transportzone-711", "segmentation_id": 711, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap58a3f865-3f", "ovs_interfaceid": "58a3f865-3fc6-462a-b685-fb32e01f6c5e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "b3373cf5-1370-4b75-b08c-f76c081ec78f", "address": "fa:16:3e:13:49:46", "network": {"id": "364cabcb-046d-4f91-b1ee-5dd3adfae6d7", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1432396967-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8713aa35bcb24b86ad0b58ca9fc991ba", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b7a73c01-1bb9-4612-a1a7-16d71b732e81", "external-id": "nsx-vlan-transportzone-711", "segmentation_id": 711, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb3373cf5-13", "ovs_interfaceid": "b3373cf5-1370-4b75-b08c-f76c081ec78f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1112.392399] env[61957]: DEBUG oslo_vmware.api [None req-7729b744-1547-433f-8fa6-259cdc94c8f9 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': task-1278118, 'name': ReconfigVM_Task, 'duration_secs': 0.551774} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1112.392929] env[61957]: DEBUG oslo_concurrency.lockutils [None req-7729b744-1547-433f-8fa6-259cdc94c8f9 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Releasing lock "b82af1e9-294a-4ba9-bcad-73b2a2aca86d" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1112.393168] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-7729b744-1547-433f-8fa6-259cdc94c8f9 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: b82af1e9-294a-4ba9-bcad-73b2a2aca86d] Reconfigured VM to attach interface {{(pid=61957) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1943}} [ 1112.528646] env[61957]: DEBUG oslo_vmware.api [None req-f475a664-751d-43c4-9965-71dc57136499 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Task: {'id': task-1278121, 'name': ReconfigVM_Task, 'duration_secs': 0.13766} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1112.528972] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-f475a664-751d-43c4-9965-71dc57136499 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] [instance: 039a6b01-a745-478c-9ff7-c8c37afe2ce4] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-274609', 'volume_id': '95dd15cb-394c-419f-bbc2-aa25920febe7', 'name': 'volume-95dd15cb-394c-419f-bbc2-aa25920febe7', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '039a6b01-a745-478c-9ff7-c8c37afe2ce4', 'attached_at': '', 'detached_at': '', 'volume_id': '95dd15cb-394c-419f-bbc2-aa25920febe7', 'serial': '95dd15cb-394c-419f-bbc2-aa25920febe7'} {{(pid=61957) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1112.567168] env[61957]: DEBUG oslo_concurrency.lockutils [req-e3709c85-801f-48b0-93a3-f747bb875c89 req-6406ef9d-5533-4d05-96d9-e7e5bc40cbac service nova] Releasing lock "refresh_cache-b82af1e9-294a-4ba9-bcad-73b2a2aca86d" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1112.897406] env[61957]: DEBUG oslo_concurrency.lockutils [None req-7729b744-1547-433f-8fa6-259cdc94c8f9 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Lock "interface-b82af1e9-294a-4ba9-bcad-73b2a2aca86d-b3373cf5-1370-4b75-b08c-f76c081ec78f" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 6.155s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1113.069986] env[61957]: DEBUG nova.objects.instance [None req-f475a664-751d-43c4-9965-71dc57136499 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Lazy-loading 'flavor' on Instance uuid 039a6b01-a745-478c-9ff7-c8c37afe2ce4 {{(pid=61957) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1114.079804] env[61957]: DEBUG oslo_concurrency.lockutils [None req-f475a664-751d-43c4-9965-71dc57136499 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Lock "039a6b01-a745-478c-9ff7-c8c37afe2ce4" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 2.749s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1114.412279] env[61957]: DEBUG oslo_concurrency.lockutils [None req-6e5a656a-a04a-46b8-b899-efd16f82d0e7 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Acquiring lock "interface-b82af1e9-294a-4ba9-bcad-73b2a2aca86d-b3373cf5-1370-4b75-b08c-f76c081ec78f" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1114.412549] env[61957]: DEBUG oslo_concurrency.lockutils [None req-6e5a656a-a04a-46b8-b899-efd16f82d0e7 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Lock "interface-b82af1e9-294a-4ba9-bcad-73b2a2aca86d-b3373cf5-1370-4b75-b08c-f76c081ec78f" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.001s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1114.659495] env[61957]: DEBUG oslo_concurrency.lockutils [None req-fbe237b1-b3df-4682-bd1d-c5ac7da593d8 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Acquiring lock "039a6b01-a745-478c-9ff7-c8c37afe2ce4" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1114.659735] env[61957]: DEBUG oslo_concurrency.lockutils [None req-fbe237b1-b3df-4682-bd1d-c5ac7da593d8 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Lock "039a6b01-a745-478c-9ff7-c8c37afe2ce4" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1114.659920] env[61957]: DEBUG oslo_concurrency.lockutils [None req-fbe237b1-b3df-4682-bd1d-c5ac7da593d8 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Acquiring lock "039a6b01-a745-478c-9ff7-c8c37afe2ce4-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1114.660137] env[61957]: DEBUG oslo_concurrency.lockutils [None req-fbe237b1-b3df-4682-bd1d-c5ac7da593d8 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Lock "039a6b01-a745-478c-9ff7-c8c37afe2ce4-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1114.660318] env[61957]: DEBUG oslo_concurrency.lockutils [None req-fbe237b1-b3df-4682-bd1d-c5ac7da593d8 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Lock "039a6b01-a745-478c-9ff7-c8c37afe2ce4-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1114.662596] env[61957]: INFO nova.compute.manager [None req-fbe237b1-b3df-4682-bd1d-c5ac7da593d8 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] [instance: 039a6b01-a745-478c-9ff7-c8c37afe2ce4] Terminating instance [ 1114.664427] env[61957]: DEBUG nova.compute.manager [None req-fbe237b1-b3df-4682-bd1d-c5ac7da593d8 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] [instance: 039a6b01-a745-478c-9ff7-c8c37afe2ce4] Start destroying the instance on the hypervisor. {{(pid=61957) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1114.664628] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-fbe237b1-b3df-4682-bd1d-c5ac7da593d8 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] [instance: 039a6b01-a745-478c-9ff7-c8c37afe2ce4] Destroying instance {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1114.665487] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7575d241-1618-4417-a14a-cd7b55224d92 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.674428] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-fbe237b1-b3df-4682-bd1d-c5ac7da593d8 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] [instance: 039a6b01-a745-478c-9ff7-c8c37afe2ce4] Powering off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1114.674651] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4ab9158c-b8da-4c20-bed2-ea71c577924e {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.680883] env[61957]: DEBUG oslo_vmware.api [None req-fbe237b1-b3df-4682-bd1d-c5ac7da593d8 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Waiting for the task: (returnval){ [ 1114.680883] env[61957]: value = "task-1278122" [ 1114.680883] env[61957]: _type = "Task" [ 1114.680883] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1114.688610] env[61957]: DEBUG oslo_vmware.api [None req-fbe237b1-b3df-4682-bd1d-c5ac7da593d8 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Task: {'id': task-1278122, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1114.815572] env[61957]: DEBUG oslo_service.periodic_task [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61957) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1114.815880] env[61957]: DEBUG oslo_service.periodic_task [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61957) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1114.816044] env[61957]: DEBUG nova.compute.manager [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Starting heal instance info cache {{(pid=61957) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 1114.914832] env[61957]: DEBUG oslo_concurrency.lockutils [None req-6e5a656a-a04a-46b8-b899-efd16f82d0e7 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Acquiring lock "b82af1e9-294a-4ba9-bcad-73b2a2aca86d" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1114.915083] env[61957]: DEBUG oslo_concurrency.lockutils [None req-6e5a656a-a04a-46b8-b899-efd16f82d0e7 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Acquired lock "b82af1e9-294a-4ba9-bcad-73b2a2aca86d" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1114.915967] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4916fe9-bcdc-47cd-82b7-7d5ff7bb823f {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.933441] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b088327-2496-4c87-bded-a5991a22a6e9 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.960521] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-6e5a656a-a04a-46b8-b899-efd16f82d0e7 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: b82af1e9-294a-4ba9-bcad-73b2a2aca86d] Reconfiguring VM to detach interface {{(pid=61957) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1974}} [ 1114.961213] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-13ffb49a-cf88-4469-8570-14051cec451f {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.979839] env[61957]: DEBUG oslo_vmware.api [None req-6e5a656a-a04a-46b8-b899-efd16f82d0e7 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Waiting for the task: (returnval){ [ 1114.979839] env[61957]: value = "task-1278123" [ 1114.979839] env[61957]: _type = "Task" [ 1114.979839] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1114.989600] env[61957]: DEBUG oslo_vmware.api [None req-6e5a656a-a04a-46b8-b899-efd16f82d0e7 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': task-1278123, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1115.190781] env[61957]: DEBUG oslo_vmware.api [None req-fbe237b1-b3df-4682-bd1d-c5ac7da593d8 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Task: {'id': task-1278122, 'name': PowerOffVM_Task, 'duration_secs': 0.178443} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1115.191129] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-fbe237b1-b3df-4682-bd1d-c5ac7da593d8 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] [instance: 039a6b01-a745-478c-9ff7-c8c37afe2ce4] Powered off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1115.191241] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-fbe237b1-b3df-4682-bd1d-c5ac7da593d8 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] [instance: 039a6b01-a745-478c-9ff7-c8c37afe2ce4] Unregistering the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1115.191480] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-38e1a35d-3ee7-41fc-a7b1-3d83e6752d11 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.258787] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-fbe237b1-b3df-4682-bd1d-c5ac7da593d8 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] [instance: 039a6b01-a745-478c-9ff7-c8c37afe2ce4] Unregistered the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1115.259077] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-fbe237b1-b3df-4682-bd1d-c5ac7da593d8 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] [instance: 039a6b01-a745-478c-9ff7-c8c37afe2ce4] Deleting contents of the VM from datastore datastore1 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1115.259293] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-fbe237b1-b3df-4682-bd1d-c5ac7da593d8 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Deleting the datastore file [datastore1] 039a6b01-a745-478c-9ff7-c8c37afe2ce4 {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1115.259496] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6a9963c7-8799-41aa-bffc-cb241b68acdb {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.266534] env[61957]: DEBUG oslo_vmware.api [None req-fbe237b1-b3df-4682-bd1d-c5ac7da593d8 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Waiting for the task: (returnval){ [ 1115.266534] env[61957]: value = "task-1278125" [ 1115.266534] env[61957]: _type = "Task" [ 1115.266534] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1115.273867] env[61957]: DEBUG oslo_vmware.api [None req-fbe237b1-b3df-4682-bd1d-c5ac7da593d8 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Task: {'id': task-1278125, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1115.489603] env[61957]: DEBUG oslo_vmware.api [None req-6e5a656a-a04a-46b8-b899-efd16f82d0e7 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': task-1278123, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1115.777453] env[61957]: DEBUG oslo_vmware.api [None req-fbe237b1-b3df-4682-bd1d-c5ac7da593d8 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Task: {'id': task-1278125, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.173795} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1115.778087] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-fbe237b1-b3df-4682-bd1d-c5ac7da593d8 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Deleted the datastore file {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1115.778348] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-fbe237b1-b3df-4682-bd1d-c5ac7da593d8 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] [instance: 039a6b01-a745-478c-9ff7-c8c37afe2ce4] Deleted contents of the VM from datastore datastore1 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1115.778544] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-fbe237b1-b3df-4682-bd1d-c5ac7da593d8 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] [instance: 039a6b01-a745-478c-9ff7-c8c37afe2ce4] Instance destroyed {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1115.778972] env[61957]: INFO nova.compute.manager [None req-fbe237b1-b3df-4682-bd1d-c5ac7da593d8 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] [instance: 039a6b01-a745-478c-9ff7-c8c37afe2ce4] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1115.779287] env[61957]: DEBUG oslo.service.loopingcall [None req-fbe237b1-b3df-4682-bd1d-c5ac7da593d8 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1115.779575] env[61957]: DEBUG nova.compute.manager [-] [instance: 039a6b01-a745-478c-9ff7-c8c37afe2ce4] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1115.779609] env[61957]: DEBUG nova.network.neutron [-] [instance: 039a6b01-a745-478c-9ff7-c8c37afe2ce4] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1115.992616] env[61957]: DEBUG oslo_vmware.api [None req-6e5a656a-a04a-46b8-b899-efd16f82d0e7 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': task-1278123, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1116.300719] env[61957]: DEBUG nova.compute.manager [req-100903f7-717f-436a-a87a-049e077b5fd7 req-d060fd5e-9c14-4c06-8d84-51faff48530e service nova] [instance: 039a6b01-a745-478c-9ff7-c8c37afe2ce4] Received event network-vif-deleted-e49b8de4-47b9-400b-a0b0-9e0a0e68dc84 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1116.301364] env[61957]: INFO nova.compute.manager [req-100903f7-717f-436a-a87a-049e077b5fd7 req-d060fd5e-9c14-4c06-8d84-51faff48530e service nova] [instance: 039a6b01-a745-478c-9ff7-c8c37afe2ce4] Neutron deleted interface e49b8de4-47b9-400b-a0b0-9e0a0e68dc84; detaching it from the instance and deleting it from the info cache [ 1116.301603] env[61957]: DEBUG nova.network.neutron [req-100903f7-717f-436a-a87a-049e077b5fd7 req-d060fd5e-9c14-4c06-8d84-51faff48530e service nova] [instance: 039a6b01-a745-478c-9ff7-c8c37afe2ce4] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1116.492366] env[61957]: DEBUG oslo_vmware.api [None req-6e5a656a-a04a-46b8-b899-efd16f82d0e7 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': task-1278123, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1116.770032] env[61957]: DEBUG nova.network.neutron [-] [instance: 039a6b01-a745-478c-9ff7-c8c37afe2ce4] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1116.804345] env[61957]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-56d4b2d9-ac30-4792-9552-5e0c3b29062b {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.814446] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55b7636a-ec79-42b9-b088-93d8aa58ba49 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.839552] env[61957]: DEBUG nova.compute.manager [req-100903f7-717f-436a-a87a-049e077b5fd7 req-d060fd5e-9c14-4c06-8d84-51faff48530e service nova] [instance: 039a6b01-a745-478c-9ff7-c8c37afe2ce4] Detach interface failed, port_id=e49b8de4-47b9-400b-a0b0-9e0a0e68dc84, reason: Instance 039a6b01-a745-478c-9ff7-c8c37afe2ce4 could not be found. {{(pid=61957) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1116.991079] env[61957]: DEBUG oslo_vmware.api [None req-6e5a656a-a04a-46b8-b899-efd16f82d0e7 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': task-1278123, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1117.271966] env[61957]: INFO nova.compute.manager [-] [instance: 039a6b01-a745-478c-9ff7-c8c37afe2ce4] Took 1.49 seconds to deallocate network for instance. [ 1117.491512] env[61957]: DEBUG oslo_vmware.api [None req-6e5a656a-a04a-46b8-b899-efd16f82d0e7 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': task-1278123, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1117.779260] env[61957]: DEBUG oslo_concurrency.lockutils [None req-fbe237b1-b3df-4682-bd1d-c5ac7da593d8 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1117.779545] env[61957]: DEBUG oslo_concurrency.lockutils [None req-fbe237b1-b3df-4682-bd1d-c5ac7da593d8 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1117.779766] env[61957]: DEBUG nova.objects.instance [None req-fbe237b1-b3df-4682-bd1d-c5ac7da593d8 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Lazy-loading 'resources' on Instance uuid 039a6b01-a745-478c-9ff7-c8c37afe2ce4 {{(pid=61957) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1117.830056] env[61957]: DEBUG nova.compute.manager [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Didn't find any instances for network info cache update. {{(pid=61957) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10012}} [ 1117.830296] env[61957]: DEBUG oslo_service.periodic_task [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61957) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1117.830456] env[61957]: DEBUG oslo_service.periodic_task [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61957) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1117.830629] env[61957]: DEBUG oslo_service.periodic_task [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61957) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1117.830793] env[61957]: DEBUG oslo_service.periodic_task [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61957) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1117.830946] env[61957]: DEBUG oslo_service.periodic_task [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61957) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1117.831043] env[61957]: DEBUG oslo_service.periodic_task [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61957) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1117.831178] env[61957]: DEBUG nova.compute.manager [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61957) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 1117.831319] env[61957]: DEBUG oslo_service.periodic_task [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61957) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1117.992642] env[61957]: DEBUG oslo_vmware.api [None req-6e5a656a-a04a-46b8-b899-efd16f82d0e7 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': task-1278123, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.334403] env[61957]: DEBUG oslo_concurrency.lockutils [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1118.335252] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b462645-9163-485a-b516-34f7972dbb03 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.342433] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0b19743-0945-45b5-ade1-47bbfd826426 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.371746] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f06d7139-619f-4aa4-812b-0e56d9cf6a4c {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.378458] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1384dddc-64f1-4a35-a29b-0011941a9567 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.393061] env[61957]: DEBUG nova.compute.provider_tree [None req-fbe237b1-b3df-4682-bd1d-c5ac7da593d8 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1118.494295] env[61957]: DEBUG oslo_vmware.api [None req-6e5a656a-a04a-46b8-b899-efd16f82d0e7 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': task-1278123, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.894037] env[61957]: DEBUG nova.scheduler.client.report [None req-fbe237b1-b3df-4682-bd1d-c5ac7da593d8 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1118.994349] env[61957]: DEBUG oslo_vmware.api [None req-6e5a656a-a04a-46b8-b899-efd16f82d0e7 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': task-1278123, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1119.400231] env[61957]: DEBUG oslo_concurrency.lockutils [None req-fbe237b1-b3df-4682-bd1d-c5ac7da593d8 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.620s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1119.404184] env[61957]: DEBUG oslo_concurrency.lockutils [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 1.068s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1119.404184] env[61957]: DEBUG oslo_concurrency.lockutils [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.001s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1119.404310] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61957) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1119.405128] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd852297-c40f-48bb-8d1c-a5253e3da41e {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.413759] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76988880-1767-41b7-b146-d59e378e0191 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.429188] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f77aa8c7-c9ac-4f6f-b362-73bd6a25157f {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.432387] env[61957]: INFO nova.scheduler.client.report [None req-fbe237b1-b3df-4682-bd1d-c5ac7da593d8 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Deleted allocations for instance 039a6b01-a745-478c-9ff7-c8c37afe2ce4 [ 1119.438500] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4aba2c3-92c2-4f81-bd1e-3c4a77d11ec3 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.467426] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181087MB free_disk=142GB free_vcpus=48 pci_devices=None {{(pid=61957) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1119.467594] env[61957]: DEBUG oslo_concurrency.lockutils [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1119.467796] env[61957]: DEBUG oslo_concurrency.lockutils [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1119.473064] env[61957]: DEBUG oslo_concurrency.lockutils [None req-66a206ff-5fcb-4812-b9a0-d20a1311acc5 tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] Acquiring lock "363e48d7-8ed1-45d5-9443-d0b715952fca" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1119.474043] env[61957]: DEBUG oslo_concurrency.lockutils [None req-66a206ff-5fcb-4812-b9a0-d20a1311acc5 tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] Lock "363e48d7-8ed1-45d5-9443-d0b715952fca" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1119.495921] env[61957]: DEBUG oslo_vmware.api [None req-6e5a656a-a04a-46b8-b899-efd16f82d0e7 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': task-1278123, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1119.942298] env[61957]: DEBUG oslo_concurrency.lockutils [None req-fbe237b1-b3df-4682-bd1d-c5ac7da593d8 tempest-AttachVolumeTestJSON-442630357 tempest-AttachVolumeTestJSON-442630357-project-member] Lock "039a6b01-a745-478c-9ff7-c8c37afe2ce4" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.282s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1119.975558] env[61957]: DEBUG nova.compute.manager [None req-66a206ff-5fcb-4812-b9a0-d20a1311acc5 tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] [instance: 363e48d7-8ed1-45d5-9443-d0b715952fca] Starting instance... {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1119.996769] env[61957]: DEBUG oslo_vmware.api [None req-6e5a656a-a04a-46b8-b899-efd16f82d0e7 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': task-1278123, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1120.496927] env[61957]: DEBUG oslo_vmware.api [None req-6e5a656a-a04a-46b8-b899-efd16f82d0e7 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': task-1278123, 'name': ReconfigVM_Task} progress is 18%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1120.498524] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Instance 4625d1dc-a621-4a33-9537-9e8cce96c914 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61957) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1120.498663] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Instance b82af1e9-294a-4ba9-bcad-73b2a2aca86d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61957) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1120.500731] env[61957]: DEBUG oslo_concurrency.lockutils [None req-66a206ff-5fcb-4812-b9a0-d20a1311acc5 tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1120.996288] env[61957]: DEBUG oslo_vmware.api [None req-6e5a656a-a04a-46b8-b899-efd16f82d0e7 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': task-1278123, 'name': ReconfigVM_Task, 'duration_secs': 5.754574} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1120.996584] env[61957]: DEBUG oslo_concurrency.lockutils [None req-6e5a656a-a04a-46b8-b899-efd16f82d0e7 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Releasing lock "b82af1e9-294a-4ba9-bcad-73b2a2aca86d" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1120.996807] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-6e5a656a-a04a-46b8-b899-efd16f82d0e7 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: b82af1e9-294a-4ba9-bcad-73b2a2aca86d] Reconfigured VM to detach interface {{(pid=61957) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1984}} [ 1121.002605] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Instance 363e48d7-8ed1-45d5-9443-d0b715952fca has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61957) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 1121.003184] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Total usable vcpus: 48, total allocated vcpus: 2 {{(pid=61957) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1121.003184] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=896MB phys_disk=200GB used_disk=2GB total_vcpus=48 used_vcpus=2 pci_stats=[] {{(pid=61957) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1121.050989] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3ce525f-d808-4ffc-b289-9b0fb15d73ef {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.059409] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51ea30d7-24cb-4391-9ca8-b76836b52e68 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.091997] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-019d02b0-2ae4-4df3-9dd2-e90d9cdbfca8 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.099467] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5c7a20e-fe29-4b3e-ab8e-4b1c2aea29ec {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.112482] env[61957]: DEBUG nova.compute.provider_tree [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1121.616216] env[61957]: DEBUG nova.scheduler.client.report [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1122.120979] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61957) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1122.121299] env[61957]: DEBUG oslo_concurrency.lockutils [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.653s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1122.121616] env[61957]: DEBUG oslo_concurrency.lockutils [None req-66a206ff-5fcb-4812-b9a0-d20a1311acc5 tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.621s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1122.123405] env[61957]: INFO nova.compute.claims [None req-66a206ff-5fcb-4812-b9a0-d20a1311acc5 tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] [instance: 363e48d7-8ed1-45d5-9443-d0b715952fca] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1122.283861] env[61957]: DEBUG oslo_concurrency.lockutils [None req-6e5a656a-a04a-46b8-b899-efd16f82d0e7 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Acquiring lock "refresh_cache-b82af1e9-294a-4ba9-bcad-73b2a2aca86d" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1122.284304] env[61957]: DEBUG oslo_concurrency.lockutils [None req-6e5a656a-a04a-46b8-b899-efd16f82d0e7 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Acquired lock "refresh_cache-b82af1e9-294a-4ba9-bcad-73b2a2aca86d" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1122.284539] env[61957]: DEBUG nova.network.neutron [None req-6e5a656a-a04a-46b8-b899-efd16f82d0e7 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: b82af1e9-294a-4ba9-bcad-73b2a2aca86d] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1122.842461] env[61957]: DEBUG oslo_concurrency.lockutils [None req-dd8126a3-9829-46ce-b79c-18844df38b6d tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Acquiring lock "b82af1e9-294a-4ba9-bcad-73b2a2aca86d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1122.842757] env[61957]: DEBUG oslo_concurrency.lockutils [None req-dd8126a3-9829-46ce-b79c-18844df38b6d tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Lock "b82af1e9-294a-4ba9-bcad-73b2a2aca86d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1122.842893] env[61957]: DEBUG oslo_concurrency.lockutils [None req-dd8126a3-9829-46ce-b79c-18844df38b6d tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Acquiring lock "b82af1e9-294a-4ba9-bcad-73b2a2aca86d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1122.843180] env[61957]: DEBUG oslo_concurrency.lockutils [None req-dd8126a3-9829-46ce-b79c-18844df38b6d tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Lock "b82af1e9-294a-4ba9-bcad-73b2a2aca86d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1122.843381] env[61957]: DEBUG oslo_concurrency.lockutils [None req-dd8126a3-9829-46ce-b79c-18844df38b6d tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Lock "b82af1e9-294a-4ba9-bcad-73b2a2aca86d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1122.845494] env[61957]: INFO nova.compute.manager [None req-dd8126a3-9829-46ce-b79c-18844df38b6d tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: b82af1e9-294a-4ba9-bcad-73b2a2aca86d] Terminating instance [ 1122.847281] env[61957]: DEBUG nova.compute.manager [None req-dd8126a3-9829-46ce-b79c-18844df38b6d tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: b82af1e9-294a-4ba9-bcad-73b2a2aca86d] Start destroying the instance on the hypervisor. {{(pid=61957) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1122.847478] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-dd8126a3-9829-46ce-b79c-18844df38b6d tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: b82af1e9-294a-4ba9-bcad-73b2a2aca86d] Destroying instance {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1122.848335] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e38474ac-d476-4e00-a6eb-0b2720891eba {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.855918] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-dd8126a3-9829-46ce-b79c-18844df38b6d tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: b82af1e9-294a-4ba9-bcad-73b2a2aca86d] Powering off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1122.856161] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7cc935e4-46a5-4060-92f7-3829014b5fad {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.862597] env[61957]: DEBUG oslo_vmware.api [None req-dd8126a3-9829-46ce-b79c-18844df38b6d tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Waiting for the task: (returnval){ [ 1122.862597] env[61957]: value = "task-1278128" [ 1122.862597] env[61957]: _type = "Task" [ 1122.862597] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1122.871818] env[61957]: DEBUG oslo_vmware.api [None req-dd8126a3-9829-46ce-b79c-18844df38b6d tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': task-1278128, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1123.014543] env[61957]: INFO nova.network.neutron [None req-6e5a656a-a04a-46b8-b899-efd16f82d0e7 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: b82af1e9-294a-4ba9-bcad-73b2a2aca86d] Port b3373cf5-1370-4b75-b08c-f76c081ec78f from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 1123.015054] env[61957]: DEBUG nova.network.neutron [None req-6e5a656a-a04a-46b8-b899-efd16f82d0e7 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: b82af1e9-294a-4ba9-bcad-73b2a2aca86d] Updating instance_info_cache with network_info: [{"id": "58a3f865-3fc6-462a-b685-fb32e01f6c5e", "address": "fa:16:3e:ae:67:c0", "network": {"id": "364cabcb-046d-4f91-b1ee-5dd3adfae6d7", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1432396967-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.232", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8713aa35bcb24b86ad0b58ca9fc991ba", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b7a73c01-1bb9-4612-a1a7-16d71b732e81", "external-id": "nsx-vlan-transportzone-711", "segmentation_id": 711, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap58a3f865-3f", "ovs_interfaceid": "58a3f865-3fc6-462a-b685-fb32e01f6c5e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1123.126501] env[61957]: DEBUG oslo_service.periodic_task [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61957) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1123.126759] env[61957]: DEBUG oslo_service.periodic_task [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61957) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1123.263145] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0610dc9c-5510-40d1-bcd0-3f5ee1521ab3 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.271086] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ea507e1-5595-49a1-956c-3243a6654a1f {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.301213] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad4b8c4c-bb01-4074-a516-aa07c38daa22 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.308280] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47c02ee2-b7fd-4f55-ba01-376c7c1e074f {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.320813] env[61957]: DEBUG nova.compute.provider_tree [None req-66a206ff-5fcb-4812-b9a0-d20a1311acc5 tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1123.371666] env[61957]: DEBUG oslo_vmware.api [None req-dd8126a3-9829-46ce-b79c-18844df38b6d tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': task-1278128, 'name': PowerOffVM_Task, 'duration_secs': 0.210495} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1123.371892] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-dd8126a3-9829-46ce-b79c-18844df38b6d tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: b82af1e9-294a-4ba9-bcad-73b2a2aca86d] Powered off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1123.372078] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-dd8126a3-9829-46ce-b79c-18844df38b6d tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: b82af1e9-294a-4ba9-bcad-73b2a2aca86d] Unregistering the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1123.372313] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4242a232-eb23-4aa6-b09b-545c7a348ff8 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.431614] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-dd8126a3-9829-46ce-b79c-18844df38b6d tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: b82af1e9-294a-4ba9-bcad-73b2a2aca86d] Unregistered the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1123.431870] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-dd8126a3-9829-46ce-b79c-18844df38b6d tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: b82af1e9-294a-4ba9-bcad-73b2a2aca86d] Deleting contents of the VM from datastore datastore2 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1123.432072] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-dd8126a3-9829-46ce-b79c-18844df38b6d tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Deleting the datastore file [datastore2] b82af1e9-294a-4ba9-bcad-73b2a2aca86d {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1123.432372] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6d60ab69-0862-4df8-9611-2f25abc8bcab {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.438216] env[61957]: DEBUG oslo_vmware.api [None req-dd8126a3-9829-46ce-b79c-18844df38b6d tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Waiting for the task: (returnval){ [ 1123.438216] env[61957]: value = "task-1278130" [ 1123.438216] env[61957]: _type = "Task" [ 1123.438216] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1123.445919] env[61957]: DEBUG oslo_vmware.api [None req-dd8126a3-9829-46ce-b79c-18844df38b6d tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': task-1278130, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1123.518196] env[61957]: DEBUG oslo_concurrency.lockutils [None req-6e5a656a-a04a-46b8-b899-efd16f82d0e7 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Releasing lock "refresh_cache-b82af1e9-294a-4ba9-bcad-73b2a2aca86d" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1123.632105] env[61957]: DEBUG oslo_service.periodic_task [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61957) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1123.632105] env[61957]: DEBUG nova.compute.manager [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Starting heal instance info cache {{(pid=61957) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 1123.632105] env[61957]: DEBUG nova.compute.manager [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Rebuilding the list of instances to heal {{(pid=61957) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9930}} [ 1123.824454] env[61957]: DEBUG nova.scheduler.client.report [None req-66a206ff-5fcb-4812-b9a0-d20a1311acc5 tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1123.948240] env[61957]: DEBUG oslo_vmware.api [None req-dd8126a3-9829-46ce-b79c-18844df38b6d tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': task-1278130, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.157652} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1123.948597] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-dd8126a3-9829-46ce-b79c-18844df38b6d tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Deleted the datastore file {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1123.948654] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-dd8126a3-9829-46ce-b79c-18844df38b6d tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: b82af1e9-294a-4ba9-bcad-73b2a2aca86d] Deleted contents of the VM from datastore datastore2 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1123.948835] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-dd8126a3-9829-46ce-b79c-18844df38b6d tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: b82af1e9-294a-4ba9-bcad-73b2a2aca86d] Instance destroyed {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1123.949069] env[61957]: INFO nova.compute.manager [None req-dd8126a3-9829-46ce-b79c-18844df38b6d tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: b82af1e9-294a-4ba9-bcad-73b2a2aca86d] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1123.949272] env[61957]: DEBUG oslo.service.loopingcall [None req-dd8126a3-9829-46ce-b79c-18844df38b6d tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1123.949477] env[61957]: DEBUG nova.compute.manager [-] [instance: b82af1e9-294a-4ba9-bcad-73b2a2aca86d] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1123.949572] env[61957]: DEBUG nova.network.neutron [-] [instance: b82af1e9-294a-4ba9-bcad-73b2a2aca86d] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1124.021741] env[61957]: DEBUG oslo_concurrency.lockutils [None req-6e5a656a-a04a-46b8-b899-efd16f82d0e7 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Lock "interface-b82af1e9-294a-4ba9-bcad-73b2a2aca86d-b3373cf5-1370-4b75-b08c-f76c081ec78f" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 9.609s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1124.135925] env[61957]: DEBUG nova.compute.manager [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] [instance: b82af1e9-294a-4ba9-bcad-73b2a2aca86d] Skipping network cache update for instance because it is being deleted. {{(pid=61957) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9943}} [ 1124.136445] env[61957]: DEBUG nova.compute.manager [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] [instance: 363e48d7-8ed1-45d5-9443-d0b715952fca] Skipping network cache update for instance because it is Building. {{(pid=61957) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 1124.184391] env[61957]: DEBUG oslo_concurrency.lockutils [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Acquiring lock "refresh_cache-4625d1dc-a621-4a33-9537-9e8cce96c914" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1124.184602] env[61957]: DEBUG oslo_concurrency.lockutils [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Acquired lock "refresh_cache-4625d1dc-a621-4a33-9537-9e8cce96c914" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1124.184789] env[61957]: DEBUG nova.network.neutron [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] [instance: 4625d1dc-a621-4a33-9537-9e8cce96c914] Forcefully refreshing network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 1124.184981] env[61957]: DEBUG nova.objects.instance [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Lazy-loading 'info_cache' on Instance uuid 4625d1dc-a621-4a33-9537-9e8cce96c914 {{(pid=61957) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1124.329438] env[61957]: DEBUG oslo_concurrency.lockutils [None req-66a206ff-5fcb-4812-b9a0-d20a1311acc5 tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.208s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1124.329996] env[61957]: DEBUG nova.compute.manager [None req-66a206ff-5fcb-4812-b9a0-d20a1311acc5 tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] [instance: 363e48d7-8ed1-45d5-9443-d0b715952fca] Start building networks asynchronously for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1124.641804] env[61957]: DEBUG nova.compute.manager [req-6cb2aba0-95f4-43a4-9c85-d192923dffeb req-4f65baf6-ac2f-4436-96b7-53decfdbc5a0 service nova] [instance: b82af1e9-294a-4ba9-bcad-73b2a2aca86d] Received event network-vif-deleted-58a3f865-3fc6-462a-b685-fb32e01f6c5e {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1124.642055] env[61957]: INFO nova.compute.manager [req-6cb2aba0-95f4-43a4-9c85-d192923dffeb req-4f65baf6-ac2f-4436-96b7-53decfdbc5a0 service nova] [instance: b82af1e9-294a-4ba9-bcad-73b2a2aca86d] Neutron deleted interface 58a3f865-3fc6-462a-b685-fb32e01f6c5e; detaching it from the instance and deleting it from the info cache [ 1124.642240] env[61957]: DEBUG nova.network.neutron [req-6cb2aba0-95f4-43a4-9c85-d192923dffeb req-4f65baf6-ac2f-4436-96b7-53decfdbc5a0 service nova] [instance: b82af1e9-294a-4ba9-bcad-73b2a2aca86d] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1124.835288] env[61957]: DEBUG nova.compute.utils [None req-66a206ff-5fcb-4812-b9a0-d20a1311acc5 tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] Using /dev/sd instead of None {{(pid=61957) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1124.837200] env[61957]: DEBUG nova.compute.manager [None req-66a206ff-5fcb-4812-b9a0-d20a1311acc5 tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] [instance: 363e48d7-8ed1-45d5-9443-d0b715952fca] Allocating IP information in the background. {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1124.837421] env[61957]: DEBUG nova.network.neutron [None req-66a206ff-5fcb-4812-b9a0-d20a1311acc5 tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] [instance: 363e48d7-8ed1-45d5-9443-d0b715952fca] allocate_for_instance() {{(pid=61957) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1124.879629] env[61957]: DEBUG nova.policy [None req-66a206ff-5fcb-4812-b9a0-d20a1311acc5 tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2aac3f334b9c4abc8ebd6414463995d3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '29f70feed7fe4f259bcc8aa8633f0aa8', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61957) authorize /opt/stack/nova/nova/policy.py:203}} [ 1125.120811] env[61957]: DEBUG nova.network.neutron [-] [instance: b82af1e9-294a-4ba9-bcad-73b2a2aca86d] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1125.128702] env[61957]: DEBUG nova.network.neutron [None req-66a206ff-5fcb-4812-b9a0-d20a1311acc5 tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] [instance: 363e48d7-8ed1-45d5-9443-d0b715952fca] Successfully created port: 1c634ca3-05f4-4ac4-a523-27b6d5d64e60 {{(pid=61957) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1125.144734] env[61957]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7fa868b2-9f25-4510-bda7-e85121da7724 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.155054] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca12765b-b0d3-447a-86e6-328c594694cd {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.176523] env[61957]: DEBUG nova.compute.manager [req-6cb2aba0-95f4-43a4-9c85-d192923dffeb req-4f65baf6-ac2f-4436-96b7-53decfdbc5a0 service nova] [instance: b82af1e9-294a-4ba9-bcad-73b2a2aca86d] Detach interface failed, port_id=58a3f865-3fc6-462a-b685-fb32e01f6c5e, reason: Instance b82af1e9-294a-4ba9-bcad-73b2a2aca86d could not be found. {{(pid=61957) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1125.340808] env[61957]: DEBUG nova.compute.manager [None req-66a206ff-5fcb-4812-b9a0-d20a1311acc5 tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] [instance: 363e48d7-8ed1-45d5-9443-d0b715952fca] Start building block device mappings for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1125.622882] env[61957]: INFO nova.compute.manager [-] [instance: b82af1e9-294a-4ba9-bcad-73b2a2aca86d] Took 1.67 seconds to deallocate network for instance. [ 1125.898392] env[61957]: DEBUG nova.network.neutron [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] [instance: 4625d1dc-a621-4a33-9537-9e8cce96c914] Updating instance_info_cache with network_info: [{"id": "8365f86e-8b14-47de-a6f4-297107be492e", "address": "fa:16:3e:e8:59:9d", "network": {"id": "364cabcb-046d-4f91-b1ee-5dd3adfae6d7", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1432396967-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8713aa35bcb24b86ad0b58ca9fc991ba", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b7a73c01-1bb9-4612-a1a7-16d71b732e81", "external-id": "nsx-vlan-transportzone-711", "segmentation_id": 711, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8365f86e-8b", "ovs_interfaceid": "8365f86e-8b14-47de-a6f4-297107be492e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1126.129071] env[61957]: DEBUG oslo_concurrency.lockutils [None req-dd8126a3-9829-46ce-b79c-18844df38b6d tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1126.129355] env[61957]: DEBUG oslo_concurrency.lockutils [None req-dd8126a3-9829-46ce-b79c-18844df38b6d tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1126.129548] env[61957]: DEBUG nova.objects.instance [None req-dd8126a3-9829-46ce-b79c-18844df38b6d tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Lazy-loading 'resources' on Instance uuid b82af1e9-294a-4ba9-bcad-73b2a2aca86d {{(pid=61957) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1126.349879] env[61957]: DEBUG nova.compute.manager [None req-66a206ff-5fcb-4812-b9a0-d20a1311acc5 tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] [instance: 363e48d7-8ed1-45d5-9443-d0b715952fca] Start spawning the instance on the hypervisor. {{(pid=61957) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1126.375373] env[61957]: DEBUG nova.virt.hardware [None req-66a206ff-5fcb-4812-b9a0-d20a1311acc5 tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T17:22:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T17:21:50Z,direct_url=,disk_format='vmdk',id=11c76a2c-f705-470a-ba9d-4657858bab38,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='81d3096d0f094373913fc3dc8f5c3c6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T17:21:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1126.375623] env[61957]: DEBUG nova.virt.hardware [None req-66a206ff-5fcb-4812-b9a0-d20a1311acc5 tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] Flavor limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1126.375781] env[61957]: DEBUG nova.virt.hardware [None req-66a206ff-5fcb-4812-b9a0-d20a1311acc5 tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] Image limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1126.375965] env[61957]: DEBUG nova.virt.hardware [None req-66a206ff-5fcb-4812-b9a0-d20a1311acc5 tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] Flavor pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1126.376127] env[61957]: DEBUG nova.virt.hardware [None req-66a206ff-5fcb-4812-b9a0-d20a1311acc5 tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] Image pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1126.376281] env[61957]: DEBUG nova.virt.hardware [None req-66a206ff-5fcb-4812-b9a0-d20a1311acc5 tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1126.376486] env[61957]: DEBUG nova.virt.hardware [None req-66a206ff-5fcb-4812-b9a0-d20a1311acc5 tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1126.376646] env[61957]: DEBUG nova.virt.hardware [None req-66a206ff-5fcb-4812-b9a0-d20a1311acc5 tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1126.376811] env[61957]: DEBUG nova.virt.hardware [None req-66a206ff-5fcb-4812-b9a0-d20a1311acc5 tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] Got 1 possible topologies {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1126.376972] env[61957]: DEBUG nova.virt.hardware [None req-66a206ff-5fcb-4812-b9a0-d20a1311acc5 tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1126.377184] env[61957]: DEBUG nova.virt.hardware [None req-66a206ff-5fcb-4812-b9a0-d20a1311acc5 tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1126.378047] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bbdb5f4-396a-4f36-bcef-0fa30c7f4034 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.385691] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64a6d1d6-9c1c-4014-bcc1-304e94aaa95e {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.400406] env[61957]: DEBUG oslo_concurrency.lockutils [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Releasing lock "refresh_cache-4625d1dc-a621-4a33-9537-9e8cce96c914" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1126.400598] env[61957]: DEBUG nova.compute.manager [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] [instance: 4625d1dc-a621-4a33-9537-9e8cce96c914] Updated the network info_cache for instance {{(pid=61957) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9997}} [ 1126.400978] env[61957]: DEBUG oslo_service.periodic_task [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61957) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1126.401199] env[61957]: DEBUG oslo_service.periodic_task [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61957) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1126.401366] env[61957]: DEBUG oslo_service.periodic_task [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61957) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1126.401510] env[61957]: DEBUG oslo_service.periodic_task [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61957) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1126.401663] env[61957]: DEBUG oslo_service.periodic_task [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61957) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1126.401791] env[61957]: DEBUG nova.compute.manager [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61957) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 1126.401933] env[61957]: DEBUG oslo_service.periodic_task [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61957) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1126.630090] env[61957]: DEBUG nova.network.neutron [None req-66a206ff-5fcb-4812-b9a0-d20a1311acc5 tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] [instance: 363e48d7-8ed1-45d5-9443-d0b715952fca] Successfully updated port: 1c634ca3-05f4-4ac4-a523-27b6d5d64e60 {{(pid=61957) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1126.666032] env[61957]: DEBUG nova.compute.manager [req-8fba7081-d087-4085-93eb-fb7ab55b5714 req-372cc409-abf5-44df-9083-68d8f021e181 service nova] [instance: 363e48d7-8ed1-45d5-9443-d0b715952fca] Received event network-vif-plugged-1c634ca3-05f4-4ac4-a523-27b6d5d64e60 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1126.666273] env[61957]: DEBUG oslo_concurrency.lockutils [req-8fba7081-d087-4085-93eb-fb7ab55b5714 req-372cc409-abf5-44df-9083-68d8f021e181 service nova] Acquiring lock "363e48d7-8ed1-45d5-9443-d0b715952fca-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1126.666479] env[61957]: DEBUG oslo_concurrency.lockutils [req-8fba7081-d087-4085-93eb-fb7ab55b5714 req-372cc409-abf5-44df-9083-68d8f021e181 service nova] Lock "363e48d7-8ed1-45d5-9443-d0b715952fca-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1126.666642] env[61957]: DEBUG oslo_concurrency.lockutils [req-8fba7081-d087-4085-93eb-fb7ab55b5714 req-372cc409-abf5-44df-9083-68d8f021e181 service nova] Lock "363e48d7-8ed1-45d5-9443-d0b715952fca-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1126.666806] env[61957]: DEBUG nova.compute.manager [req-8fba7081-d087-4085-93eb-fb7ab55b5714 req-372cc409-abf5-44df-9083-68d8f021e181 service nova] [instance: 363e48d7-8ed1-45d5-9443-d0b715952fca] No waiting events found dispatching network-vif-plugged-1c634ca3-05f4-4ac4-a523-27b6d5d64e60 {{(pid=61957) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1126.666968] env[61957]: WARNING nova.compute.manager [req-8fba7081-d087-4085-93eb-fb7ab55b5714 req-372cc409-abf5-44df-9083-68d8f021e181 service nova] [instance: 363e48d7-8ed1-45d5-9443-d0b715952fca] Received unexpected event network-vif-plugged-1c634ca3-05f4-4ac4-a523-27b6d5d64e60 for instance with vm_state building and task_state spawning. [ 1126.667140] env[61957]: DEBUG nova.compute.manager [req-8fba7081-d087-4085-93eb-fb7ab55b5714 req-372cc409-abf5-44df-9083-68d8f021e181 service nova] [instance: 363e48d7-8ed1-45d5-9443-d0b715952fca] Received event network-changed-1c634ca3-05f4-4ac4-a523-27b6d5d64e60 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1126.667394] env[61957]: DEBUG nova.compute.manager [req-8fba7081-d087-4085-93eb-fb7ab55b5714 req-372cc409-abf5-44df-9083-68d8f021e181 service nova] [instance: 363e48d7-8ed1-45d5-9443-d0b715952fca] Refreshing instance network info cache due to event network-changed-1c634ca3-05f4-4ac4-a523-27b6d5d64e60. {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1126.667483] env[61957]: DEBUG oslo_concurrency.lockutils [req-8fba7081-d087-4085-93eb-fb7ab55b5714 req-372cc409-abf5-44df-9083-68d8f021e181 service nova] Acquiring lock "refresh_cache-363e48d7-8ed1-45d5-9443-d0b715952fca" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1126.667614] env[61957]: DEBUG oslo_concurrency.lockutils [req-8fba7081-d087-4085-93eb-fb7ab55b5714 req-372cc409-abf5-44df-9083-68d8f021e181 service nova] Acquired lock "refresh_cache-363e48d7-8ed1-45d5-9443-d0b715952fca" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1126.667767] env[61957]: DEBUG nova.network.neutron [req-8fba7081-d087-4085-93eb-fb7ab55b5714 req-372cc409-abf5-44df-9083-68d8f021e181 service nova] [instance: 363e48d7-8ed1-45d5-9443-d0b715952fca] Refreshing network info cache for port 1c634ca3-05f4-4ac4-a523-27b6d5d64e60 {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1126.684386] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2a8da27-8b91-44e7-9c58-09924cf7f0d4 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.691568] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f14c331-fb9c-45c2-9f24-759299bf5513 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.720991] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7d991e4-3a1b-4c12-9c64-62fa2c91cf6a {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.727826] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51030dd1-64d1-45fe-801b-e249f7c9b374 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.740300] env[61957]: DEBUG nova.compute.provider_tree [None req-dd8126a3-9829-46ce-b79c-18844df38b6d tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1126.904969] env[61957]: DEBUG oslo_concurrency.lockutils [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1127.134808] env[61957]: DEBUG oslo_concurrency.lockutils [None req-66a206ff-5fcb-4812-b9a0-d20a1311acc5 tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] Acquiring lock "refresh_cache-363e48d7-8ed1-45d5-9443-d0b715952fca" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1127.198075] env[61957]: DEBUG nova.network.neutron [req-8fba7081-d087-4085-93eb-fb7ab55b5714 req-372cc409-abf5-44df-9083-68d8f021e181 service nova] [instance: 363e48d7-8ed1-45d5-9443-d0b715952fca] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1127.242983] env[61957]: DEBUG nova.scheduler.client.report [None req-dd8126a3-9829-46ce-b79c-18844df38b6d tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1127.266885] env[61957]: DEBUG nova.network.neutron [req-8fba7081-d087-4085-93eb-fb7ab55b5714 req-372cc409-abf5-44df-9083-68d8f021e181 service nova] [instance: 363e48d7-8ed1-45d5-9443-d0b715952fca] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1127.748115] env[61957]: DEBUG oslo_concurrency.lockutils [None req-dd8126a3-9829-46ce-b79c-18844df38b6d tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.619s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1127.750581] env[61957]: DEBUG oslo_concurrency.lockutils [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.846s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1127.750777] env[61957]: DEBUG oslo_concurrency.lockutils [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1127.750937] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61957) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1127.751812] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd9e5a80-2171-4390-a009-2d2624646df2 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.760946] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9be455d-43df-41d5-b7c6-ac8c9a054d32 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.776216] env[61957]: DEBUG oslo_concurrency.lockutils [req-8fba7081-d087-4085-93eb-fb7ab55b5714 req-372cc409-abf5-44df-9083-68d8f021e181 service nova] Releasing lock "refresh_cache-363e48d7-8ed1-45d5-9443-d0b715952fca" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1127.777540] env[61957]: INFO nova.scheduler.client.report [None req-dd8126a3-9829-46ce-b79c-18844df38b6d tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Deleted allocations for instance b82af1e9-294a-4ba9-bcad-73b2a2aca86d [ 1127.778465] env[61957]: DEBUG oslo_concurrency.lockutils [None req-66a206ff-5fcb-4812-b9a0-d20a1311acc5 tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] Acquired lock "refresh_cache-363e48d7-8ed1-45d5-9443-d0b715952fca" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1127.778615] env[61957]: DEBUG nova.network.neutron [None req-66a206ff-5fcb-4812-b9a0-d20a1311acc5 tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] [instance: 363e48d7-8ed1-45d5-9443-d0b715952fca] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1127.779973] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b265e3c3-5050-44bf-aeb0-c2b5bf45bdb4 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.789188] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c15c8e44-2f07-4447-84aa-c022d597c25a {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.819307] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181087MB free_disk=142GB free_vcpus=48 pci_devices=None {{(pid=61957) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1127.819510] env[61957]: DEBUG oslo_concurrency.lockutils [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1127.819740] env[61957]: DEBUG oslo_concurrency.lockutils [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1128.290090] env[61957]: DEBUG oslo_concurrency.lockutils [None req-dd8126a3-9829-46ce-b79c-18844df38b6d tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Lock "b82af1e9-294a-4ba9-bcad-73b2a2aca86d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.447s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1128.315338] env[61957]: DEBUG nova.network.neutron [None req-66a206ff-5fcb-4812-b9a0-d20a1311acc5 tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] [instance: 363e48d7-8ed1-45d5-9443-d0b715952fca] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1128.435239] env[61957]: DEBUG nova.network.neutron [None req-66a206ff-5fcb-4812-b9a0-d20a1311acc5 tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] [instance: 363e48d7-8ed1-45d5-9443-d0b715952fca] Updating instance_info_cache with network_info: [{"id": "1c634ca3-05f4-4ac4-a523-27b6d5d64e60", "address": "fa:16:3e:1f:40:85", "network": {"id": "a712bee6-193d-44ff-862a-fe63b0ee15ee", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-460165686-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "29f70feed7fe4f259bcc8aa8633f0aa8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9630cae2-7dd9-42b7-8b53-91ab254af243", "external-id": "nsx-vlan-transportzone-900", "segmentation_id": 900, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1c634ca3-05", "ovs_interfaceid": "1c634ca3-05f4-4ac4-a523-27b6d5d64e60", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1128.842570] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Instance 4625d1dc-a621-4a33-9537-9e8cce96c914 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61957) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1128.842731] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Instance 363e48d7-8ed1-45d5-9443-d0b715952fca actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61957) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1128.842908] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Total usable vcpus: 48, total allocated vcpus: 2 {{(pid=61957) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1128.843101] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=896MB phys_disk=200GB used_disk=2GB total_vcpus=48 used_vcpus=2 pci_stats=[] {{(pid=61957) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1128.882616] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70ee0961-9842-413c-883f-69102f8ed817 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.890130] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f41710b-e9fc-4d36-a989-22d96bd59e35 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.920099] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab95cb7e-b59e-4d44-a4fe-b1ee6ae52c4f {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.930248] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a8f68e7-f5ae-4b0d-9198-e55716d236ae {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.943614] env[61957]: DEBUG oslo_concurrency.lockutils [None req-66a206ff-5fcb-4812-b9a0-d20a1311acc5 tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] Releasing lock "refresh_cache-363e48d7-8ed1-45d5-9443-d0b715952fca" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1128.943934] env[61957]: DEBUG nova.compute.manager [None req-66a206ff-5fcb-4812-b9a0-d20a1311acc5 tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] [instance: 363e48d7-8ed1-45d5-9443-d0b715952fca] Instance network_info: |[{"id": "1c634ca3-05f4-4ac4-a523-27b6d5d64e60", "address": "fa:16:3e:1f:40:85", "network": {"id": "a712bee6-193d-44ff-862a-fe63b0ee15ee", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-460165686-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "29f70feed7fe4f259bcc8aa8633f0aa8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9630cae2-7dd9-42b7-8b53-91ab254af243", "external-id": "nsx-vlan-transportzone-900", "segmentation_id": 900, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1c634ca3-05", "ovs_interfaceid": "1c634ca3-05f4-4ac4-a523-27b6d5d64e60", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1128.944594] env[61957]: DEBUG nova.compute.provider_tree [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1128.945900] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-66a206ff-5fcb-4812-b9a0-d20a1311acc5 tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] [instance: 363e48d7-8ed1-45d5-9443-d0b715952fca] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:1f:40:85', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9630cae2-7dd9-42b7-8b53-91ab254af243', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1c634ca3-05f4-4ac4-a523-27b6d5d64e60', 'vif_model': 'vmxnet3'}] {{(pid=61957) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1128.955039] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-66a206ff-5fcb-4812-b9a0-d20a1311acc5 tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] Creating folder: Project (29f70feed7fe4f259bcc8aa8633f0aa8). Parent ref: group-v274445. {{(pid=61957) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1128.955039] env[61957]: DEBUG nova.scheduler.client.report [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1128.957713] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-27ca9690-77b5-4368-8ffc-a5aca294d775 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.968638] env[61957]: INFO nova.virt.vmwareapi.vm_util [None req-66a206ff-5fcb-4812-b9a0-d20a1311acc5 tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] Created folder: Project (29f70feed7fe4f259bcc8aa8633f0aa8) in parent group-v274445. [ 1128.968977] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-66a206ff-5fcb-4812-b9a0-d20a1311acc5 tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] Creating folder: Instances. Parent ref: group-v274610. {{(pid=61957) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1128.969560] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3e4cf708-eb84-453f-b33a-6eb8c37ec966 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.978881] env[61957]: INFO nova.virt.vmwareapi.vm_util [None req-66a206ff-5fcb-4812-b9a0-d20a1311acc5 tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] Created folder: Instances in parent group-v274610. [ 1128.979178] env[61957]: DEBUG oslo.service.loopingcall [None req-66a206ff-5fcb-4812-b9a0-d20a1311acc5 tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1128.979381] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 363e48d7-8ed1-45d5-9443-d0b715952fca] Creating VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1128.979576] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-fd695eea-f707-42c8-a897-5d5aa35a52b9 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.998223] env[61957]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1128.998223] env[61957]: value = "task-1278134" [ 1128.998223] env[61957]: _type = "Task" [ 1128.998223] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1129.005567] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1278134, 'name': CreateVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1129.426020] env[61957]: DEBUG oslo_concurrency.lockutils [None req-9c924a22-9e9c-4de5-b18e-12e9214683e8 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Acquiring lock "4625d1dc-a621-4a33-9537-9e8cce96c914" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1129.426452] env[61957]: DEBUG oslo_concurrency.lockutils [None req-9c924a22-9e9c-4de5-b18e-12e9214683e8 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Lock "4625d1dc-a621-4a33-9537-9e8cce96c914" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1129.426527] env[61957]: DEBUG oslo_concurrency.lockutils [None req-9c924a22-9e9c-4de5-b18e-12e9214683e8 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Acquiring lock "4625d1dc-a621-4a33-9537-9e8cce96c914-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1129.426748] env[61957]: DEBUG oslo_concurrency.lockutils [None req-9c924a22-9e9c-4de5-b18e-12e9214683e8 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Lock "4625d1dc-a621-4a33-9537-9e8cce96c914-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1129.426930] env[61957]: DEBUG oslo_concurrency.lockutils [None req-9c924a22-9e9c-4de5-b18e-12e9214683e8 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Lock "4625d1dc-a621-4a33-9537-9e8cce96c914-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1129.429096] env[61957]: INFO nova.compute.manager [None req-9c924a22-9e9c-4de5-b18e-12e9214683e8 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 4625d1dc-a621-4a33-9537-9e8cce96c914] Terminating instance [ 1129.431264] env[61957]: DEBUG nova.compute.manager [None req-9c924a22-9e9c-4de5-b18e-12e9214683e8 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 4625d1dc-a621-4a33-9537-9e8cce96c914] Start destroying the instance on the hypervisor. {{(pid=61957) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1129.431264] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-9c924a22-9e9c-4de5-b18e-12e9214683e8 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 4625d1dc-a621-4a33-9537-9e8cce96c914] Destroying instance {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1129.431914] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9078ba3d-2c38-4267-afbb-ca3a2ee298b9 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.441360] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-9c924a22-9e9c-4de5-b18e-12e9214683e8 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 4625d1dc-a621-4a33-9537-9e8cce96c914] Powering off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1129.441670] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b56d0563-117c-4a73-a72b-495313f3b45f {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.447371] env[61957]: DEBUG oslo_vmware.api [None req-9c924a22-9e9c-4de5-b18e-12e9214683e8 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Waiting for the task: (returnval){ [ 1129.447371] env[61957]: value = "task-1278135" [ 1129.447371] env[61957]: _type = "Task" [ 1129.447371] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1129.454455] env[61957]: DEBUG oslo_vmware.api [None req-9c924a22-9e9c-4de5-b18e-12e9214683e8 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': task-1278135, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1129.461290] env[61957]: DEBUG nova.compute.resource_tracker [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61957) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1129.461472] env[61957]: DEBUG oslo_concurrency.lockutils [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 1.642s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1129.461669] env[61957]: DEBUG oslo_service.periodic_task [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=61957) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1129.461801] env[61957]: DEBUG nova.compute.manager [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Cleaning up deleted instances {{(pid=61957) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11226}} [ 1129.508590] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1278134, 'name': CreateVM_Task, 'duration_secs': 0.313795} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1129.508849] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 363e48d7-8ed1-45d5-9443-d0b715952fca] Created VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1129.509613] env[61957]: DEBUG oslo_concurrency.lockutils [None req-66a206ff-5fcb-4812-b9a0-d20a1311acc5 tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1129.509801] env[61957]: DEBUG oslo_concurrency.lockutils [None req-66a206ff-5fcb-4812-b9a0-d20a1311acc5 tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1129.510232] env[61957]: DEBUG oslo_concurrency.lockutils [None req-66a206ff-5fcb-4812-b9a0-d20a1311acc5 tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1129.510523] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fda710f3-b83b-484f-a081-f751269f00a1 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.515468] env[61957]: DEBUG oslo_vmware.api [None req-66a206ff-5fcb-4812-b9a0-d20a1311acc5 tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] Waiting for the task: (returnval){ [ 1129.515468] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52e3112b-df26-5015-adc7-a71176293611" [ 1129.515468] env[61957]: _type = "Task" [ 1129.515468] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1129.523399] env[61957]: DEBUG oslo_vmware.api [None req-66a206ff-5fcb-4812-b9a0-d20a1311acc5 tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52e3112b-df26-5015-adc7-a71176293611, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1129.957835] env[61957]: DEBUG oslo_vmware.api [None req-9c924a22-9e9c-4de5-b18e-12e9214683e8 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': task-1278135, 'name': PowerOffVM_Task, 'duration_secs': 0.194911} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1129.958138] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-9c924a22-9e9c-4de5-b18e-12e9214683e8 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 4625d1dc-a621-4a33-9537-9e8cce96c914] Powered off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1129.958317] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-9c924a22-9e9c-4de5-b18e-12e9214683e8 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 4625d1dc-a621-4a33-9537-9e8cce96c914] Unregistering the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1129.958579] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8d14cd7b-2a75-4837-8172-3c55d85f44f1 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.977265] env[61957]: DEBUG nova.compute.manager [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] There are 54 instances to clean {{(pid=61957) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11235}} [ 1129.977265] env[61957]: DEBUG nova.compute.manager [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] [instance: 36769881-ff61-4bf3-b7e4-e7af19275805] Instance has had 0 of 5 cleanup attempts {{(pid=61957) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1130.026602] env[61957]: DEBUG oslo_vmware.api [None req-66a206ff-5fcb-4812-b9a0-d20a1311acc5 tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52e3112b-df26-5015-adc7-a71176293611, 'name': SearchDatastore_Task, 'duration_secs': 0.009512} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1130.026920] env[61957]: DEBUG oslo_concurrency.lockutils [None req-66a206ff-5fcb-4812-b9a0-d20a1311acc5 tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1130.027171] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-66a206ff-5fcb-4812-b9a0-d20a1311acc5 tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] [instance: 363e48d7-8ed1-45d5-9443-d0b715952fca] Processing image 11c76a2c-f705-470a-ba9d-4657858bab38 {{(pid=61957) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1130.027413] env[61957]: DEBUG oslo_concurrency.lockutils [None req-66a206ff-5fcb-4812-b9a0-d20a1311acc5 tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1130.027636] env[61957]: DEBUG oslo_concurrency.lockutils [None req-66a206ff-5fcb-4812-b9a0-d20a1311acc5 tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1130.027749] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-66a206ff-5fcb-4812-b9a0-d20a1311acc5 tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1130.027994] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1249fd03-ffb8-432a-b3cb-72aeafcb61a9 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.031104] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-9c924a22-9e9c-4de5-b18e-12e9214683e8 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 4625d1dc-a621-4a33-9537-9e8cce96c914] Unregistered the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1130.031315] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-9c924a22-9e9c-4de5-b18e-12e9214683e8 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 4625d1dc-a621-4a33-9537-9e8cce96c914] Deleting contents of the VM from datastore datastore1 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1130.031491] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-9c924a22-9e9c-4de5-b18e-12e9214683e8 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Deleting the datastore file [datastore1] 4625d1dc-a621-4a33-9537-9e8cce96c914 {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1130.031718] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8a999976-237d-46cc-abd1-062d487a153a {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.038471] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-66a206ff-5fcb-4812-b9a0-d20a1311acc5 tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1130.038760] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-66a206ff-5fcb-4812-b9a0-d20a1311acc5 tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61957) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1130.040271] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4d037b81-c3bd-41ca-a2d6-8f035614f496 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.043819] env[61957]: DEBUG oslo_vmware.api [None req-9c924a22-9e9c-4de5-b18e-12e9214683e8 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Waiting for the task: (returnval){ [ 1130.043819] env[61957]: value = "task-1278137" [ 1130.043819] env[61957]: _type = "Task" [ 1130.043819] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1130.049395] env[61957]: DEBUG oslo_vmware.api [None req-66a206ff-5fcb-4812-b9a0-d20a1311acc5 tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] Waiting for the task: (returnval){ [ 1130.049395] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52635482-2446-1c99-6e5a-a9a55a420f50" [ 1130.049395] env[61957]: _type = "Task" [ 1130.049395] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1130.054604] env[61957]: DEBUG oslo_vmware.api [None req-9c924a22-9e9c-4de5-b18e-12e9214683e8 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': task-1278137, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1130.059519] env[61957]: DEBUG oslo_vmware.api [None req-66a206ff-5fcb-4812-b9a0-d20a1311acc5 tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52635482-2446-1c99-6e5a-a9a55a420f50, 'name': SearchDatastore_Task, 'duration_secs': 0.008672} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1130.060246] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e8b50a06-f913-4405-ac68-0fa0d195fe13 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.064948] env[61957]: DEBUG oslo_vmware.api [None req-66a206ff-5fcb-4812-b9a0-d20a1311acc5 tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] Waiting for the task: (returnval){ [ 1130.064948] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]521e629c-7b47-9471-64fc-f8bb416a5053" [ 1130.064948] env[61957]: _type = "Task" [ 1130.064948] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1130.072037] env[61957]: DEBUG oslo_vmware.api [None req-66a206ff-5fcb-4812-b9a0-d20a1311acc5 tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]521e629c-7b47-9471-64fc-f8bb416a5053, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1130.482758] env[61957]: DEBUG nova.compute.manager [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] [instance: 84cd5619-4ac9-41ef-9368-a023ad9ae66d] Instance has had 0 of 5 cleanup attempts {{(pid=61957) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1130.554522] env[61957]: DEBUG oslo_vmware.api [None req-9c924a22-9e9c-4de5-b18e-12e9214683e8 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Task: {'id': task-1278137, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.169959} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1130.554812] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-9c924a22-9e9c-4de5-b18e-12e9214683e8 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Deleted the datastore file {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1130.555011] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-9c924a22-9e9c-4de5-b18e-12e9214683e8 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 4625d1dc-a621-4a33-9537-9e8cce96c914] Deleted contents of the VM from datastore datastore1 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1130.555210] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-9c924a22-9e9c-4de5-b18e-12e9214683e8 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 4625d1dc-a621-4a33-9537-9e8cce96c914] Instance destroyed {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1130.555433] env[61957]: INFO nova.compute.manager [None req-9c924a22-9e9c-4de5-b18e-12e9214683e8 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] [instance: 4625d1dc-a621-4a33-9537-9e8cce96c914] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1130.555664] env[61957]: DEBUG oslo.service.loopingcall [None req-9c924a22-9e9c-4de5-b18e-12e9214683e8 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1130.556221] env[61957]: DEBUG nova.compute.manager [-] [instance: 4625d1dc-a621-4a33-9537-9e8cce96c914] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1130.556221] env[61957]: DEBUG nova.network.neutron [-] [instance: 4625d1dc-a621-4a33-9537-9e8cce96c914] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1130.574733] env[61957]: DEBUG oslo_vmware.api [None req-66a206ff-5fcb-4812-b9a0-d20a1311acc5 tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]521e629c-7b47-9471-64fc-f8bb416a5053, 'name': SearchDatastore_Task, 'duration_secs': 0.008799} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1130.574983] env[61957]: DEBUG oslo_concurrency.lockutils [None req-66a206ff-5fcb-4812-b9a0-d20a1311acc5 tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1130.575256] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-66a206ff-5fcb-4812-b9a0-d20a1311acc5 tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore2] 363e48d7-8ed1-45d5-9443-d0b715952fca/363e48d7-8ed1-45d5-9443-d0b715952fca.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1130.575517] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-dc746a02-6cd1-44bf-ad15-0f85fb031206 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.582256] env[61957]: DEBUG oslo_vmware.api [None req-66a206ff-5fcb-4812-b9a0-d20a1311acc5 tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] Waiting for the task: (returnval){ [ 1130.582256] env[61957]: value = "task-1278138" [ 1130.582256] env[61957]: _type = "Task" [ 1130.582256] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1130.591020] env[61957]: DEBUG oslo_vmware.api [None req-66a206ff-5fcb-4812-b9a0-d20a1311acc5 tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] Task: {'id': task-1278138, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1130.989364] env[61957]: DEBUG nova.compute.manager [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] [instance: b82af1e9-294a-4ba9-bcad-73b2a2aca86d] Instance has had 0 of 5 cleanup attempts {{(pid=61957) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1131.032349] env[61957]: DEBUG nova.compute.manager [req-f81c44bf-4b51-4430-bd4c-6ffe89bc9fea req-bf6e2967-3d68-4525-9827-7b1d82a5716c service nova] [instance: 4625d1dc-a621-4a33-9537-9e8cce96c914] Received event network-vif-deleted-8365f86e-8b14-47de-a6f4-297107be492e {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1131.033012] env[61957]: INFO nova.compute.manager [req-f81c44bf-4b51-4430-bd4c-6ffe89bc9fea req-bf6e2967-3d68-4525-9827-7b1d82a5716c service nova] [instance: 4625d1dc-a621-4a33-9537-9e8cce96c914] Neutron deleted interface 8365f86e-8b14-47de-a6f4-297107be492e; detaching it from the instance and deleting it from the info cache [ 1131.033394] env[61957]: DEBUG nova.network.neutron [req-f81c44bf-4b51-4430-bd4c-6ffe89bc9fea req-bf6e2967-3d68-4525-9827-7b1d82a5716c service nova] [instance: 4625d1dc-a621-4a33-9537-9e8cce96c914] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1131.091907] env[61957]: DEBUG oslo_vmware.api [None req-66a206ff-5fcb-4812-b9a0-d20a1311acc5 tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] Task: {'id': task-1278138, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.432046} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1131.092290] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-66a206ff-5fcb-4812-b9a0-d20a1311acc5 tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore2] 363e48d7-8ed1-45d5-9443-d0b715952fca/363e48d7-8ed1-45d5-9443-d0b715952fca.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1131.092566] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-66a206ff-5fcb-4812-b9a0-d20a1311acc5 tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] [instance: 363e48d7-8ed1-45d5-9443-d0b715952fca] Extending root virtual disk to 1048576 {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1131.092849] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ce799035-f767-43d4-9daa-0936a43c8cac {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.100259] env[61957]: DEBUG oslo_vmware.api [None req-66a206ff-5fcb-4812-b9a0-d20a1311acc5 tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] Waiting for the task: (returnval){ [ 1131.100259] env[61957]: value = "task-1278139" [ 1131.100259] env[61957]: _type = "Task" [ 1131.100259] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1131.108076] env[61957]: DEBUG oslo_vmware.api [None req-66a206ff-5fcb-4812-b9a0-d20a1311acc5 tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] Task: {'id': task-1278139, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1131.491015] env[61957]: DEBUG nova.compute.manager [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] [instance: 11d67afa-1d96-4a9b-8439-383d586d8e17] Instance has had 0 of 5 cleanup attempts {{(pid=61957) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1131.507975] env[61957]: DEBUG nova.network.neutron [-] [instance: 4625d1dc-a621-4a33-9537-9e8cce96c914] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1131.535938] env[61957]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f2f35c38-d9d4-41e5-8428-b3dfac2085b6 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.545096] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ad99c97-e946-40d9-94f9-8c1e573a2695 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.567060] env[61957]: DEBUG nova.compute.manager [req-f81c44bf-4b51-4430-bd4c-6ffe89bc9fea req-bf6e2967-3d68-4525-9827-7b1d82a5716c service nova] [instance: 4625d1dc-a621-4a33-9537-9e8cce96c914] Detach interface failed, port_id=8365f86e-8b14-47de-a6f4-297107be492e, reason: Instance 4625d1dc-a621-4a33-9537-9e8cce96c914 could not be found. {{(pid=61957) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1131.608335] env[61957]: DEBUG oslo_vmware.api [None req-66a206ff-5fcb-4812-b9a0-d20a1311acc5 tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] Task: {'id': task-1278139, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.07139} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1131.608597] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-66a206ff-5fcb-4812-b9a0-d20a1311acc5 tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] [instance: 363e48d7-8ed1-45d5-9443-d0b715952fca] Extended root virtual disk {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1131.609342] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a486494a-2d5f-4fa5-ba85-0ef4ad382377 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.630305] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-66a206ff-5fcb-4812-b9a0-d20a1311acc5 tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] [instance: 363e48d7-8ed1-45d5-9443-d0b715952fca] Reconfiguring VM instance instance-0000006a to attach disk [datastore2] 363e48d7-8ed1-45d5-9443-d0b715952fca/363e48d7-8ed1-45d5-9443-d0b715952fca.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1131.630532] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8ba1afa9-1c1d-452b-a3c3-1be8bcb68f53 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.648700] env[61957]: DEBUG oslo_vmware.api [None req-66a206ff-5fcb-4812-b9a0-d20a1311acc5 tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] Waiting for the task: (returnval){ [ 1131.648700] env[61957]: value = "task-1278140" [ 1131.648700] env[61957]: _type = "Task" [ 1131.648700] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1131.656182] env[61957]: DEBUG oslo_vmware.api [None req-66a206ff-5fcb-4812-b9a0-d20a1311acc5 tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] Task: {'id': task-1278140, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1131.995774] env[61957]: DEBUG nova.compute.manager [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] [instance: a7419b30-1539-4977-a5d9-93a49d84914a] Instance has had 0 of 5 cleanup attempts {{(pid=61957) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1132.010506] env[61957]: INFO nova.compute.manager [-] [instance: 4625d1dc-a621-4a33-9537-9e8cce96c914] Took 1.45 seconds to deallocate network for instance. [ 1132.159151] env[61957]: DEBUG oslo_vmware.api [None req-66a206ff-5fcb-4812-b9a0-d20a1311acc5 tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] Task: {'id': task-1278140, 'name': ReconfigVM_Task, 'duration_secs': 0.456973} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1132.159411] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-66a206ff-5fcb-4812-b9a0-d20a1311acc5 tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] [instance: 363e48d7-8ed1-45d5-9443-d0b715952fca] Reconfigured VM instance instance-0000006a to attach disk [datastore2] 363e48d7-8ed1-45d5-9443-d0b715952fca/363e48d7-8ed1-45d5-9443-d0b715952fca.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1132.160029] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e36c797e-9e5b-4fff-bc25-43b7418b1e88 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.165832] env[61957]: DEBUG oslo_vmware.api [None req-66a206ff-5fcb-4812-b9a0-d20a1311acc5 tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] Waiting for the task: (returnval){ [ 1132.165832] env[61957]: value = "task-1278141" [ 1132.165832] env[61957]: _type = "Task" [ 1132.165832] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1132.173476] env[61957]: DEBUG oslo_vmware.api [None req-66a206ff-5fcb-4812-b9a0-d20a1311acc5 tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] Task: {'id': task-1278141, 'name': Rename_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1132.499217] env[61957]: DEBUG nova.compute.manager [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] [instance: 42860944-c700-43be-b291-970803fc7955] Instance has had 0 of 5 cleanup attempts {{(pid=61957) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1132.517616] env[61957]: DEBUG oslo_concurrency.lockutils [None req-9c924a22-9e9c-4de5-b18e-12e9214683e8 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1132.517949] env[61957]: DEBUG oslo_concurrency.lockutils [None req-9c924a22-9e9c-4de5-b18e-12e9214683e8 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1132.518192] env[61957]: DEBUG nova.objects.instance [None req-9c924a22-9e9c-4de5-b18e-12e9214683e8 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Lazy-loading 'resources' on Instance uuid 4625d1dc-a621-4a33-9537-9e8cce96c914 {{(pid=61957) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1132.675582] env[61957]: DEBUG oslo_vmware.api [None req-66a206ff-5fcb-4812-b9a0-d20a1311acc5 tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] Task: {'id': task-1278141, 'name': Rename_Task, 'duration_secs': 0.138447} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1132.675881] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-66a206ff-5fcb-4812-b9a0-d20a1311acc5 tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] [instance: 363e48d7-8ed1-45d5-9443-d0b715952fca] Powering on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1132.676143] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-727ca2cb-23d0-4c47-8f53-7c04f8848144 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.681873] env[61957]: DEBUG oslo_vmware.api [None req-66a206ff-5fcb-4812-b9a0-d20a1311acc5 tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] Waiting for the task: (returnval){ [ 1132.681873] env[61957]: value = "task-1278142" [ 1132.681873] env[61957]: _type = "Task" [ 1132.681873] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1132.688976] env[61957]: DEBUG oslo_vmware.api [None req-66a206ff-5fcb-4812-b9a0-d20a1311acc5 tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] Task: {'id': task-1278142, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.002622] env[61957]: DEBUG nova.compute.manager [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] [instance: 039a6b01-a745-478c-9ff7-c8c37afe2ce4] Instance has had 0 of 5 cleanup attempts {{(pid=61957) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1133.064118] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ff29916-f1c1-407f-83c6-bdc12e31bd44 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.071918] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb67a743-86f8-48a6-a368-6de24652b317 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.102723] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9f88770-490a-4bc9-9f28-6e94c41cb38f {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.110105] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f69420a9-3f27-4685-83f9-789726d576a5 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.124735] env[61957]: DEBUG nova.compute.provider_tree [None req-9c924a22-9e9c-4de5-b18e-12e9214683e8 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1133.190944] env[61957]: DEBUG oslo_vmware.api [None req-66a206ff-5fcb-4812-b9a0-d20a1311acc5 tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] Task: {'id': task-1278142, 'name': PowerOnVM_Task, 'duration_secs': 0.405689} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1133.191245] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-66a206ff-5fcb-4812-b9a0-d20a1311acc5 tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] [instance: 363e48d7-8ed1-45d5-9443-d0b715952fca] Powered on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1133.191453] env[61957]: INFO nova.compute.manager [None req-66a206ff-5fcb-4812-b9a0-d20a1311acc5 tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] [instance: 363e48d7-8ed1-45d5-9443-d0b715952fca] Took 6.84 seconds to spawn the instance on the hypervisor. [ 1133.191635] env[61957]: DEBUG nova.compute.manager [None req-66a206ff-5fcb-4812-b9a0-d20a1311acc5 tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] [instance: 363e48d7-8ed1-45d5-9443-d0b715952fca] Checking state {{(pid=61957) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1133.192390] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ee32f16-374e-40d5-a02b-a434d639febb {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.506319] env[61957]: DEBUG nova.compute.manager [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] [instance: 3f144012-c0cc-43ce-8c73-4422fe8c1fe6] Instance has had 0 of 5 cleanup attempts {{(pid=61957) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1133.628145] env[61957]: DEBUG nova.scheduler.client.report [None req-9c924a22-9e9c-4de5-b18e-12e9214683e8 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1133.707019] env[61957]: INFO nova.compute.manager [None req-66a206ff-5fcb-4812-b9a0-d20a1311acc5 tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] [instance: 363e48d7-8ed1-45d5-9443-d0b715952fca] Took 13.22 seconds to build instance. [ 1133.929082] env[61957]: INFO nova.compute.manager [None req-ad676c14-83a8-4682-b842-12b0dc0fabcc tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] [instance: 363e48d7-8ed1-45d5-9443-d0b715952fca] Rescuing [ 1133.929359] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ad676c14-83a8-4682-b842-12b0dc0fabcc tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] Acquiring lock "refresh_cache-363e48d7-8ed1-45d5-9443-d0b715952fca" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1133.929513] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ad676c14-83a8-4682-b842-12b0dc0fabcc tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] Acquired lock "refresh_cache-363e48d7-8ed1-45d5-9443-d0b715952fca" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1133.929691] env[61957]: DEBUG nova.network.neutron [None req-ad676c14-83a8-4682-b842-12b0dc0fabcc tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] [instance: 363e48d7-8ed1-45d5-9443-d0b715952fca] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1134.009360] env[61957]: DEBUG nova.compute.manager [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] [instance: 927af4c3-c7e8-404c-839e-d25af5e384a9] Instance has had 0 of 5 cleanup attempts {{(pid=61957) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1134.132311] env[61957]: DEBUG oslo_concurrency.lockutils [None req-9c924a22-9e9c-4de5-b18e-12e9214683e8 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.614s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1134.153838] env[61957]: INFO nova.scheduler.client.report [None req-9c924a22-9e9c-4de5-b18e-12e9214683e8 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Deleted allocations for instance 4625d1dc-a621-4a33-9537-9e8cce96c914 [ 1134.208905] env[61957]: DEBUG oslo_concurrency.lockutils [None req-66a206ff-5fcb-4812-b9a0-d20a1311acc5 tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] Lock "363e48d7-8ed1-45d5-9443-d0b715952fca" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.735s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1134.512745] env[61957]: DEBUG nova.compute.manager [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] [instance: 84cc9489-2a11-4510-8ad6-13dc41814856] Instance has had 0 of 5 cleanup attempts {{(pid=61957) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1134.624332] env[61957]: DEBUG nova.network.neutron [None req-ad676c14-83a8-4682-b842-12b0dc0fabcc tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] [instance: 363e48d7-8ed1-45d5-9443-d0b715952fca] Updating instance_info_cache with network_info: [{"id": "1c634ca3-05f4-4ac4-a523-27b6d5d64e60", "address": "fa:16:3e:1f:40:85", "network": {"id": "a712bee6-193d-44ff-862a-fe63b0ee15ee", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-460165686-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "29f70feed7fe4f259bcc8aa8633f0aa8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9630cae2-7dd9-42b7-8b53-91ab254af243", "external-id": "nsx-vlan-transportzone-900", "segmentation_id": 900, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1c634ca3-05", "ovs_interfaceid": "1c634ca3-05f4-4ac4-a523-27b6d5d64e60", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1134.661249] env[61957]: DEBUG oslo_concurrency.lockutils [None req-9c924a22-9e9c-4de5-b18e-12e9214683e8 tempest-AttachInterfacesTestJSON-1769166894 tempest-AttachInterfacesTestJSON-1769166894-project-member] Lock "4625d1dc-a621-4a33-9537-9e8cce96c914" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.235s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1135.017231] env[61957]: DEBUG nova.compute.manager [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] [instance: e2107d5c-8453-4d24-b364-2b5c1d125062] Instance has had 0 of 5 cleanup attempts {{(pid=61957) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1135.126565] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ad676c14-83a8-4682-b842-12b0dc0fabcc tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] Releasing lock "refresh_cache-363e48d7-8ed1-45d5-9443-d0b715952fca" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1135.519387] env[61957]: DEBUG nova.compute.manager [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] [instance: 6509707e-afc6-4c4a-88f9-a471865d510c] Instance has had 0 of 5 cleanup attempts {{(pid=61957) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1135.658062] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-ad676c14-83a8-4682-b842-12b0dc0fabcc tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] [instance: 363e48d7-8ed1-45d5-9443-d0b715952fca] Powering off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1135.658645] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-971ae359-e1a8-44fe-b1d3-eb55007e62b7 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.666219] env[61957]: DEBUG oslo_vmware.api [None req-ad676c14-83a8-4682-b842-12b0dc0fabcc tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] Waiting for the task: (returnval){ [ 1135.666219] env[61957]: value = "task-1278143" [ 1135.666219] env[61957]: _type = "Task" [ 1135.666219] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1135.674686] env[61957]: DEBUG oslo_vmware.api [None req-ad676c14-83a8-4682-b842-12b0dc0fabcc tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] Task: {'id': task-1278143, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1136.023397] env[61957]: DEBUG nova.compute.manager [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] [instance: fa4c150e-7a85-41a4-8890-9f2deb3b34b1] Instance has had 0 of 5 cleanup attempts {{(pid=61957) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1136.176382] env[61957]: DEBUG oslo_vmware.api [None req-ad676c14-83a8-4682-b842-12b0dc0fabcc tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] Task: {'id': task-1278143, 'name': PowerOffVM_Task, 'duration_secs': 0.196152} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1136.176695] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-ad676c14-83a8-4682-b842-12b0dc0fabcc tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] [instance: 363e48d7-8ed1-45d5-9443-d0b715952fca] Powered off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1136.177482] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-189479c4-384c-4326-99a4-0385148f5deb {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.198186] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e45fbd5-f77e-4da1-87a4-f8d0689d2d13 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.235026] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-ad676c14-83a8-4682-b842-12b0dc0fabcc tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] [instance: 363e48d7-8ed1-45d5-9443-d0b715952fca] Powering off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1136.235026] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-972994f6-3e80-497b-be00-a5756b98d999 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.246642] env[61957]: DEBUG oslo_vmware.api [None req-ad676c14-83a8-4682-b842-12b0dc0fabcc tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] Waiting for the task: (returnval){ [ 1136.246642] env[61957]: value = "task-1278144" [ 1136.246642] env[61957]: _type = "Task" [ 1136.246642] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1136.256091] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-ad676c14-83a8-4682-b842-12b0dc0fabcc tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] [instance: 363e48d7-8ed1-45d5-9443-d0b715952fca] VM already powered off {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 1136.256372] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-ad676c14-83a8-4682-b842-12b0dc0fabcc tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] [instance: 363e48d7-8ed1-45d5-9443-d0b715952fca] Processing image 11c76a2c-f705-470a-ba9d-4657858bab38 {{(pid=61957) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1136.256586] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ad676c14-83a8-4682-b842-12b0dc0fabcc tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1136.256716] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ad676c14-83a8-4682-b842-12b0dc0fabcc tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1136.256942] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-ad676c14-83a8-4682-b842-12b0dc0fabcc tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1136.257193] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-618d63ae-5d06-421b-bbc7-78b1725014ee {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.266381] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-ad676c14-83a8-4682-b842-12b0dc0fabcc tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1136.266381] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-ad676c14-83a8-4682-b842-12b0dc0fabcc tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61957) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1136.267115] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dfb02dc2-77f4-4994-88d1-8419f94b3078 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.273444] env[61957]: DEBUG oslo_vmware.api [None req-ad676c14-83a8-4682-b842-12b0dc0fabcc tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] Waiting for the task: (returnval){ [ 1136.273444] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52ff4ab3-60af-1c86-dc35-13236d8ffc1c" [ 1136.273444] env[61957]: _type = "Task" [ 1136.273444] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1136.279646] env[61957]: DEBUG oslo_vmware.api [None req-ad676c14-83a8-4682-b842-12b0dc0fabcc tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52ff4ab3-60af-1c86-dc35-13236d8ffc1c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1136.527110] env[61957]: DEBUG nova.compute.manager [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] [instance: f65458b2-0db6-4d98-8c82-b27b6945a3fd] Instance has had 0 of 5 cleanup attempts {{(pid=61957) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1136.783599] env[61957]: DEBUG oslo_vmware.api [None req-ad676c14-83a8-4682-b842-12b0dc0fabcc tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52ff4ab3-60af-1c86-dc35-13236d8ffc1c, 'name': SearchDatastore_Task, 'duration_secs': 0.008943} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1136.784497] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7d67099d-8361-4dfa-81c8-422847bf6e48 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.789922] env[61957]: DEBUG oslo_vmware.api [None req-ad676c14-83a8-4682-b842-12b0dc0fabcc tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] Waiting for the task: (returnval){ [ 1136.789922] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52f238fd-f47b-e08e-44b7-cde9d4abf695" [ 1136.789922] env[61957]: _type = "Task" [ 1136.789922] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1136.797692] env[61957]: DEBUG oslo_vmware.api [None req-ad676c14-83a8-4682-b842-12b0dc0fabcc tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52f238fd-f47b-e08e-44b7-cde9d4abf695, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1137.030412] env[61957]: DEBUG nova.compute.manager [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] [instance: 6f45c7c8-de07-4d21-bccc-e4afe3c8ad3d] Instance has had 0 of 5 cleanup attempts {{(pid=61957) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1137.300550] env[61957]: DEBUG oslo_vmware.api [None req-ad676c14-83a8-4682-b842-12b0dc0fabcc tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52f238fd-f47b-e08e-44b7-cde9d4abf695, 'name': SearchDatastore_Task, 'duration_secs': 0.009387} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1137.300825] env[61957]: DEBUG oslo_concurrency.lockutils [None req-ad676c14-83a8-4682-b842-12b0dc0fabcc tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1137.301134] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-ad676c14-83a8-4682-b842-12b0dc0fabcc tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore2] 363e48d7-8ed1-45d5-9443-d0b715952fca/11c76a2c-f705-470a-ba9d-4657858bab38-rescue.vmdk. {{(pid=61957) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 1137.301371] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-371908b5-9d3f-4793-b58a-ecbb4b8bd764 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.309905] env[61957]: DEBUG oslo_vmware.api [None req-ad676c14-83a8-4682-b842-12b0dc0fabcc tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] Waiting for the task: (returnval){ [ 1137.309905] env[61957]: value = "task-1278145" [ 1137.309905] env[61957]: _type = "Task" [ 1137.309905] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1137.317704] env[61957]: DEBUG oslo_vmware.api [None req-ad676c14-83a8-4682-b842-12b0dc0fabcc tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] Task: {'id': task-1278145, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1137.535156] env[61957]: DEBUG nova.compute.manager [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] [instance: 964c8c9d-0721-4ad3-a146-fce1fd1d6513] Instance has had 0 of 5 cleanup attempts {{(pid=61957) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1137.819818] env[61957]: DEBUG oslo_vmware.api [None req-ad676c14-83a8-4682-b842-12b0dc0fabcc tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] Task: {'id': task-1278145, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.45647} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1137.819946] env[61957]: INFO nova.virt.vmwareapi.ds_util [None req-ad676c14-83a8-4682-b842-12b0dc0fabcc tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore2] 363e48d7-8ed1-45d5-9443-d0b715952fca/11c76a2c-f705-470a-ba9d-4657858bab38-rescue.vmdk. [ 1137.820674] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4637f4d4-363a-46b6-945f-23f0aee33a37 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.844895] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-ad676c14-83a8-4682-b842-12b0dc0fabcc tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] [instance: 363e48d7-8ed1-45d5-9443-d0b715952fca] Reconfiguring VM instance instance-0000006a to attach disk [datastore2] 363e48d7-8ed1-45d5-9443-d0b715952fca/11c76a2c-f705-470a-ba9d-4657858bab38-rescue.vmdk or device None with type thin {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1137.845172] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-46557cca-ca33-4a30-a6ce-437bfd5f1c7a {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.863888] env[61957]: DEBUG oslo_vmware.api [None req-ad676c14-83a8-4682-b842-12b0dc0fabcc tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] Waiting for the task: (returnval){ [ 1137.863888] env[61957]: value = "task-1278146" [ 1137.863888] env[61957]: _type = "Task" [ 1137.863888] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1137.874437] env[61957]: DEBUG oslo_vmware.api [None req-ad676c14-83a8-4682-b842-12b0dc0fabcc tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] Task: {'id': task-1278146, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1138.040173] env[61957]: DEBUG nova.compute.manager [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] [instance: d6c5c70d-86de-4dea-8b9e-76f321947a35] Instance has had 0 of 5 cleanup attempts {{(pid=61957) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1138.373916] env[61957]: DEBUG oslo_vmware.api [None req-ad676c14-83a8-4682-b842-12b0dc0fabcc tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] Task: {'id': task-1278146, 'name': ReconfigVM_Task, 'duration_secs': 0.256413} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1138.374473] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-ad676c14-83a8-4682-b842-12b0dc0fabcc tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] [instance: 363e48d7-8ed1-45d5-9443-d0b715952fca] Reconfigured VM instance instance-0000006a to attach disk [datastore2] 363e48d7-8ed1-45d5-9443-d0b715952fca/11c76a2c-f705-470a-ba9d-4657858bab38-rescue.vmdk or device None with type thin {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1138.375474] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89519ecd-3e52-427f-982c-abc012394a2a {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.408760] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-092d7ea7-e38b-49a6-9bce-d69c46a293d0 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.424865] env[61957]: DEBUG oslo_vmware.api [None req-ad676c14-83a8-4682-b842-12b0dc0fabcc tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] Waiting for the task: (returnval){ [ 1138.424865] env[61957]: value = "task-1278147" [ 1138.424865] env[61957]: _type = "Task" [ 1138.424865] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1138.433440] env[61957]: DEBUG oslo_vmware.api [None req-ad676c14-83a8-4682-b842-12b0dc0fabcc tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] Task: {'id': task-1278147, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1138.543230] env[61957]: DEBUG nova.compute.manager [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] [instance: 00ecc8c2-398e-4a29-b19f-a2013985b481] Instance has had 0 of 5 cleanup attempts {{(pid=61957) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1138.936089] env[61957]: DEBUG oslo_vmware.api [None req-ad676c14-83a8-4682-b842-12b0dc0fabcc tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] Task: {'id': task-1278147, 'name': ReconfigVM_Task, 'duration_secs': 0.154489} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1138.936562] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-ad676c14-83a8-4682-b842-12b0dc0fabcc tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] [instance: 363e48d7-8ed1-45d5-9443-d0b715952fca] Powering on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1138.936562] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-135a9354-3b86-4ad7-abf8-8828b7e3788e {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.942828] env[61957]: DEBUG oslo_vmware.api [None req-ad676c14-83a8-4682-b842-12b0dc0fabcc tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] Waiting for the task: (returnval){ [ 1138.942828] env[61957]: value = "task-1278148" [ 1138.942828] env[61957]: _type = "Task" [ 1138.942828] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1138.950653] env[61957]: DEBUG oslo_vmware.api [None req-ad676c14-83a8-4682-b842-12b0dc0fabcc tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] Task: {'id': task-1278148, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1139.048476] env[61957]: DEBUG nova.compute.manager [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] [instance: 08f3b59c-91cc-459b-a6ad-3cb9f4a7cec2] Instance has had 0 of 5 cleanup attempts {{(pid=61957) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1139.455909] env[61957]: DEBUG oslo_vmware.api [None req-ad676c14-83a8-4682-b842-12b0dc0fabcc tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] Task: {'id': task-1278148, 'name': PowerOnVM_Task, 'duration_secs': 0.429055} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1139.456400] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-ad676c14-83a8-4682-b842-12b0dc0fabcc tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] [instance: 363e48d7-8ed1-45d5-9443-d0b715952fca] Powered on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1139.459597] env[61957]: DEBUG nova.compute.manager [None req-ad676c14-83a8-4682-b842-12b0dc0fabcc tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] [instance: 363e48d7-8ed1-45d5-9443-d0b715952fca] Checking state {{(pid=61957) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1139.460375] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1220d3ff-4be9-43f1-844d-02c3ff32fbd4 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.553010] env[61957]: DEBUG nova.compute.manager [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] [instance: 7793aa07-6aa2-459d-8a91-56c6b0412d3b] Instance has had 0 of 5 cleanup attempts {{(pid=61957) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1139.983771] env[61957]: DEBUG oslo_concurrency.lockutils [None req-024ba6d8-b0a9-4c02-8f92-8b55e68f2ea5 tempest-ServerAddressesNegativeTestJSON-995848772 tempest-ServerAddressesNegativeTestJSON-995848772-project-member] Acquiring lock "2fe04bcb-569c-48c3-b300-e67b4abe9fdf" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1139.984111] env[61957]: DEBUG oslo_concurrency.lockutils [None req-024ba6d8-b0a9-4c02-8f92-8b55e68f2ea5 tempest-ServerAddressesNegativeTestJSON-995848772 tempest-ServerAddressesNegativeTestJSON-995848772-project-member] Lock "2fe04bcb-569c-48c3-b300-e67b4abe9fdf" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1140.057374] env[61957]: DEBUG nova.compute.manager [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] [instance: 2e32aae5-edd7-402f-98ad-75a93d26f7a1] Instance has had 0 of 5 cleanup attempts {{(pid=61957) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1140.487054] env[61957]: DEBUG nova.compute.manager [None req-024ba6d8-b0a9-4c02-8f92-8b55e68f2ea5 tempest-ServerAddressesNegativeTestJSON-995848772 tempest-ServerAddressesNegativeTestJSON-995848772-project-member] [instance: 2fe04bcb-569c-48c3-b300-e67b4abe9fdf] Starting instance... {{(pid=61957) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1140.562022] env[61957]: DEBUG nova.compute.manager [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] [instance: 50e09db7-9f8f-452b-8232-bd7473e9e63a] Instance has had 0 of 5 cleanup attempts {{(pid=61957) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1140.926325] env[61957]: DEBUG nova.compute.manager [req-ac40299e-5fc4-4a1e-a536-33297b0fe6b1 req-19234346-a788-4210-8db9-877d01836aa1 service nova] [instance: 363e48d7-8ed1-45d5-9443-d0b715952fca] Received event network-changed-1c634ca3-05f4-4ac4-a523-27b6d5d64e60 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1140.926507] env[61957]: DEBUG nova.compute.manager [req-ac40299e-5fc4-4a1e-a536-33297b0fe6b1 req-19234346-a788-4210-8db9-877d01836aa1 service nova] [instance: 363e48d7-8ed1-45d5-9443-d0b715952fca] Refreshing instance network info cache due to event network-changed-1c634ca3-05f4-4ac4-a523-27b6d5d64e60. {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1140.927233] env[61957]: DEBUG oslo_concurrency.lockutils [req-ac40299e-5fc4-4a1e-a536-33297b0fe6b1 req-19234346-a788-4210-8db9-877d01836aa1 service nova] Acquiring lock "refresh_cache-363e48d7-8ed1-45d5-9443-d0b715952fca" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1140.927233] env[61957]: DEBUG oslo_concurrency.lockutils [req-ac40299e-5fc4-4a1e-a536-33297b0fe6b1 req-19234346-a788-4210-8db9-877d01836aa1 service nova] Acquired lock "refresh_cache-363e48d7-8ed1-45d5-9443-d0b715952fca" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1140.927233] env[61957]: DEBUG nova.network.neutron [req-ac40299e-5fc4-4a1e-a536-33297b0fe6b1 req-19234346-a788-4210-8db9-877d01836aa1 service nova] [instance: 363e48d7-8ed1-45d5-9443-d0b715952fca] Refreshing network info cache for port 1c634ca3-05f4-4ac4-a523-27b6d5d64e60 {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1141.013518] env[61957]: DEBUG oslo_concurrency.lockutils [None req-024ba6d8-b0a9-4c02-8f92-8b55e68f2ea5 tempest-ServerAddressesNegativeTestJSON-995848772 tempest-ServerAddressesNegativeTestJSON-995848772-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1141.013797] env[61957]: DEBUG oslo_concurrency.lockutils [None req-024ba6d8-b0a9-4c02-8f92-8b55e68f2ea5 tempest-ServerAddressesNegativeTestJSON-995848772 tempest-ServerAddressesNegativeTestJSON-995848772-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1141.015386] env[61957]: INFO nova.compute.claims [None req-024ba6d8-b0a9-4c02-8f92-8b55e68f2ea5 tempest-ServerAddressesNegativeTestJSON-995848772 tempest-ServerAddressesNegativeTestJSON-995848772-project-member] [instance: 2fe04bcb-569c-48c3-b300-e67b4abe9fdf] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1141.066239] env[61957]: DEBUG nova.compute.manager [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] [instance: 9b0053fe-420e-4b92-86a8-ed44f6a6ec49] Instance has had 0 of 5 cleanup attempts {{(pid=61957) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1141.569325] env[61957]: DEBUG nova.compute.manager [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] [instance: 7e2db5f6-1419-415e-b957-66cf2e3ec19d] Instance has had 0 of 5 cleanup attempts {{(pid=61957) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1141.642076] env[61957]: DEBUG nova.network.neutron [req-ac40299e-5fc4-4a1e-a536-33297b0fe6b1 req-19234346-a788-4210-8db9-877d01836aa1 service nova] [instance: 363e48d7-8ed1-45d5-9443-d0b715952fca] Updated VIF entry in instance network info cache for port 1c634ca3-05f4-4ac4-a523-27b6d5d64e60. {{(pid=61957) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1141.642482] env[61957]: DEBUG nova.network.neutron [req-ac40299e-5fc4-4a1e-a536-33297b0fe6b1 req-19234346-a788-4210-8db9-877d01836aa1 service nova] [instance: 363e48d7-8ed1-45d5-9443-d0b715952fca] Updating instance_info_cache with network_info: [{"id": "1c634ca3-05f4-4ac4-a523-27b6d5d64e60", "address": "fa:16:3e:1f:40:85", "network": {"id": "a712bee6-193d-44ff-862a-fe63b0ee15ee", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-460165686-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "29f70feed7fe4f259bcc8aa8633f0aa8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9630cae2-7dd9-42b7-8b53-91ab254af243", "external-id": "nsx-vlan-transportzone-900", "segmentation_id": 900, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1c634ca3-05", "ovs_interfaceid": "1c634ca3-05f4-4ac4-a523-27b6d5d64e60", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1142.067563] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9dbe728-23ee-4e2a-9ccd-1dce348abe36 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.074876] env[61957]: DEBUG nova.compute.manager [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] [instance: b598b8bb-919e-4404-b264-7b76161b0f79] Instance has had 0 of 5 cleanup attempts {{(pid=61957) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1142.077916] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60000931-93b5-497d-8a80-d7cb4e096305 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.110813] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7aecdb24-4bd4-4696-9d15-f557938a76d0 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.118709] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f65f2f2-92fd-4ecc-a5b2-0993ae4e9a8f {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.132020] env[61957]: DEBUG nova.compute.provider_tree [None req-024ba6d8-b0a9-4c02-8f92-8b55e68f2ea5 tempest-ServerAddressesNegativeTestJSON-995848772 tempest-ServerAddressesNegativeTestJSON-995848772-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1142.144809] env[61957]: DEBUG oslo_concurrency.lockutils [req-ac40299e-5fc4-4a1e-a536-33297b0fe6b1 req-19234346-a788-4210-8db9-877d01836aa1 service nova] Releasing lock "refresh_cache-363e48d7-8ed1-45d5-9443-d0b715952fca" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1142.578096] env[61957]: DEBUG nova.compute.manager [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] [instance: d4674fbc-47b1-42d1-aaba-e86d46c51e8f] Instance has had 0 of 5 cleanup attempts {{(pid=61957) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1142.634968] env[61957]: DEBUG nova.scheduler.client.report [None req-024ba6d8-b0a9-4c02-8f92-8b55e68f2ea5 tempest-ServerAddressesNegativeTestJSON-995848772 tempest-ServerAddressesNegativeTestJSON-995848772-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1142.955015] env[61957]: DEBUG nova.compute.manager [req-6c461701-b708-4231-bb00-e10e61f1ea24 req-7d48bd1f-bcc8-4198-95d3-f954f5e3a51b service nova] [instance: 363e48d7-8ed1-45d5-9443-d0b715952fca] Received event network-changed-1c634ca3-05f4-4ac4-a523-27b6d5d64e60 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1142.955194] env[61957]: DEBUG nova.compute.manager [req-6c461701-b708-4231-bb00-e10e61f1ea24 req-7d48bd1f-bcc8-4198-95d3-f954f5e3a51b service nova] [instance: 363e48d7-8ed1-45d5-9443-d0b715952fca] Refreshing instance network info cache due to event network-changed-1c634ca3-05f4-4ac4-a523-27b6d5d64e60. {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1142.955419] env[61957]: DEBUG oslo_concurrency.lockutils [req-6c461701-b708-4231-bb00-e10e61f1ea24 req-7d48bd1f-bcc8-4198-95d3-f954f5e3a51b service nova] Acquiring lock "refresh_cache-363e48d7-8ed1-45d5-9443-d0b715952fca" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1142.955568] env[61957]: DEBUG oslo_concurrency.lockutils [req-6c461701-b708-4231-bb00-e10e61f1ea24 req-7d48bd1f-bcc8-4198-95d3-f954f5e3a51b service nova] Acquired lock "refresh_cache-363e48d7-8ed1-45d5-9443-d0b715952fca" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1142.955729] env[61957]: DEBUG nova.network.neutron [req-6c461701-b708-4231-bb00-e10e61f1ea24 req-7d48bd1f-bcc8-4198-95d3-f954f5e3a51b service nova] [instance: 363e48d7-8ed1-45d5-9443-d0b715952fca] Refreshing network info cache for port 1c634ca3-05f4-4ac4-a523-27b6d5d64e60 {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1143.082422] env[61957]: DEBUG nova.compute.manager [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] [instance: 19966b0a-53b7-48c5-849c-a9d00dc024f8] Instance has had 0 of 5 cleanup attempts {{(pid=61957) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1143.139301] env[61957]: DEBUG oslo_concurrency.lockutils [None req-024ba6d8-b0a9-4c02-8f92-8b55e68f2ea5 tempest-ServerAddressesNegativeTestJSON-995848772 tempest-ServerAddressesNegativeTestJSON-995848772-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.125s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1143.139790] env[61957]: DEBUG nova.compute.manager [None req-024ba6d8-b0a9-4c02-8f92-8b55e68f2ea5 tempest-ServerAddressesNegativeTestJSON-995848772 tempest-ServerAddressesNegativeTestJSON-995848772-project-member] [instance: 2fe04bcb-569c-48c3-b300-e67b4abe9fdf] Start building networks asynchronously for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1143.587784] env[61957]: DEBUG nova.compute.manager [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] [instance: 39b8b1df-efdd-4c22-9f31-85c85be4f0eb] Instance has had 0 of 5 cleanup attempts {{(pid=61957) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1143.640509] env[61957]: DEBUG nova.network.neutron [req-6c461701-b708-4231-bb00-e10e61f1ea24 req-7d48bd1f-bcc8-4198-95d3-f954f5e3a51b service nova] [instance: 363e48d7-8ed1-45d5-9443-d0b715952fca] Updated VIF entry in instance network info cache for port 1c634ca3-05f4-4ac4-a523-27b6d5d64e60. {{(pid=61957) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1143.640872] env[61957]: DEBUG nova.network.neutron [req-6c461701-b708-4231-bb00-e10e61f1ea24 req-7d48bd1f-bcc8-4198-95d3-f954f5e3a51b service nova] [instance: 363e48d7-8ed1-45d5-9443-d0b715952fca] Updating instance_info_cache with network_info: [{"id": "1c634ca3-05f4-4ac4-a523-27b6d5d64e60", "address": "fa:16:3e:1f:40:85", "network": {"id": "a712bee6-193d-44ff-862a-fe63b0ee15ee", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-460165686-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "29f70feed7fe4f259bcc8aa8633f0aa8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9630cae2-7dd9-42b7-8b53-91ab254af243", "external-id": "nsx-vlan-transportzone-900", "segmentation_id": 900, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1c634ca3-05", "ovs_interfaceid": "1c634ca3-05f4-4ac4-a523-27b6d5d64e60", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1143.644562] env[61957]: DEBUG nova.compute.utils [None req-024ba6d8-b0a9-4c02-8f92-8b55e68f2ea5 tempest-ServerAddressesNegativeTestJSON-995848772 tempest-ServerAddressesNegativeTestJSON-995848772-project-member] Using /dev/sd instead of None {{(pid=61957) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1143.645937] env[61957]: DEBUG nova.compute.manager [None req-024ba6d8-b0a9-4c02-8f92-8b55e68f2ea5 tempest-ServerAddressesNegativeTestJSON-995848772 tempest-ServerAddressesNegativeTestJSON-995848772-project-member] [instance: 2fe04bcb-569c-48c3-b300-e67b4abe9fdf] Allocating IP information in the background. {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1143.646113] env[61957]: DEBUG nova.network.neutron [None req-024ba6d8-b0a9-4c02-8f92-8b55e68f2ea5 tempest-ServerAddressesNegativeTestJSON-995848772 tempest-ServerAddressesNegativeTestJSON-995848772-project-member] [instance: 2fe04bcb-569c-48c3-b300-e67b4abe9fdf] allocate_for_instance() {{(pid=61957) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1143.684857] env[61957]: DEBUG nova.policy [None req-024ba6d8-b0a9-4c02-8f92-8b55e68f2ea5 tempest-ServerAddressesNegativeTestJSON-995848772 tempest-ServerAddressesNegativeTestJSON-995848772-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5fe29a3f418b463ba4e16c5086247567', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8da650621f9e41299eca9f3f3d07e085', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61957) authorize /opt/stack/nova/nova/policy.py:203}} [ 1143.936133] env[61957]: DEBUG nova.network.neutron [None req-024ba6d8-b0a9-4c02-8f92-8b55e68f2ea5 tempest-ServerAddressesNegativeTestJSON-995848772 tempest-ServerAddressesNegativeTestJSON-995848772-project-member] [instance: 2fe04bcb-569c-48c3-b300-e67b4abe9fdf] Successfully created port: ff49168a-f538-451b-84d1-a2efd419a3ba {{(pid=61957) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1144.089936] env[61957]: DEBUG nova.compute.manager [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] [instance: dbeeb200-70b9-4cb4-b5a4-182389d21918] Instance has had 0 of 5 cleanup attempts {{(pid=61957) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1144.143099] env[61957]: DEBUG oslo_concurrency.lockutils [req-6c461701-b708-4231-bb00-e10e61f1ea24 req-7d48bd1f-bcc8-4198-95d3-f954f5e3a51b service nova] Releasing lock "refresh_cache-363e48d7-8ed1-45d5-9443-d0b715952fca" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1144.143443] env[61957]: DEBUG nova.compute.manager [req-6c461701-b708-4231-bb00-e10e61f1ea24 req-7d48bd1f-bcc8-4198-95d3-f954f5e3a51b service nova] [instance: 363e48d7-8ed1-45d5-9443-d0b715952fca] Received event network-changed-1c634ca3-05f4-4ac4-a523-27b6d5d64e60 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1144.143619] env[61957]: DEBUG nova.compute.manager [req-6c461701-b708-4231-bb00-e10e61f1ea24 req-7d48bd1f-bcc8-4198-95d3-f954f5e3a51b service nova] [instance: 363e48d7-8ed1-45d5-9443-d0b715952fca] Refreshing instance network info cache due to event network-changed-1c634ca3-05f4-4ac4-a523-27b6d5d64e60. {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1144.143836] env[61957]: DEBUG oslo_concurrency.lockutils [req-6c461701-b708-4231-bb00-e10e61f1ea24 req-7d48bd1f-bcc8-4198-95d3-f954f5e3a51b service nova] Acquiring lock "refresh_cache-363e48d7-8ed1-45d5-9443-d0b715952fca" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1144.143990] env[61957]: DEBUG oslo_concurrency.lockutils [req-6c461701-b708-4231-bb00-e10e61f1ea24 req-7d48bd1f-bcc8-4198-95d3-f954f5e3a51b service nova] Acquired lock "refresh_cache-363e48d7-8ed1-45d5-9443-d0b715952fca" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1144.144174] env[61957]: DEBUG nova.network.neutron [req-6c461701-b708-4231-bb00-e10e61f1ea24 req-7d48bd1f-bcc8-4198-95d3-f954f5e3a51b service nova] [instance: 363e48d7-8ed1-45d5-9443-d0b715952fca] Refreshing network info cache for port 1c634ca3-05f4-4ac4-a523-27b6d5d64e60 {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1144.148619] env[61957]: DEBUG nova.compute.manager [None req-024ba6d8-b0a9-4c02-8f92-8b55e68f2ea5 tempest-ServerAddressesNegativeTestJSON-995848772 tempest-ServerAddressesNegativeTestJSON-995848772-project-member] [instance: 2fe04bcb-569c-48c3-b300-e67b4abe9fdf] Start building block device mappings for instance. {{(pid=61957) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1144.266689] env[61957]: DEBUG oslo_concurrency.lockutils [None req-b4f9918b-3627-43df-a7c4-9b2be3813404 tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] Acquiring lock "363e48d7-8ed1-45d5-9443-d0b715952fca" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1144.266953] env[61957]: DEBUG oslo_concurrency.lockutils [None req-b4f9918b-3627-43df-a7c4-9b2be3813404 tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] Lock "363e48d7-8ed1-45d5-9443-d0b715952fca" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1144.267191] env[61957]: DEBUG oslo_concurrency.lockutils [None req-b4f9918b-3627-43df-a7c4-9b2be3813404 tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] Acquiring lock "363e48d7-8ed1-45d5-9443-d0b715952fca-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1144.268039] env[61957]: DEBUG oslo_concurrency.lockutils [None req-b4f9918b-3627-43df-a7c4-9b2be3813404 tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] Lock "363e48d7-8ed1-45d5-9443-d0b715952fca-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1144.268039] env[61957]: DEBUG oslo_concurrency.lockutils [None req-b4f9918b-3627-43df-a7c4-9b2be3813404 tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] Lock "363e48d7-8ed1-45d5-9443-d0b715952fca-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1144.269580] env[61957]: INFO nova.compute.manager [None req-b4f9918b-3627-43df-a7c4-9b2be3813404 tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] [instance: 363e48d7-8ed1-45d5-9443-d0b715952fca] Terminating instance [ 1144.271429] env[61957]: DEBUG nova.compute.manager [None req-b4f9918b-3627-43df-a7c4-9b2be3813404 tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] [instance: 363e48d7-8ed1-45d5-9443-d0b715952fca] Start destroying the instance on the hypervisor. {{(pid=61957) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1144.271633] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-b4f9918b-3627-43df-a7c4-9b2be3813404 tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] [instance: 363e48d7-8ed1-45d5-9443-d0b715952fca] Destroying instance {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1144.272477] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca8e6e41-e590-4414-b60f-479978d4b6f3 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.280511] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-b4f9918b-3627-43df-a7c4-9b2be3813404 tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] [instance: 363e48d7-8ed1-45d5-9443-d0b715952fca] Powering off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1144.280755] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-62f1cc50-c894-44a3-82fc-dd52155d2c30 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.287451] env[61957]: DEBUG oslo_vmware.api [None req-b4f9918b-3627-43df-a7c4-9b2be3813404 tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] Waiting for the task: (returnval){ [ 1144.287451] env[61957]: value = "task-1278149" [ 1144.287451] env[61957]: _type = "Task" [ 1144.287451] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1144.295059] env[61957]: DEBUG oslo_vmware.api [None req-b4f9918b-3627-43df-a7c4-9b2be3813404 tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] Task: {'id': task-1278149, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1144.593032] env[61957]: DEBUG nova.compute.manager [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] [instance: 736eabcf-70f8-4e1e-9830-270d9971fc9a] Instance has had 0 of 5 cleanup attempts {{(pid=61957) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1144.797264] env[61957]: DEBUG oslo_vmware.api [None req-b4f9918b-3627-43df-a7c4-9b2be3813404 tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] Task: {'id': task-1278149, 'name': PowerOffVM_Task, 'duration_secs': 0.189288} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1144.797620] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-b4f9918b-3627-43df-a7c4-9b2be3813404 tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] [instance: 363e48d7-8ed1-45d5-9443-d0b715952fca] Powered off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1144.797721] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-b4f9918b-3627-43df-a7c4-9b2be3813404 tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] [instance: 363e48d7-8ed1-45d5-9443-d0b715952fca] Unregistering the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1144.797968] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-11470fa0-47ec-495d-925d-97179483c4ff {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.823137] env[61957]: DEBUG nova.network.neutron [req-6c461701-b708-4231-bb00-e10e61f1ea24 req-7d48bd1f-bcc8-4198-95d3-f954f5e3a51b service nova] [instance: 363e48d7-8ed1-45d5-9443-d0b715952fca] Updated VIF entry in instance network info cache for port 1c634ca3-05f4-4ac4-a523-27b6d5d64e60. {{(pid=61957) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1144.823525] env[61957]: DEBUG nova.network.neutron [req-6c461701-b708-4231-bb00-e10e61f1ea24 req-7d48bd1f-bcc8-4198-95d3-f954f5e3a51b service nova] [instance: 363e48d7-8ed1-45d5-9443-d0b715952fca] Updating instance_info_cache with network_info: [{"id": "1c634ca3-05f4-4ac4-a523-27b6d5d64e60", "address": "fa:16:3e:1f:40:85", "network": {"id": "a712bee6-193d-44ff-862a-fe63b0ee15ee", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-460165686-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "29f70feed7fe4f259bcc8aa8633f0aa8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9630cae2-7dd9-42b7-8b53-91ab254af243", "external-id": "nsx-vlan-transportzone-900", "segmentation_id": 900, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1c634ca3-05", "ovs_interfaceid": "1c634ca3-05f4-4ac4-a523-27b6d5d64e60", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1144.869299] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-b4f9918b-3627-43df-a7c4-9b2be3813404 tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] [instance: 363e48d7-8ed1-45d5-9443-d0b715952fca] Unregistered the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1144.869535] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-b4f9918b-3627-43df-a7c4-9b2be3813404 tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] [instance: 363e48d7-8ed1-45d5-9443-d0b715952fca] Deleting contents of the VM from datastore datastore2 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1144.869719] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-b4f9918b-3627-43df-a7c4-9b2be3813404 tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] Deleting the datastore file [datastore2] 363e48d7-8ed1-45d5-9443-d0b715952fca {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1144.870025] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-89c20edc-c468-401b-88c0-c7c40e3905b3 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.875916] env[61957]: DEBUG oslo_vmware.api [None req-b4f9918b-3627-43df-a7c4-9b2be3813404 tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] Waiting for the task: (returnval){ [ 1144.875916] env[61957]: value = "task-1278151" [ 1144.875916] env[61957]: _type = "Task" [ 1144.875916] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1144.883576] env[61957]: DEBUG oslo_vmware.api [None req-b4f9918b-3627-43df-a7c4-9b2be3813404 tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] Task: {'id': task-1278151, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1144.996301] env[61957]: DEBUG nova.compute.manager [req-570b1395-46d0-4f04-91c3-1fb4c2d56e35 req-2ce3c41e-b94a-4261-995c-b9984b2e9cc8 service nova] [instance: 363e48d7-8ed1-45d5-9443-d0b715952fca] Received event network-changed-1c634ca3-05f4-4ac4-a523-27b6d5d64e60 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1144.996472] env[61957]: DEBUG nova.compute.manager [req-570b1395-46d0-4f04-91c3-1fb4c2d56e35 req-2ce3c41e-b94a-4261-995c-b9984b2e9cc8 service nova] [instance: 363e48d7-8ed1-45d5-9443-d0b715952fca] Refreshing instance network info cache due to event network-changed-1c634ca3-05f4-4ac4-a523-27b6d5d64e60. {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1144.996673] env[61957]: DEBUG oslo_concurrency.lockutils [req-570b1395-46d0-4f04-91c3-1fb4c2d56e35 req-2ce3c41e-b94a-4261-995c-b9984b2e9cc8 service nova] Acquiring lock "refresh_cache-363e48d7-8ed1-45d5-9443-d0b715952fca" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1145.096834] env[61957]: DEBUG nova.compute.manager [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] [instance: a8ca1ad6-d636-4fa2-b0ac-53b020e392ce] Instance has had 0 of 5 cleanup attempts {{(pid=61957) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1145.157795] env[61957]: DEBUG nova.compute.manager [None req-024ba6d8-b0a9-4c02-8f92-8b55e68f2ea5 tempest-ServerAddressesNegativeTestJSON-995848772 tempest-ServerAddressesNegativeTestJSON-995848772-project-member] [instance: 2fe04bcb-569c-48c3-b300-e67b4abe9fdf] Start spawning the instance on the hypervisor. {{(pid=61957) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1145.183046] env[61957]: DEBUG nova.virt.hardware [None req-024ba6d8-b0a9-4c02-8f92-8b55e68f2ea5 tempest-ServerAddressesNegativeTestJSON-995848772 tempest-ServerAddressesNegativeTestJSON-995848772-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T17:22:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T17:21:50Z,direct_url=,disk_format='vmdk',id=11c76a2c-f705-470a-ba9d-4657858bab38,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='81d3096d0f094373913fc3dc8f5c3c6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T17:21:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1145.183345] env[61957]: DEBUG nova.virt.hardware [None req-024ba6d8-b0a9-4c02-8f92-8b55e68f2ea5 tempest-ServerAddressesNegativeTestJSON-995848772 tempest-ServerAddressesNegativeTestJSON-995848772-project-member] Flavor limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1145.183519] env[61957]: DEBUG nova.virt.hardware [None req-024ba6d8-b0a9-4c02-8f92-8b55e68f2ea5 tempest-ServerAddressesNegativeTestJSON-995848772 tempest-ServerAddressesNegativeTestJSON-995848772-project-member] Image limits 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1145.183710] env[61957]: DEBUG nova.virt.hardware [None req-024ba6d8-b0a9-4c02-8f92-8b55e68f2ea5 tempest-ServerAddressesNegativeTestJSON-995848772 tempest-ServerAddressesNegativeTestJSON-995848772-project-member] Flavor pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1145.183952] env[61957]: DEBUG nova.virt.hardware [None req-024ba6d8-b0a9-4c02-8f92-8b55e68f2ea5 tempest-ServerAddressesNegativeTestJSON-995848772 tempest-ServerAddressesNegativeTestJSON-995848772-project-member] Image pref 0:0:0 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1145.184148] env[61957]: DEBUG nova.virt.hardware [None req-024ba6d8-b0a9-4c02-8f92-8b55e68f2ea5 tempest-ServerAddressesNegativeTestJSON-995848772 tempest-ServerAddressesNegativeTestJSON-995848772-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61957) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1145.184387] env[61957]: DEBUG nova.virt.hardware [None req-024ba6d8-b0a9-4c02-8f92-8b55e68f2ea5 tempest-ServerAddressesNegativeTestJSON-995848772 tempest-ServerAddressesNegativeTestJSON-995848772-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1145.184570] env[61957]: DEBUG nova.virt.hardware [None req-024ba6d8-b0a9-4c02-8f92-8b55e68f2ea5 tempest-ServerAddressesNegativeTestJSON-995848772 tempest-ServerAddressesNegativeTestJSON-995848772-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1145.184746] env[61957]: DEBUG nova.virt.hardware [None req-024ba6d8-b0a9-4c02-8f92-8b55e68f2ea5 tempest-ServerAddressesNegativeTestJSON-995848772 tempest-ServerAddressesNegativeTestJSON-995848772-project-member] Got 1 possible topologies {{(pid=61957) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1145.184903] env[61957]: DEBUG nova.virt.hardware [None req-024ba6d8-b0a9-4c02-8f92-8b55e68f2ea5 tempest-ServerAddressesNegativeTestJSON-995848772 tempest-ServerAddressesNegativeTestJSON-995848772-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1145.185090] env[61957]: DEBUG nova.virt.hardware [None req-024ba6d8-b0a9-4c02-8f92-8b55e68f2ea5 tempest-ServerAddressesNegativeTestJSON-995848772 tempest-ServerAddressesNegativeTestJSON-995848772-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61957) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1145.186014] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1249b5dc-fa6d-4d2d-a7a8-b05879281c17 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.194024] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b34d7aa-274e-4b9f-9c27-7e1130966003 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.304419] env[61957]: DEBUG nova.compute.manager [req-51762159-afa2-4b1c-b464-83757a66e209 req-9828dc4f-6c26-46a0-9cff-3f5269d25b15 service nova] [instance: 2fe04bcb-569c-48c3-b300-e67b4abe9fdf] Received event network-vif-plugged-ff49168a-f538-451b-84d1-a2efd419a3ba {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1145.304678] env[61957]: DEBUG oslo_concurrency.lockutils [req-51762159-afa2-4b1c-b464-83757a66e209 req-9828dc4f-6c26-46a0-9cff-3f5269d25b15 service nova] Acquiring lock "2fe04bcb-569c-48c3-b300-e67b4abe9fdf-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1145.304859] env[61957]: DEBUG oslo_concurrency.lockutils [req-51762159-afa2-4b1c-b464-83757a66e209 req-9828dc4f-6c26-46a0-9cff-3f5269d25b15 service nova] Lock "2fe04bcb-569c-48c3-b300-e67b4abe9fdf-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1145.305018] env[61957]: DEBUG oslo_concurrency.lockutils [req-51762159-afa2-4b1c-b464-83757a66e209 req-9828dc4f-6c26-46a0-9cff-3f5269d25b15 service nova] Lock "2fe04bcb-569c-48c3-b300-e67b4abe9fdf-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1145.305199] env[61957]: DEBUG nova.compute.manager [req-51762159-afa2-4b1c-b464-83757a66e209 req-9828dc4f-6c26-46a0-9cff-3f5269d25b15 service nova] [instance: 2fe04bcb-569c-48c3-b300-e67b4abe9fdf] No waiting events found dispatching network-vif-plugged-ff49168a-f538-451b-84d1-a2efd419a3ba {{(pid=61957) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1145.305370] env[61957]: WARNING nova.compute.manager [req-51762159-afa2-4b1c-b464-83757a66e209 req-9828dc4f-6c26-46a0-9cff-3f5269d25b15 service nova] [instance: 2fe04bcb-569c-48c3-b300-e67b4abe9fdf] Received unexpected event network-vif-plugged-ff49168a-f538-451b-84d1-a2efd419a3ba for instance with vm_state building and task_state spawning. [ 1145.326012] env[61957]: DEBUG oslo_concurrency.lockutils [req-6c461701-b708-4231-bb00-e10e61f1ea24 req-7d48bd1f-bcc8-4198-95d3-f954f5e3a51b service nova] Releasing lock "refresh_cache-363e48d7-8ed1-45d5-9443-d0b715952fca" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1145.326360] env[61957]: DEBUG oslo_concurrency.lockutils [req-570b1395-46d0-4f04-91c3-1fb4c2d56e35 req-2ce3c41e-b94a-4261-995c-b9984b2e9cc8 service nova] Acquired lock "refresh_cache-363e48d7-8ed1-45d5-9443-d0b715952fca" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1145.326547] env[61957]: DEBUG nova.network.neutron [req-570b1395-46d0-4f04-91c3-1fb4c2d56e35 req-2ce3c41e-b94a-4261-995c-b9984b2e9cc8 service nova] [instance: 363e48d7-8ed1-45d5-9443-d0b715952fca] Refreshing network info cache for port 1c634ca3-05f4-4ac4-a523-27b6d5d64e60 {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1145.385762] env[61957]: DEBUG oslo_vmware.api [None req-b4f9918b-3627-43df-a7c4-9b2be3813404 tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] Task: {'id': task-1278151, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.17248} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1145.386571] env[61957]: DEBUG nova.network.neutron [None req-024ba6d8-b0a9-4c02-8f92-8b55e68f2ea5 tempest-ServerAddressesNegativeTestJSON-995848772 tempest-ServerAddressesNegativeTestJSON-995848772-project-member] [instance: 2fe04bcb-569c-48c3-b300-e67b4abe9fdf] Successfully updated port: ff49168a-f538-451b-84d1-a2efd419a3ba {{(pid=61957) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1145.387801] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-b4f9918b-3627-43df-a7c4-9b2be3813404 tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] Deleted the datastore file {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1145.388118] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-b4f9918b-3627-43df-a7c4-9b2be3813404 tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] [instance: 363e48d7-8ed1-45d5-9443-d0b715952fca] Deleted contents of the VM from datastore datastore2 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1145.388409] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-b4f9918b-3627-43df-a7c4-9b2be3813404 tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] [instance: 363e48d7-8ed1-45d5-9443-d0b715952fca] Instance destroyed {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1145.388639] env[61957]: INFO nova.compute.manager [None req-b4f9918b-3627-43df-a7c4-9b2be3813404 tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] [instance: 363e48d7-8ed1-45d5-9443-d0b715952fca] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1145.388880] env[61957]: DEBUG oslo.service.loopingcall [None req-b4f9918b-3627-43df-a7c4-9b2be3813404 tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1145.389279] env[61957]: DEBUG nova.compute.manager [-] [instance: 363e48d7-8ed1-45d5-9443-d0b715952fca] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1145.389430] env[61957]: DEBUG nova.network.neutron [-] [instance: 363e48d7-8ed1-45d5-9443-d0b715952fca] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1145.599930] env[61957]: DEBUG nova.compute.manager [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] [instance: 0d0f2d34-de35-4e80-8d9f-12693add0786] Instance has had 0 of 5 cleanup attempts {{(pid=61957) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1145.845274] env[61957]: INFO nova.network.neutron [req-570b1395-46d0-4f04-91c3-1fb4c2d56e35 req-2ce3c41e-b94a-4261-995c-b9984b2e9cc8 service nova] [instance: 363e48d7-8ed1-45d5-9443-d0b715952fca] Port 1c634ca3-05f4-4ac4-a523-27b6d5d64e60 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 1145.845615] env[61957]: DEBUG nova.network.neutron [req-570b1395-46d0-4f04-91c3-1fb4c2d56e35 req-2ce3c41e-b94a-4261-995c-b9984b2e9cc8 service nova] [instance: 363e48d7-8ed1-45d5-9443-d0b715952fca] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1145.889935] env[61957]: DEBUG oslo_concurrency.lockutils [None req-024ba6d8-b0a9-4c02-8f92-8b55e68f2ea5 tempest-ServerAddressesNegativeTestJSON-995848772 tempest-ServerAddressesNegativeTestJSON-995848772-project-member] Acquiring lock "refresh_cache-2fe04bcb-569c-48c3-b300-e67b4abe9fdf" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1145.890195] env[61957]: DEBUG oslo_concurrency.lockutils [None req-024ba6d8-b0a9-4c02-8f92-8b55e68f2ea5 tempest-ServerAddressesNegativeTestJSON-995848772 tempest-ServerAddressesNegativeTestJSON-995848772-project-member] Acquired lock "refresh_cache-2fe04bcb-569c-48c3-b300-e67b4abe9fdf" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1145.890441] env[61957]: DEBUG nova.network.neutron [None req-024ba6d8-b0a9-4c02-8f92-8b55e68f2ea5 tempest-ServerAddressesNegativeTestJSON-995848772 tempest-ServerAddressesNegativeTestJSON-995848772-project-member] [instance: 2fe04bcb-569c-48c3-b300-e67b4abe9fdf] Building network info cache for instance {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1146.098362] env[61957]: DEBUG nova.network.neutron [-] [instance: 363e48d7-8ed1-45d5-9443-d0b715952fca] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1146.102240] env[61957]: DEBUG nova.compute.manager [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] [instance: d5d10a31-0e78-4ed7-b944-9208138a4861] Instance has had 0 of 5 cleanup attempts {{(pid=61957) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1146.348258] env[61957]: DEBUG oslo_concurrency.lockutils [req-570b1395-46d0-4f04-91c3-1fb4c2d56e35 req-2ce3c41e-b94a-4261-995c-b9984b2e9cc8 service nova] Releasing lock "refresh_cache-363e48d7-8ed1-45d5-9443-d0b715952fca" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1146.425337] env[61957]: DEBUG nova.network.neutron [None req-024ba6d8-b0a9-4c02-8f92-8b55e68f2ea5 tempest-ServerAddressesNegativeTestJSON-995848772 tempest-ServerAddressesNegativeTestJSON-995848772-project-member] [instance: 2fe04bcb-569c-48c3-b300-e67b4abe9fdf] Instance cache missing network info. {{(pid=61957) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1146.552550] env[61957]: DEBUG nova.network.neutron [None req-024ba6d8-b0a9-4c02-8f92-8b55e68f2ea5 tempest-ServerAddressesNegativeTestJSON-995848772 tempest-ServerAddressesNegativeTestJSON-995848772-project-member] [instance: 2fe04bcb-569c-48c3-b300-e67b4abe9fdf] Updating instance_info_cache with network_info: [{"id": "ff49168a-f538-451b-84d1-a2efd419a3ba", "address": "fa:16:3e:35:44:05", "network": {"id": "18f21039-d87d-4acb-9963-ba9853e797b3", "bridge": "br-int", "label": "tempest-ServerAddressesNegativeTestJSON-1597681559-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8da650621f9e41299eca9f3f3d07e085", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "790c811b-3fa6-49f8-87ac-c51450911137", "external-id": "nsx-vlan-transportzone-908", "segmentation_id": 908, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapff49168a-f5", "ovs_interfaceid": "ff49168a-f538-451b-84d1-a2efd419a3ba", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1146.601051] env[61957]: INFO nova.compute.manager [-] [instance: 363e48d7-8ed1-45d5-9443-d0b715952fca] Took 1.21 seconds to deallocate network for instance. [ 1146.605455] env[61957]: DEBUG nova.compute.manager [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] [instance: 28ab7d23-487f-4ae1-8fe4-58db55b59918] Instance has had 0 of 5 cleanup attempts {{(pid=61957) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1147.055659] env[61957]: DEBUG oslo_concurrency.lockutils [None req-024ba6d8-b0a9-4c02-8f92-8b55e68f2ea5 tempest-ServerAddressesNegativeTestJSON-995848772 tempest-ServerAddressesNegativeTestJSON-995848772-project-member] Releasing lock "refresh_cache-2fe04bcb-569c-48c3-b300-e67b4abe9fdf" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1147.056048] env[61957]: DEBUG nova.compute.manager [None req-024ba6d8-b0a9-4c02-8f92-8b55e68f2ea5 tempest-ServerAddressesNegativeTestJSON-995848772 tempest-ServerAddressesNegativeTestJSON-995848772-project-member] [instance: 2fe04bcb-569c-48c3-b300-e67b4abe9fdf] Instance network_info: |[{"id": "ff49168a-f538-451b-84d1-a2efd419a3ba", "address": "fa:16:3e:35:44:05", "network": {"id": "18f21039-d87d-4acb-9963-ba9853e797b3", "bridge": "br-int", "label": "tempest-ServerAddressesNegativeTestJSON-1597681559-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8da650621f9e41299eca9f3f3d07e085", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "790c811b-3fa6-49f8-87ac-c51450911137", "external-id": "nsx-vlan-transportzone-908", "segmentation_id": 908, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapff49168a-f5", "ovs_interfaceid": "ff49168a-f538-451b-84d1-a2efd419a3ba", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61957) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1147.056514] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-024ba6d8-b0a9-4c02-8f92-8b55e68f2ea5 tempest-ServerAddressesNegativeTestJSON-995848772 tempest-ServerAddressesNegativeTestJSON-995848772-project-member] [instance: 2fe04bcb-569c-48c3-b300-e67b4abe9fdf] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:35:44:05', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '790c811b-3fa6-49f8-87ac-c51450911137', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ff49168a-f538-451b-84d1-a2efd419a3ba', 'vif_model': 'vmxnet3'}] {{(pid=61957) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1147.063951] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-024ba6d8-b0a9-4c02-8f92-8b55e68f2ea5 tempest-ServerAddressesNegativeTestJSON-995848772 tempest-ServerAddressesNegativeTestJSON-995848772-project-member] Creating folder: Project (8da650621f9e41299eca9f3f3d07e085). Parent ref: group-v274445. {{(pid=61957) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1147.064240] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-fb03111b-3744-4d18-884f-42c3e23ff85e {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.076193] env[61957]: INFO nova.virt.vmwareapi.vm_util [None req-024ba6d8-b0a9-4c02-8f92-8b55e68f2ea5 tempest-ServerAddressesNegativeTestJSON-995848772 tempest-ServerAddressesNegativeTestJSON-995848772-project-member] Created folder: Project (8da650621f9e41299eca9f3f3d07e085) in parent group-v274445. [ 1147.076414] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-024ba6d8-b0a9-4c02-8f92-8b55e68f2ea5 tempest-ServerAddressesNegativeTestJSON-995848772 tempest-ServerAddressesNegativeTestJSON-995848772-project-member] Creating folder: Instances. Parent ref: group-v274613. {{(pid=61957) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1147.076658] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9ad731fd-fa02-4312-bc6f-c990597280f7 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.085471] env[61957]: INFO nova.virt.vmwareapi.vm_util [None req-024ba6d8-b0a9-4c02-8f92-8b55e68f2ea5 tempest-ServerAddressesNegativeTestJSON-995848772 tempest-ServerAddressesNegativeTestJSON-995848772-project-member] Created folder: Instances in parent group-v274613. [ 1147.085703] env[61957]: DEBUG oslo.service.loopingcall [None req-024ba6d8-b0a9-4c02-8f92-8b55e68f2ea5 tempest-ServerAddressesNegativeTestJSON-995848772 tempest-ServerAddressesNegativeTestJSON-995848772-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1147.085890] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2fe04bcb-569c-48c3-b300-e67b4abe9fdf] Creating VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1147.086102] env[61957]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ea0bac72-3a2d-4023-aab2-5e8e94ca1926 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.103645] env[61957]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1147.103645] env[61957]: value = "task-1278154" [ 1147.103645] env[61957]: _type = "Task" [ 1147.103645] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1147.107715] env[61957]: DEBUG oslo_concurrency.lockutils [None req-b4f9918b-3627-43df-a7c4-9b2be3813404 tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1147.107972] env[61957]: DEBUG oslo_concurrency.lockutils [None req-b4f9918b-3627-43df-a7c4-9b2be3813404 tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1147.108231] env[61957]: DEBUG nova.objects.instance [None req-b4f9918b-3627-43df-a7c4-9b2be3813404 tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] Lazy-loading 'resources' on Instance uuid 363e48d7-8ed1-45d5-9443-d0b715952fca {{(pid=61957) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1147.111911] env[61957]: DEBUG nova.compute.manager [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] [instance: 1a0327c2-5671-4970-9db7-c7cc912d8678] Instance has had 0 of 5 cleanup attempts {{(pid=61957) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1147.113446] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1278154, 'name': CreateVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1147.332725] env[61957]: DEBUG nova.compute.manager [req-10488da0-9bf6-41ec-a72c-2bea148420d4 req-f3d70f83-4d6b-493b-b7e9-e2b65558a3c7 service nova] [instance: 2fe04bcb-569c-48c3-b300-e67b4abe9fdf] Received event network-changed-ff49168a-f538-451b-84d1-a2efd419a3ba {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1147.332962] env[61957]: DEBUG nova.compute.manager [req-10488da0-9bf6-41ec-a72c-2bea148420d4 req-f3d70f83-4d6b-493b-b7e9-e2b65558a3c7 service nova] [instance: 2fe04bcb-569c-48c3-b300-e67b4abe9fdf] Refreshing instance network info cache due to event network-changed-ff49168a-f538-451b-84d1-a2efd419a3ba. {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1147.333262] env[61957]: DEBUG oslo_concurrency.lockutils [req-10488da0-9bf6-41ec-a72c-2bea148420d4 req-f3d70f83-4d6b-493b-b7e9-e2b65558a3c7 service nova] Acquiring lock "refresh_cache-2fe04bcb-569c-48c3-b300-e67b4abe9fdf" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1147.333420] env[61957]: DEBUG oslo_concurrency.lockutils [req-10488da0-9bf6-41ec-a72c-2bea148420d4 req-f3d70f83-4d6b-493b-b7e9-e2b65558a3c7 service nova] Acquired lock "refresh_cache-2fe04bcb-569c-48c3-b300-e67b4abe9fdf" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1147.333607] env[61957]: DEBUG nova.network.neutron [req-10488da0-9bf6-41ec-a72c-2bea148420d4 req-f3d70f83-4d6b-493b-b7e9-e2b65558a3c7 service nova] [instance: 2fe04bcb-569c-48c3-b300-e67b4abe9fdf] Refreshing network info cache for port ff49168a-f538-451b-84d1-a2efd419a3ba {{(pid=61957) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1147.613689] env[61957]: DEBUG oslo_vmware.api [-] Task: {'id': task-1278154, 'name': CreateVM_Task, 'duration_secs': 0.346395} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1147.614160] env[61957]: DEBUG nova.compute.manager [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] [instance: e40fa112-4648-428f-a403-b3e3b8319ea8] Instance has had 0 of 5 cleanup attempts {{(pid=61957) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1147.615824] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2fe04bcb-569c-48c3-b300-e67b4abe9fdf] Created VM on the ESX host {{(pid=61957) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1147.618682] env[61957]: DEBUG oslo_concurrency.lockutils [None req-024ba6d8-b0a9-4c02-8f92-8b55e68f2ea5 tempest-ServerAddressesNegativeTestJSON-995848772 tempest-ServerAddressesNegativeTestJSON-995848772-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1147.618855] env[61957]: DEBUG oslo_concurrency.lockutils [None req-024ba6d8-b0a9-4c02-8f92-8b55e68f2ea5 tempest-ServerAddressesNegativeTestJSON-995848772 tempest-ServerAddressesNegativeTestJSON-995848772-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1147.619185] env[61957]: DEBUG oslo_concurrency.lockutils [None req-024ba6d8-b0a9-4c02-8f92-8b55e68f2ea5 tempest-ServerAddressesNegativeTestJSON-995848772 tempest-ServerAddressesNegativeTestJSON-995848772-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1147.619987] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8420e8f6-4461-4fb4-b6fb-65ed0163e600 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.624875] env[61957]: DEBUG oslo_vmware.api [None req-024ba6d8-b0a9-4c02-8f92-8b55e68f2ea5 tempest-ServerAddressesNegativeTestJSON-995848772 tempest-ServerAddressesNegativeTestJSON-995848772-project-member] Waiting for the task: (returnval){ [ 1147.624875] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52cb7d3a-4cfe-a6ea-6cda-3b2317b5196a" [ 1147.624875] env[61957]: _type = "Task" [ 1147.624875] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1147.632432] env[61957]: DEBUG oslo_vmware.api [None req-024ba6d8-b0a9-4c02-8f92-8b55e68f2ea5 tempest-ServerAddressesNegativeTestJSON-995848772 tempest-ServerAddressesNegativeTestJSON-995848772-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52cb7d3a-4cfe-a6ea-6cda-3b2317b5196a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1147.658524] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50e3e45c-de5d-42a8-8455-ebc2fbb5ed06 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.665866] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdf7fefd-ba9b-42eb-9eeb-aff69f3a229b {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.696181] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eda2c574-5cf0-4b27-bee9-df63a69ea311 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.703303] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e78c9839-bfab-4cc9-9546-021749c6e952 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.716271] env[61957]: DEBUG nova.compute.provider_tree [None req-b4f9918b-3627-43df-a7c4-9b2be3813404 tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1148.018561] env[61957]: DEBUG nova.network.neutron [req-10488da0-9bf6-41ec-a72c-2bea148420d4 req-f3d70f83-4d6b-493b-b7e9-e2b65558a3c7 service nova] [instance: 2fe04bcb-569c-48c3-b300-e67b4abe9fdf] Updated VIF entry in instance network info cache for port ff49168a-f538-451b-84d1-a2efd419a3ba. {{(pid=61957) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1148.018728] env[61957]: DEBUG nova.network.neutron [req-10488da0-9bf6-41ec-a72c-2bea148420d4 req-f3d70f83-4d6b-493b-b7e9-e2b65558a3c7 service nova] [instance: 2fe04bcb-569c-48c3-b300-e67b4abe9fdf] Updating instance_info_cache with network_info: [{"id": "ff49168a-f538-451b-84d1-a2efd419a3ba", "address": "fa:16:3e:35:44:05", "network": {"id": "18f21039-d87d-4acb-9963-ba9853e797b3", "bridge": "br-int", "label": "tempest-ServerAddressesNegativeTestJSON-1597681559-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8da650621f9e41299eca9f3f3d07e085", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "790c811b-3fa6-49f8-87ac-c51450911137", "external-id": "nsx-vlan-transportzone-908", "segmentation_id": 908, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapff49168a-f5", "ovs_interfaceid": "ff49168a-f538-451b-84d1-a2efd419a3ba", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1148.120313] env[61957]: DEBUG nova.compute.manager [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] [instance: 2e4a5344-600f-4b61-826e-c15f96b50af2] Instance has had 0 of 5 cleanup attempts {{(pid=61957) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1148.134848] env[61957]: DEBUG oslo_vmware.api [None req-024ba6d8-b0a9-4c02-8f92-8b55e68f2ea5 tempest-ServerAddressesNegativeTestJSON-995848772 tempest-ServerAddressesNegativeTestJSON-995848772-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52cb7d3a-4cfe-a6ea-6cda-3b2317b5196a, 'name': SearchDatastore_Task, 'duration_secs': 0.008949} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1148.135132] env[61957]: DEBUG oslo_concurrency.lockutils [None req-024ba6d8-b0a9-4c02-8f92-8b55e68f2ea5 tempest-ServerAddressesNegativeTestJSON-995848772 tempest-ServerAddressesNegativeTestJSON-995848772-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1148.135602] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-024ba6d8-b0a9-4c02-8f92-8b55e68f2ea5 tempest-ServerAddressesNegativeTestJSON-995848772 tempest-ServerAddressesNegativeTestJSON-995848772-project-member] [instance: 2fe04bcb-569c-48c3-b300-e67b4abe9fdf] Processing image 11c76a2c-f705-470a-ba9d-4657858bab38 {{(pid=61957) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1148.135602] env[61957]: DEBUG oslo_concurrency.lockutils [None req-024ba6d8-b0a9-4c02-8f92-8b55e68f2ea5 tempest-ServerAddressesNegativeTestJSON-995848772 tempest-ServerAddressesNegativeTestJSON-995848772-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1148.135712] env[61957]: DEBUG oslo_concurrency.lockutils [None req-024ba6d8-b0a9-4c02-8f92-8b55e68f2ea5 tempest-ServerAddressesNegativeTestJSON-995848772 tempest-ServerAddressesNegativeTestJSON-995848772-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1148.135894] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-024ba6d8-b0a9-4c02-8f92-8b55e68f2ea5 tempest-ServerAddressesNegativeTestJSON-995848772 tempest-ServerAddressesNegativeTestJSON-995848772-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1148.136161] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f1098b16-3999-46fc-80ce-0c5d2c09e057 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.144297] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-024ba6d8-b0a9-4c02-8f92-8b55e68f2ea5 tempest-ServerAddressesNegativeTestJSON-995848772 tempest-ServerAddressesNegativeTestJSON-995848772-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61957) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1148.144495] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-024ba6d8-b0a9-4c02-8f92-8b55e68f2ea5 tempest-ServerAddressesNegativeTestJSON-995848772 tempest-ServerAddressesNegativeTestJSON-995848772-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61957) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1148.145409] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2f58d555-d76f-4bc3-94ae-5f9ed9140570 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.150618] env[61957]: DEBUG oslo_vmware.api [None req-024ba6d8-b0a9-4c02-8f92-8b55e68f2ea5 tempest-ServerAddressesNegativeTestJSON-995848772 tempest-ServerAddressesNegativeTestJSON-995848772-project-member] Waiting for the task: (returnval){ [ 1148.150618] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52326222-7423-4119-159b-46a6d73d1ceb" [ 1148.150618] env[61957]: _type = "Task" [ 1148.150618] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1148.158070] env[61957]: DEBUG oslo_vmware.api [None req-024ba6d8-b0a9-4c02-8f92-8b55e68f2ea5 tempest-ServerAddressesNegativeTestJSON-995848772 tempest-ServerAddressesNegativeTestJSON-995848772-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52326222-7423-4119-159b-46a6d73d1ceb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1148.219319] env[61957]: DEBUG nova.scheduler.client.report [None req-b4f9918b-3627-43df-a7c4-9b2be3813404 tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1148.521258] env[61957]: DEBUG oslo_concurrency.lockutils [req-10488da0-9bf6-41ec-a72c-2bea148420d4 req-f3d70f83-4d6b-493b-b7e9-e2b65558a3c7 service nova] Releasing lock "refresh_cache-2fe04bcb-569c-48c3-b300-e67b4abe9fdf" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1148.521538] env[61957]: DEBUG nova.compute.manager [req-10488da0-9bf6-41ec-a72c-2bea148420d4 req-f3d70f83-4d6b-493b-b7e9-e2b65558a3c7 service nova] [instance: 363e48d7-8ed1-45d5-9443-d0b715952fca] Received event network-vif-deleted-1c634ca3-05f4-4ac4-a523-27b6d5d64e60 {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1148.623443] env[61957]: DEBUG nova.compute.manager [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] [instance: 8e04a098-25de-4d57-9e3c-ea44d234d57e] Instance has had 0 of 5 cleanup attempts {{(pid=61957) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1148.660898] env[61957]: DEBUG oslo_vmware.api [None req-024ba6d8-b0a9-4c02-8f92-8b55e68f2ea5 tempest-ServerAddressesNegativeTestJSON-995848772 tempest-ServerAddressesNegativeTestJSON-995848772-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52326222-7423-4119-159b-46a6d73d1ceb, 'name': SearchDatastore_Task, 'duration_secs': 0.00782} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1148.661715] env[61957]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a40027a0-1860-40e9-abf2-de7fbac823ec {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.666714] env[61957]: DEBUG oslo_vmware.api [None req-024ba6d8-b0a9-4c02-8f92-8b55e68f2ea5 tempest-ServerAddressesNegativeTestJSON-995848772 tempest-ServerAddressesNegativeTestJSON-995848772-project-member] Waiting for the task: (returnval){ [ 1148.666714] env[61957]: value = "session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52f7265d-3781-fa4e-dafe-5dfbe880fce6" [ 1148.666714] env[61957]: _type = "Task" [ 1148.666714] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1148.673506] env[61957]: DEBUG oslo_vmware.api [None req-024ba6d8-b0a9-4c02-8f92-8b55e68f2ea5 tempest-ServerAddressesNegativeTestJSON-995848772 tempest-ServerAddressesNegativeTestJSON-995848772-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52f7265d-3781-fa4e-dafe-5dfbe880fce6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1148.723923] env[61957]: DEBUG oslo_concurrency.lockutils [None req-b4f9918b-3627-43df-a7c4-9b2be3813404 tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.616s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1148.740857] env[61957]: INFO nova.scheduler.client.report [None req-b4f9918b-3627-43df-a7c4-9b2be3813404 tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] Deleted allocations for instance 363e48d7-8ed1-45d5-9443-d0b715952fca [ 1149.126401] env[61957]: DEBUG nova.compute.manager [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] [instance: 526ff179-62a6-4763-ab25-797617c4ed57] Instance has had 0 of 5 cleanup attempts {{(pid=61957) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1149.177854] env[61957]: DEBUG oslo_vmware.api [None req-024ba6d8-b0a9-4c02-8f92-8b55e68f2ea5 tempest-ServerAddressesNegativeTestJSON-995848772 tempest-ServerAddressesNegativeTestJSON-995848772-project-member] Task: {'id': session[52b36adf-61b6-ff57-a567-2f39fe530cd3]52f7265d-3781-fa4e-dafe-5dfbe880fce6, 'name': SearchDatastore_Task, 'duration_secs': 0.008574} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1149.178144] env[61957]: DEBUG oslo_concurrency.lockutils [None req-024ba6d8-b0a9-4c02-8f92-8b55e68f2ea5 tempest-ServerAddressesNegativeTestJSON-995848772 tempest-ServerAddressesNegativeTestJSON-995848772-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk" {{(pid=61957) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1149.178401] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-024ba6d8-b0a9-4c02-8f92-8b55e68f2ea5 tempest-ServerAddressesNegativeTestJSON-995848772 tempest-ServerAddressesNegativeTestJSON-995848772-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore2] 2fe04bcb-569c-48c3-b300-e67b4abe9fdf/2fe04bcb-569c-48c3-b300-e67b4abe9fdf.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1149.178660] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-34d1c770-6d61-447a-a7af-393e99902e0e {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.185614] env[61957]: DEBUG oslo_vmware.api [None req-024ba6d8-b0a9-4c02-8f92-8b55e68f2ea5 tempest-ServerAddressesNegativeTestJSON-995848772 tempest-ServerAddressesNegativeTestJSON-995848772-project-member] Waiting for the task: (returnval){ [ 1149.185614] env[61957]: value = "task-1278155" [ 1149.185614] env[61957]: _type = "Task" [ 1149.185614] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1149.192831] env[61957]: DEBUG oslo_vmware.api [None req-024ba6d8-b0a9-4c02-8f92-8b55e68f2ea5 tempest-ServerAddressesNegativeTestJSON-995848772 tempest-ServerAddressesNegativeTestJSON-995848772-project-member] Task: {'id': task-1278155, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1149.248843] env[61957]: DEBUG oslo_concurrency.lockutils [None req-b4f9918b-3627-43df-a7c4-9b2be3813404 tempest-ServerRescueTestJSONUnderV235-54199146 tempest-ServerRescueTestJSONUnderV235-54199146-project-member] Lock "363e48d7-8ed1-45d5-9443-d0b715952fca" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.982s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1149.629657] env[61957]: DEBUG nova.compute.manager [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] [instance: 281b9644-0e7d-48b9-a7b6-45fd6102d558] Instance has had 0 of 5 cleanup attempts {{(pid=61957) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1149.695859] env[61957]: DEBUG oslo_vmware.api [None req-024ba6d8-b0a9-4c02-8f92-8b55e68f2ea5 tempest-ServerAddressesNegativeTestJSON-995848772 tempest-ServerAddressesNegativeTestJSON-995848772-project-member] Task: {'id': task-1278155, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.424961} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1149.696150] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-024ba6d8-b0a9-4c02-8f92-8b55e68f2ea5 tempest-ServerAddressesNegativeTestJSON-995848772 tempest-ServerAddressesNegativeTestJSON-995848772-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/11c76a2c-f705-470a-ba9d-4657858bab38/11c76a2c-f705-470a-ba9d-4657858bab38.vmdk to [datastore2] 2fe04bcb-569c-48c3-b300-e67b4abe9fdf/2fe04bcb-569c-48c3-b300-e67b4abe9fdf.vmdk {{(pid=61957) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1149.696379] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-024ba6d8-b0a9-4c02-8f92-8b55e68f2ea5 tempest-ServerAddressesNegativeTestJSON-995848772 tempest-ServerAddressesNegativeTestJSON-995848772-project-member] [instance: 2fe04bcb-569c-48c3-b300-e67b4abe9fdf] Extending root virtual disk to 1048576 {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1149.696674] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-63368575-21f2-424c-934d-7d27264d45a6 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.702943] env[61957]: DEBUG oslo_vmware.api [None req-024ba6d8-b0a9-4c02-8f92-8b55e68f2ea5 tempest-ServerAddressesNegativeTestJSON-995848772 tempest-ServerAddressesNegativeTestJSON-995848772-project-member] Waiting for the task: (returnval){ [ 1149.702943] env[61957]: value = "task-1278156" [ 1149.702943] env[61957]: _type = "Task" [ 1149.702943] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1149.711377] env[61957]: DEBUG oslo_vmware.api [None req-024ba6d8-b0a9-4c02-8f92-8b55e68f2ea5 tempest-ServerAddressesNegativeTestJSON-995848772 tempest-ServerAddressesNegativeTestJSON-995848772-project-member] Task: {'id': task-1278156, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1150.134056] env[61957]: DEBUG nova.compute.manager [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] [instance: c23141ee-0cbb-4d1b-8390-c3073fe354f1] Instance has had 0 of 5 cleanup attempts {{(pid=61957) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1150.212786] env[61957]: DEBUG oslo_vmware.api [None req-024ba6d8-b0a9-4c02-8f92-8b55e68f2ea5 tempest-ServerAddressesNegativeTestJSON-995848772 tempest-ServerAddressesNegativeTestJSON-995848772-project-member] Task: {'id': task-1278156, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.055302} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1150.213092] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-024ba6d8-b0a9-4c02-8f92-8b55e68f2ea5 tempest-ServerAddressesNegativeTestJSON-995848772 tempest-ServerAddressesNegativeTestJSON-995848772-project-member] [instance: 2fe04bcb-569c-48c3-b300-e67b4abe9fdf] Extended root virtual disk {{(pid=61957) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1150.213963] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc926f07-af50-4e44-af0f-fc0b543b9321 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.236052] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-024ba6d8-b0a9-4c02-8f92-8b55e68f2ea5 tempest-ServerAddressesNegativeTestJSON-995848772 tempest-ServerAddressesNegativeTestJSON-995848772-project-member] [instance: 2fe04bcb-569c-48c3-b300-e67b4abe9fdf] Reconfiguring VM instance instance-0000006b to attach disk [datastore2] 2fe04bcb-569c-48c3-b300-e67b4abe9fdf/2fe04bcb-569c-48c3-b300-e67b4abe9fdf.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1150.236619] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d66cf863-e2ac-4320-a6ed-5e806356d2ac {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.255289] env[61957]: DEBUG oslo_vmware.api [None req-024ba6d8-b0a9-4c02-8f92-8b55e68f2ea5 tempest-ServerAddressesNegativeTestJSON-995848772 tempest-ServerAddressesNegativeTestJSON-995848772-project-member] Waiting for the task: (returnval){ [ 1150.255289] env[61957]: value = "task-1278157" [ 1150.255289] env[61957]: _type = "Task" [ 1150.255289] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1150.262570] env[61957]: DEBUG oslo_vmware.api [None req-024ba6d8-b0a9-4c02-8f92-8b55e68f2ea5 tempest-ServerAddressesNegativeTestJSON-995848772 tempest-ServerAddressesNegativeTestJSON-995848772-project-member] Task: {'id': task-1278157, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1150.636764] env[61957]: DEBUG nova.compute.manager [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] [instance: f66db265-887e-4d61-b848-c609e5c884cb] Instance has had 0 of 5 cleanup attempts {{(pid=61957) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1150.766053] env[61957]: DEBUG oslo_vmware.api [None req-024ba6d8-b0a9-4c02-8f92-8b55e68f2ea5 tempest-ServerAddressesNegativeTestJSON-995848772 tempest-ServerAddressesNegativeTestJSON-995848772-project-member] Task: {'id': task-1278157, 'name': ReconfigVM_Task, 'duration_secs': 0.277288} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1150.766382] env[61957]: DEBUG nova.virt.vmwareapi.volumeops [None req-024ba6d8-b0a9-4c02-8f92-8b55e68f2ea5 tempest-ServerAddressesNegativeTestJSON-995848772 tempest-ServerAddressesNegativeTestJSON-995848772-project-member] [instance: 2fe04bcb-569c-48c3-b300-e67b4abe9fdf] Reconfigured VM instance instance-0000006b to attach disk [datastore2] 2fe04bcb-569c-48c3-b300-e67b4abe9fdf/2fe04bcb-569c-48c3-b300-e67b4abe9fdf.vmdk or device None with type sparse {{(pid=61957) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1150.767068] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f3c77d95-a8a5-4c36-8bd6-f160bc9abf56 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.773827] env[61957]: DEBUG oslo_vmware.api [None req-024ba6d8-b0a9-4c02-8f92-8b55e68f2ea5 tempest-ServerAddressesNegativeTestJSON-995848772 tempest-ServerAddressesNegativeTestJSON-995848772-project-member] Waiting for the task: (returnval){ [ 1150.773827] env[61957]: value = "task-1278158" [ 1150.773827] env[61957]: _type = "Task" [ 1150.773827] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1150.781761] env[61957]: DEBUG oslo_vmware.api [None req-024ba6d8-b0a9-4c02-8f92-8b55e68f2ea5 tempest-ServerAddressesNegativeTestJSON-995848772 tempest-ServerAddressesNegativeTestJSON-995848772-project-member] Task: {'id': task-1278158, 'name': Rename_Task} progress is 5%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1151.139936] env[61957]: DEBUG nova.compute.manager [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] [instance: 6438fe40-046c-45d5-9986-8f182ecde49f] Instance has had 0 of 5 cleanup attempts {{(pid=61957) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1151.284247] env[61957]: DEBUG oslo_vmware.api [None req-024ba6d8-b0a9-4c02-8f92-8b55e68f2ea5 tempest-ServerAddressesNegativeTestJSON-995848772 tempest-ServerAddressesNegativeTestJSON-995848772-project-member] Task: {'id': task-1278158, 'name': Rename_Task, 'duration_secs': 0.13708} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1151.284522] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-024ba6d8-b0a9-4c02-8f92-8b55e68f2ea5 tempest-ServerAddressesNegativeTestJSON-995848772 tempest-ServerAddressesNegativeTestJSON-995848772-project-member] [instance: 2fe04bcb-569c-48c3-b300-e67b4abe9fdf] Powering on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1151.284765] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5cf56902-fed0-459b-a740-b25b79b6a3b5 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.291011] env[61957]: DEBUG oslo_vmware.api [None req-024ba6d8-b0a9-4c02-8f92-8b55e68f2ea5 tempest-ServerAddressesNegativeTestJSON-995848772 tempest-ServerAddressesNegativeTestJSON-995848772-project-member] Waiting for the task: (returnval){ [ 1151.291011] env[61957]: value = "task-1278159" [ 1151.291011] env[61957]: _type = "Task" [ 1151.291011] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1151.299773] env[61957]: DEBUG oslo_vmware.api [None req-024ba6d8-b0a9-4c02-8f92-8b55e68f2ea5 tempest-ServerAddressesNegativeTestJSON-995848772 tempest-ServerAddressesNegativeTestJSON-995848772-project-member] Task: {'id': task-1278159, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1151.643995] env[61957]: DEBUG nova.compute.manager [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] [instance: 0adae8e1-8c2f-4110-805b-1f286debc833] Instance has had 0 of 5 cleanup attempts {{(pid=61957) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1151.800510] env[61957]: DEBUG oslo_vmware.api [None req-024ba6d8-b0a9-4c02-8f92-8b55e68f2ea5 tempest-ServerAddressesNegativeTestJSON-995848772 tempest-ServerAddressesNegativeTestJSON-995848772-project-member] Task: {'id': task-1278159, 'name': PowerOnVM_Task, 'duration_secs': 0.42675} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1151.800780] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-024ba6d8-b0a9-4c02-8f92-8b55e68f2ea5 tempest-ServerAddressesNegativeTestJSON-995848772 tempest-ServerAddressesNegativeTestJSON-995848772-project-member] [instance: 2fe04bcb-569c-48c3-b300-e67b4abe9fdf] Powered on the VM {{(pid=61957) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1151.800981] env[61957]: INFO nova.compute.manager [None req-024ba6d8-b0a9-4c02-8f92-8b55e68f2ea5 tempest-ServerAddressesNegativeTestJSON-995848772 tempest-ServerAddressesNegativeTestJSON-995848772-project-member] [instance: 2fe04bcb-569c-48c3-b300-e67b4abe9fdf] Took 6.64 seconds to spawn the instance on the hypervisor. [ 1151.801180] env[61957]: DEBUG nova.compute.manager [None req-024ba6d8-b0a9-4c02-8f92-8b55e68f2ea5 tempest-ServerAddressesNegativeTestJSON-995848772 tempest-ServerAddressesNegativeTestJSON-995848772-project-member] [instance: 2fe04bcb-569c-48c3-b300-e67b4abe9fdf] Checking state {{(pid=61957) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1151.801928] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbedbaf6-ad91-466f-9856-6595366bd98b {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.147230] env[61957]: DEBUG nova.compute.manager [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] [instance: 37a9e1c1-9f6c-4047-9ff5-e141d0ca383e] Instance has had 0 of 5 cleanup attempts {{(pid=61957) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1152.316968] env[61957]: INFO nova.compute.manager [None req-024ba6d8-b0a9-4c02-8f92-8b55e68f2ea5 tempest-ServerAddressesNegativeTestJSON-995848772 tempest-ServerAddressesNegativeTestJSON-995848772-project-member] [instance: 2fe04bcb-569c-48c3-b300-e67b4abe9fdf] Took 11.32 seconds to build instance. [ 1152.650683] env[61957]: DEBUG nova.compute.manager [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] [instance: 37ce46a1-8f9a-4d15-bd81-e40845a0e48a] Instance has had 0 of 5 cleanup attempts {{(pid=61957) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1152.818552] env[61957]: DEBUG oslo_concurrency.lockutils [None req-024ba6d8-b0a9-4c02-8f92-8b55e68f2ea5 tempest-ServerAddressesNegativeTestJSON-995848772 tempest-ServerAddressesNegativeTestJSON-995848772-project-member] Lock "2fe04bcb-569c-48c3-b300-e67b4abe9fdf" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 12.834s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1153.153866] env[61957]: DEBUG nova.compute.manager [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] [instance: ead6aae1-36b5-4f57-9129-3bb02cf103ce] Instance has had 0 of 5 cleanup attempts {{(pid=61957) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1153.197315] env[61957]: DEBUG oslo_concurrency.lockutils [None req-c8c54d25-41af-4d85-93b7-80b04c1c19e6 tempest-ServerAddressesNegativeTestJSON-995848772 tempest-ServerAddressesNegativeTestJSON-995848772-project-member] Acquiring lock "2fe04bcb-569c-48c3-b300-e67b4abe9fdf" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1153.197586] env[61957]: DEBUG oslo_concurrency.lockutils [None req-c8c54d25-41af-4d85-93b7-80b04c1c19e6 tempest-ServerAddressesNegativeTestJSON-995848772 tempest-ServerAddressesNegativeTestJSON-995848772-project-member] Lock "2fe04bcb-569c-48c3-b300-e67b4abe9fdf" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1153.197842] env[61957]: DEBUG oslo_concurrency.lockutils [None req-c8c54d25-41af-4d85-93b7-80b04c1c19e6 tempest-ServerAddressesNegativeTestJSON-995848772 tempest-ServerAddressesNegativeTestJSON-995848772-project-member] Acquiring lock "2fe04bcb-569c-48c3-b300-e67b4abe9fdf-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1153.198079] env[61957]: DEBUG oslo_concurrency.lockutils [None req-c8c54d25-41af-4d85-93b7-80b04c1c19e6 tempest-ServerAddressesNegativeTestJSON-995848772 tempest-ServerAddressesNegativeTestJSON-995848772-project-member] Lock "2fe04bcb-569c-48c3-b300-e67b4abe9fdf-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1153.198260] env[61957]: DEBUG oslo_concurrency.lockutils [None req-c8c54d25-41af-4d85-93b7-80b04c1c19e6 tempest-ServerAddressesNegativeTestJSON-995848772 tempest-ServerAddressesNegativeTestJSON-995848772-project-member] Lock "2fe04bcb-569c-48c3-b300-e67b4abe9fdf-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1153.200586] env[61957]: INFO nova.compute.manager [None req-c8c54d25-41af-4d85-93b7-80b04c1c19e6 tempest-ServerAddressesNegativeTestJSON-995848772 tempest-ServerAddressesNegativeTestJSON-995848772-project-member] [instance: 2fe04bcb-569c-48c3-b300-e67b4abe9fdf] Terminating instance [ 1153.202357] env[61957]: DEBUG nova.compute.manager [None req-c8c54d25-41af-4d85-93b7-80b04c1c19e6 tempest-ServerAddressesNegativeTestJSON-995848772 tempest-ServerAddressesNegativeTestJSON-995848772-project-member] [instance: 2fe04bcb-569c-48c3-b300-e67b4abe9fdf] Start destroying the instance on the hypervisor. {{(pid=61957) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1153.202565] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-c8c54d25-41af-4d85-93b7-80b04c1c19e6 tempest-ServerAddressesNegativeTestJSON-995848772 tempest-ServerAddressesNegativeTestJSON-995848772-project-member] [instance: 2fe04bcb-569c-48c3-b300-e67b4abe9fdf] Destroying instance {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1153.203425] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf725913-d482-45c3-a43e-8b8ca3189823 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.210654] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-c8c54d25-41af-4d85-93b7-80b04c1c19e6 tempest-ServerAddressesNegativeTestJSON-995848772 tempest-ServerAddressesNegativeTestJSON-995848772-project-member] [instance: 2fe04bcb-569c-48c3-b300-e67b4abe9fdf] Powering off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1153.210871] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c8b1a6fb-43ff-491e-ba95-45c9a1dc182e {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.218360] env[61957]: DEBUG oslo_vmware.api [None req-c8c54d25-41af-4d85-93b7-80b04c1c19e6 tempest-ServerAddressesNegativeTestJSON-995848772 tempest-ServerAddressesNegativeTestJSON-995848772-project-member] Waiting for the task: (returnval){ [ 1153.218360] env[61957]: value = "task-1278160" [ 1153.218360] env[61957]: _type = "Task" [ 1153.218360] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1153.225502] env[61957]: DEBUG oslo_vmware.api [None req-c8c54d25-41af-4d85-93b7-80b04c1c19e6 tempest-ServerAddressesNegativeTestJSON-995848772 tempest-ServerAddressesNegativeTestJSON-995848772-project-member] Task: {'id': task-1278160, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1153.657444] env[61957]: DEBUG nova.compute.manager [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] [instance: 4951132e-7247-4772-8f88-3664c6a7e61e] Instance has had 0 of 5 cleanup attempts {{(pid=61957) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1153.728321] env[61957]: DEBUG oslo_vmware.api [None req-c8c54d25-41af-4d85-93b7-80b04c1c19e6 tempest-ServerAddressesNegativeTestJSON-995848772 tempest-ServerAddressesNegativeTestJSON-995848772-project-member] Task: {'id': task-1278160, 'name': PowerOffVM_Task, 'duration_secs': 0.169566} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1153.728639] env[61957]: DEBUG nova.virt.vmwareapi.vm_util [None req-c8c54d25-41af-4d85-93b7-80b04c1c19e6 tempest-ServerAddressesNegativeTestJSON-995848772 tempest-ServerAddressesNegativeTestJSON-995848772-project-member] [instance: 2fe04bcb-569c-48c3-b300-e67b4abe9fdf] Powered off the VM {{(pid=61957) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1153.728847] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-c8c54d25-41af-4d85-93b7-80b04c1c19e6 tempest-ServerAddressesNegativeTestJSON-995848772 tempest-ServerAddressesNegativeTestJSON-995848772-project-member] [instance: 2fe04bcb-569c-48c3-b300-e67b4abe9fdf] Unregistering the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1153.729146] env[61957]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ca0c2548-0968-4ca4-aa74-231fc9c55e02 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.786413] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-c8c54d25-41af-4d85-93b7-80b04c1c19e6 tempest-ServerAddressesNegativeTestJSON-995848772 tempest-ServerAddressesNegativeTestJSON-995848772-project-member] [instance: 2fe04bcb-569c-48c3-b300-e67b4abe9fdf] Unregistered the VM {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1153.786680] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-c8c54d25-41af-4d85-93b7-80b04c1c19e6 tempest-ServerAddressesNegativeTestJSON-995848772 tempest-ServerAddressesNegativeTestJSON-995848772-project-member] [instance: 2fe04bcb-569c-48c3-b300-e67b4abe9fdf] Deleting contents of the VM from datastore datastore2 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1153.786900] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-c8c54d25-41af-4d85-93b7-80b04c1c19e6 tempest-ServerAddressesNegativeTestJSON-995848772 tempest-ServerAddressesNegativeTestJSON-995848772-project-member] Deleting the datastore file [datastore2] 2fe04bcb-569c-48c3-b300-e67b4abe9fdf {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1153.787212] env[61957]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6632432d-5f18-4319-abd1-31fc55a89852 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.793386] env[61957]: DEBUG oslo_vmware.api [None req-c8c54d25-41af-4d85-93b7-80b04c1c19e6 tempest-ServerAddressesNegativeTestJSON-995848772 tempest-ServerAddressesNegativeTestJSON-995848772-project-member] Waiting for the task: (returnval){ [ 1153.793386] env[61957]: value = "task-1278162" [ 1153.793386] env[61957]: _type = "Task" [ 1153.793386] env[61957]: } to complete. {{(pid=61957) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1153.801763] env[61957]: DEBUG oslo_vmware.api [None req-c8c54d25-41af-4d85-93b7-80b04c1c19e6 tempest-ServerAddressesNegativeTestJSON-995848772 tempest-ServerAddressesNegativeTestJSON-995848772-project-member] Task: {'id': task-1278162, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1154.161124] env[61957]: DEBUG nova.compute.manager [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] [instance: 17f00b20-9d3b-45e6-919d-6fab9999ec77] Instance has had 0 of 5 cleanup attempts {{(pid=61957) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1154.304482] env[61957]: DEBUG oslo_vmware.api [None req-c8c54d25-41af-4d85-93b7-80b04c1c19e6 tempest-ServerAddressesNegativeTestJSON-995848772 tempest-ServerAddressesNegativeTestJSON-995848772-project-member] Task: {'id': task-1278162, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.1692} completed successfully. {{(pid=61957) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1154.304762] env[61957]: DEBUG nova.virt.vmwareapi.ds_util [None req-c8c54d25-41af-4d85-93b7-80b04c1c19e6 tempest-ServerAddressesNegativeTestJSON-995848772 tempest-ServerAddressesNegativeTestJSON-995848772-project-member] Deleted the datastore file {{(pid=61957) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1154.304950] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-c8c54d25-41af-4d85-93b7-80b04c1c19e6 tempest-ServerAddressesNegativeTestJSON-995848772 tempest-ServerAddressesNegativeTestJSON-995848772-project-member] [instance: 2fe04bcb-569c-48c3-b300-e67b4abe9fdf] Deleted contents of the VM from datastore datastore2 {{(pid=61957) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1154.305142] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-c8c54d25-41af-4d85-93b7-80b04c1c19e6 tempest-ServerAddressesNegativeTestJSON-995848772 tempest-ServerAddressesNegativeTestJSON-995848772-project-member] [instance: 2fe04bcb-569c-48c3-b300-e67b4abe9fdf] Instance destroyed {{(pid=61957) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1154.305320] env[61957]: INFO nova.compute.manager [None req-c8c54d25-41af-4d85-93b7-80b04c1c19e6 tempest-ServerAddressesNegativeTestJSON-995848772 tempest-ServerAddressesNegativeTestJSON-995848772-project-member] [instance: 2fe04bcb-569c-48c3-b300-e67b4abe9fdf] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1154.305564] env[61957]: DEBUG oslo.service.loopingcall [None req-c8c54d25-41af-4d85-93b7-80b04c1c19e6 tempest-ServerAddressesNegativeTestJSON-995848772 tempest-ServerAddressesNegativeTestJSON-995848772-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61957) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1154.305760] env[61957]: DEBUG nova.compute.manager [-] [instance: 2fe04bcb-569c-48c3-b300-e67b4abe9fdf] Deallocating network for instance {{(pid=61957) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1154.305854] env[61957]: DEBUG nova.network.neutron [-] [instance: 2fe04bcb-569c-48c3-b300-e67b4abe9fdf] deallocate_for_instance() {{(pid=61957) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1154.557193] env[61957]: DEBUG nova.compute.manager [req-bf9c70c7-bbcf-4b97-b59e-ea934e0a3c1e req-668b19a5-3538-4471-87dc-47417b9fa45c service nova] [instance: 2fe04bcb-569c-48c3-b300-e67b4abe9fdf] Received event network-vif-deleted-ff49168a-f538-451b-84d1-a2efd419a3ba {{(pid=61957) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1154.557193] env[61957]: INFO nova.compute.manager [req-bf9c70c7-bbcf-4b97-b59e-ea934e0a3c1e req-668b19a5-3538-4471-87dc-47417b9fa45c service nova] [instance: 2fe04bcb-569c-48c3-b300-e67b4abe9fdf] Neutron deleted interface ff49168a-f538-451b-84d1-a2efd419a3ba; detaching it from the instance and deleting it from the info cache [ 1154.557333] env[61957]: DEBUG nova.network.neutron [req-bf9c70c7-bbcf-4b97-b59e-ea934e0a3c1e req-668b19a5-3538-4471-87dc-47417b9fa45c service nova] [instance: 2fe04bcb-569c-48c3-b300-e67b4abe9fdf] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1154.665626] env[61957]: DEBUG nova.compute.manager [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] [instance: 14ba830d-4e2c-4e9a-a059-3c86209f0127] Instance has had 0 of 5 cleanup attempts {{(pid=61957) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1155.035977] env[61957]: DEBUG nova.network.neutron [-] [instance: 2fe04bcb-569c-48c3-b300-e67b4abe9fdf] Updating instance_info_cache with network_info: [] {{(pid=61957) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1155.060029] env[61957]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-87067141-63d0-46cc-aa33-f5d5db82542e {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.069472] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68ec5711-2565-4d23-aec6-c1b978db7cbf {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.091869] env[61957]: DEBUG nova.compute.manager [req-bf9c70c7-bbcf-4b97-b59e-ea934e0a3c1e req-668b19a5-3538-4471-87dc-47417b9fa45c service nova] [instance: 2fe04bcb-569c-48c3-b300-e67b4abe9fdf] Detach interface failed, port_id=ff49168a-f538-451b-84d1-a2efd419a3ba, reason: Instance 2fe04bcb-569c-48c3-b300-e67b4abe9fdf could not be found. {{(pid=61957) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1155.169187] env[61957]: DEBUG nova.compute.manager [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] [instance: e9ac118f-08b3-430b-848c-461c2b2e3e02] Instance has had 0 of 5 cleanup attempts {{(pid=61957) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1155.538361] env[61957]: INFO nova.compute.manager [-] [instance: 2fe04bcb-569c-48c3-b300-e67b4abe9fdf] Took 1.23 seconds to deallocate network for instance. [ 1155.672098] env[61957]: DEBUG nova.compute.manager [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] [instance: 615dbbed-2b02-4351-9e03-8c13f424a133] Instance has had 0 of 5 cleanup attempts {{(pid=61957) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1156.045467] env[61957]: DEBUG oslo_concurrency.lockutils [None req-c8c54d25-41af-4d85-93b7-80b04c1c19e6 tempest-ServerAddressesNegativeTestJSON-995848772 tempest-ServerAddressesNegativeTestJSON-995848772-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1156.045467] env[61957]: DEBUG oslo_concurrency.lockutils [None req-c8c54d25-41af-4d85-93b7-80b04c1c19e6 tempest-ServerAddressesNegativeTestJSON-995848772 tempest-ServerAddressesNegativeTestJSON-995848772-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1156.045616] env[61957]: DEBUG nova.objects.instance [None req-c8c54d25-41af-4d85-93b7-80b04c1c19e6 tempest-ServerAddressesNegativeTestJSON-995848772 tempest-ServerAddressesNegativeTestJSON-995848772-project-member] Lazy-loading 'resources' on Instance uuid 2fe04bcb-569c-48c3-b300-e67b4abe9fdf {{(pid=61957) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1156.175578] env[61957]: DEBUG nova.compute.manager [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] [instance: e0249936-d616-4ffb-8f77-d8107633c42a] Instance has had 0 of 5 cleanup attempts {{(pid=61957) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1156.580897] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11a4e347-d2dd-416b-83a8-dd1cec6ac88e {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.588802] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b47b5aa9-71ec-41b5-9e89-d33e8b2aa5e1 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.618825] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aacbae16-9dd1-43c5-8e6c-ff8de7b35c06 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.625493] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ef2211a-b89e-4171-b3b3-c26c9df0a87e {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.638528] env[61957]: DEBUG nova.compute.provider_tree [None req-c8c54d25-41af-4d85-93b7-80b04c1c19e6 tempest-ServerAddressesNegativeTestJSON-995848772 tempest-ServerAddressesNegativeTestJSON-995848772-project-member] Inventory has not changed in ProviderTree for provider: 0ceb6c9e-61c6-496d-8579-9d32627e96da {{(pid=61957) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1156.678323] env[61957]: DEBUG nova.compute.manager [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] [instance: 6e7bd89c-2c2a-450a-9858-3526d96c28ab] Instance has had 0 of 5 cleanup attempts {{(pid=61957) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1157.141634] env[61957]: DEBUG nova.scheduler.client.report [None req-c8c54d25-41af-4d85-93b7-80b04c1c19e6 tempest-ServerAddressesNegativeTestJSON-995848772 tempest-ServerAddressesNegativeTestJSON-995848772-project-member] Inventory has not changed for provider 0ceb6c9e-61c6-496d-8579-9d32627e96da based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61957) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1157.181569] env[61957]: DEBUG oslo_service.periodic_task [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=61957) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1157.181733] env[61957]: DEBUG nova.compute.manager [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Cleaning up deleted instances with incomplete migration {{(pid=61957) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11264}} [ 1157.647164] env[61957]: DEBUG oslo_concurrency.lockutils [None req-c8c54d25-41af-4d85-93b7-80b04c1c19e6 tempest-ServerAddressesNegativeTestJSON-995848772 tempest-ServerAddressesNegativeTestJSON-995848772-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.602s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1157.666910] env[61957]: INFO nova.scheduler.client.report [None req-c8c54d25-41af-4d85-93b7-80b04c1c19e6 tempest-ServerAddressesNegativeTestJSON-995848772 tempest-ServerAddressesNegativeTestJSON-995848772-project-member] Deleted allocations for instance 2fe04bcb-569c-48c3-b300-e67b4abe9fdf [ 1157.684069] env[61957]: DEBUG oslo_service.periodic_task [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=61957) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1158.174158] env[61957]: DEBUG oslo_concurrency.lockutils [None req-c8c54d25-41af-4d85-93b7-80b04c1c19e6 tempest-ServerAddressesNegativeTestJSON-995848772 tempest-ServerAddressesNegativeTestJSON-995848772-project-member] Lock "2fe04bcb-569c-48c3-b300-e67b4abe9fdf" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.976s {{(pid=61957) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1160.484045] env[61957]: DEBUG oslo_service.periodic_task [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Running periodic task ComputeManager._sync_power_states {{(pid=61957) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1160.987061] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Getting list of instances from cluster (obj){ [ 1160.987061] env[61957]: value = "domain-c8" [ 1160.987061] env[61957]: _type = "ClusterComputeResource" [ 1160.987061] env[61957]: } {{(pid=61957) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2122}} [ 1160.988740] env[61957]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a9c8053-9c84-47fe-ad22-dcb86831eeb6 {{(pid=61957) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.997972] env[61957]: DEBUG nova.virt.vmwareapi.vmops [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Got total of 0 instances {{(pid=61957) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2131}} [ 1176.641900] env[61957]: DEBUG oslo_service.periodic_task [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61957) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1177.122510] env[61957]: DEBUG oslo_service.periodic_task [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61957) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1179.126317] env[61957]: DEBUG oslo_service.periodic_task [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61957) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1179.126717] env[61957]: DEBUG nova.compute.manager [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Starting heal instance info cache {{(pid=61957) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 1179.126717] env[61957]: DEBUG nova.compute.manager [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Rebuilding the list of instances to heal {{(pid=61957) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9930}} [ 1179.629587] env[61957]: DEBUG nova.compute.manager [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Didn't find any instances for network info cache update. {{(pid=61957) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10012}} [ 1180.126692] env[61957]: DEBUG oslo_service.periodic_task [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61957) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1180.127064] env[61957]: DEBUG oslo_service.periodic_task [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61957) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1180.127109] env[61957]: DEBUG oslo_service.periodic_task [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61957) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1180.127250] env[61957]: DEBUG nova.compute.manager [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61957) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 1181.127612] env[61957]: DEBUG oslo_service.periodic_task [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61957) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1181.127982] env[61957]: DEBUG oslo_service.periodic_task [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61957) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1181.128097] env[61957]: DEBUG oslo_service.periodic_task [None req-2c26e0b4-2a92-4d70-a750-4c42b70e39d2 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61957) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}}